00:00:00.000 Started by upstream project "autotest-spdk-master-vs-dpdk-v22.11" build number 2405 00:00:00.000 originally caused by: 00:00:00.001 Started by upstream project "nightly-trigger" build number 3670 00:00:00.001 originally caused by: 00:00:00.001 Started by timer 00:00:00.017 Checking out git https://review.spdk.io/gerrit/a/build_pool/jenkins_build_pool into /var/jenkins_home/workspace/raid-vg-autotest_script/33b20b30f0a51e6b52980845e0f6aa336787973ad45e341fbbf98d1b65b265d4 to read jbp/jenkins/jjb-config/jobs/autotest-downstream/autotest-vg.groovy 00:00:00.018 The recommended git tool is: git 00:00:00.019 using credential 00000000-0000-0000-0000-000000000002 00:00:00.021 > git rev-parse --resolve-git-dir /var/jenkins_home/workspace/raid-vg-autotest_script/33b20b30f0a51e6b52980845e0f6aa336787973ad45e341fbbf98d1b65b265d4/jbp/.git # timeout=10 00:00:00.037 Fetching changes from the remote Git repository 00:00:00.043 > git config remote.origin.url https://review.spdk.io/gerrit/a/build_pool/jenkins_build_pool # timeout=10 00:00:00.055 Using shallow fetch with depth 1 00:00:00.055 Fetching upstream changes from https://review.spdk.io/gerrit/a/build_pool/jenkins_build_pool 00:00:00.055 > git --version # timeout=10 00:00:00.070 > git --version # 'git version 2.39.2' 00:00:00.070 using GIT_ASKPASS to set credentials SPDKCI HTTPS Credentials 00:00:00.087 Setting http proxy: proxy-dmz.intel.com:911 00:00:00.087 > git fetch --tags --force --progress --depth=1 -- https://review.spdk.io/gerrit/a/build_pool/jenkins_build_pool refs/heads/master # timeout=5 00:00:02.256 > git rev-parse origin/FETCH_HEAD^{commit} # timeout=10 00:00:02.265 > git rev-parse FETCH_HEAD^{commit} # timeout=10 00:00:02.275 Checking out Revision db4637e8b949f278f369ec13f70585206ccd9507 (FETCH_HEAD) 00:00:02.275 > git config core.sparsecheckout # timeout=10 00:00:02.285 > git read-tree -mu HEAD # timeout=10 00:00:02.299 > git checkout -f db4637e8b949f278f369ec13f70585206ccd9507 # timeout=5 00:00:02.324 Commit message: "jenkins/jjb-config: Add missing SPDK_TEST_NVME_INTERRUPT flag" 00:00:02.325 > git rev-list --no-walk db4637e8b949f278f369ec13f70585206ccd9507 # timeout=10 00:00:02.538 [Pipeline] Start of Pipeline 00:00:02.566 [Pipeline] library 00:00:02.569 Loading library shm_lib@master 00:00:02.569 Library shm_lib@master is cached. Copying from home. 00:00:02.590 [Pipeline] node 00:00:02.603 Running on VM-host-WFP7 in /var/jenkins/workspace/raid-vg-autotest 00:00:02.604 [Pipeline] { 00:00:02.616 [Pipeline] catchError 00:00:02.617 [Pipeline] { 00:00:02.632 [Pipeline] wrap 00:00:02.641 [Pipeline] { 00:00:02.651 [Pipeline] stage 00:00:02.653 [Pipeline] { (Prologue) 00:00:02.675 [Pipeline] echo 00:00:02.677 Node: VM-host-WFP7 00:00:02.685 [Pipeline] cleanWs 00:00:02.699 [WS-CLEANUP] Deleting project workspace... 00:00:02.699 [WS-CLEANUP] Deferred wipeout is used... 00:00:02.707 [WS-CLEANUP] done 00:00:02.923 [Pipeline] setCustomBuildProperty 00:00:02.998 [Pipeline] httpRequest 00:00:03.313 [Pipeline] echo 00:00:03.314 Sorcerer 10.211.164.20 is alive 00:00:03.321 [Pipeline] retry 00:00:03.323 [Pipeline] { 00:00:03.333 [Pipeline] httpRequest 00:00:03.337 HttpMethod: GET 00:00:03.338 URL: http://10.211.164.20/packages/jbp_db4637e8b949f278f369ec13f70585206ccd9507.tar.gz 00:00:03.338 Sending request to url: http://10.211.164.20/packages/jbp_db4637e8b949f278f369ec13f70585206ccd9507.tar.gz 00:00:03.339 Response Code: HTTP/1.1 200 OK 00:00:03.339 Success: Status code 200 is in the accepted range: 200,404 00:00:03.339 Saving response body to /var/jenkins/workspace/raid-vg-autotest/jbp_db4637e8b949f278f369ec13f70585206ccd9507.tar.gz 00:00:03.485 [Pipeline] } 00:00:03.500 [Pipeline] // retry 00:00:03.505 [Pipeline] sh 00:00:03.786 + tar --no-same-owner -xf jbp_db4637e8b949f278f369ec13f70585206ccd9507.tar.gz 00:00:03.801 [Pipeline] httpRequest 00:00:04.266 [Pipeline] echo 00:00:04.267 Sorcerer 10.211.164.20 is alive 00:00:04.275 [Pipeline] retry 00:00:04.277 [Pipeline] { 00:00:04.288 [Pipeline] httpRequest 00:00:04.293 HttpMethod: GET 00:00:04.294 URL: http://10.211.164.20/packages/spdk_2f2acf4eb25cee406c156120cee22721275ca7fd.tar.gz 00:00:04.294 Sending request to url: http://10.211.164.20/packages/spdk_2f2acf4eb25cee406c156120cee22721275ca7fd.tar.gz 00:00:04.295 Response Code: HTTP/1.1 200 OK 00:00:04.295 Success: Status code 200 is in the accepted range: 200,404 00:00:04.296 Saving response body to /var/jenkins/workspace/raid-vg-autotest/spdk_2f2acf4eb25cee406c156120cee22721275ca7fd.tar.gz 00:00:20.255 [Pipeline] } 00:00:20.273 [Pipeline] // retry 00:00:20.281 [Pipeline] sh 00:00:20.568 + tar --no-same-owner -xf spdk_2f2acf4eb25cee406c156120cee22721275ca7fd.tar.gz 00:00:23.127 [Pipeline] sh 00:00:23.415 + git -C spdk log --oneline -n5 00:00:23.415 2f2acf4eb doc: move nvmf_tracing.md to tracing.md 00:00:23.415 5592070b3 doc: update nvmf_tracing.md 00:00:23.415 5ca6db5da nvme_spec: Add SPDK_NVME_IO_FLAGS_PRCHK_MASK 00:00:23.415 f7ce15267 bdev: Insert or overwrite metadata using bounce/accel buffer if NVMe PRACT is set 00:00:23.415 aa58c9e0b dif: Add spdk_dif_pi_format_get_size() to use for NVMe PRACT 00:00:23.438 [Pipeline] withCredentials 00:00:23.451 > git --version # timeout=10 00:00:23.466 > git --version # 'git version 2.39.2' 00:00:23.486 Masking supported pattern matches of $GIT_PASSWORD or $GIT_ASKPASS 00:00:23.488 [Pipeline] { 00:00:23.497 [Pipeline] retry 00:00:23.499 [Pipeline] { 00:00:23.514 [Pipeline] sh 00:00:23.801 + git ls-remote http://dpdk.org/git/dpdk-stable v22.11.4 00:00:24.074 [Pipeline] } 00:00:24.102 [Pipeline] // retry 00:00:24.106 [Pipeline] } 00:00:24.122 [Pipeline] // withCredentials 00:00:24.131 [Pipeline] httpRequest 00:00:24.519 [Pipeline] echo 00:00:24.521 Sorcerer 10.211.164.20 is alive 00:00:24.531 [Pipeline] retry 00:00:24.533 [Pipeline] { 00:00:24.548 [Pipeline] httpRequest 00:00:24.554 HttpMethod: GET 00:00:24.554 URL: http://10.211.164.20/packages/dpdk_fee0f13c213d0584f0c42a51d0e0625d99a0b2f1.tar.gz 00:00:24.555 Sending request to url: http://10.211.164.20/packages/dpdk_fee0f13c213d0584f0c42a51d0e0625d99a0b2f1.tar.gz 00:00:24.571 Response Code: HTTP/1.1 200 OK 00:00:24.572 Success: Status code 200 is in the accepted range: 200,404 00:00:24.572 Saving response body to /var/jenkins/workspace/raid-vg-autotest/dpdk_fee0f13c213d0584f0c42a51d0e0625d99a0b2f1.tar.gz 00:00:38.536 [Pipeline] } 00:00:38.559 [Pipeline] // retry 00:00:38.569 [Pipeline] sh 00:00:38.875 + tar --no-same-owner -xf dpdk_fee0f13c213d0584f0c42a51d0e0625d99a0b2f1.tar.gz 00:00:40.286 [Pipeline] sh 00:00:40.577 + git -C dpdk log --oneline -n5 00:00:40.577 caf0f5d395 version: 22.11.4 00:00:40.577 7d6f1cc05f Revert "net/iavf: fix abnormal disable HW interrupt" 00:00:40.577 dc9c799c7d vhost: fix missing spinlock unlock 00:00:40.578 4307659a90 net/mlx5: fix LACP redirection in Rx domain 00:00:40.578 6ef77f2a5e net/gve: fix RX buffer size alignment 00:00:40.596 [Pipeline] writeFile 00:00:40.611 [Pipeline] sh 00:00:40.897 + jbp/jenkins/jjb-config/jobs/scripts/autorun_quirks.sh 00:00:40.910 [Pipeline] sh 00:00:41.195 + cat autorun-spdk.conf 00:00:41.195 SPDK_RUN_FUNCTIONAL_TEST=1 00:00:41.195 SPDK_RUN_ASAN=1 00:00:41.195 SPDK_RUN_UBSAN=1 00:00:41.195 SPDK_TEST_RAID=1 00:00:41.195 SPDK_TEST_NATIVE_DPDK=v22.11.4 00:00:41.195 SPDK_RUN_EXTERNAL_DPDK=/home/vagrant/spdk_repo/dpdk/build 00:00:41.195 SPDK_ABI_DIR=/home/vagrant/spdk_repo/spdk-abi 00:00:41.202 RUN_NIGHTLY=1 00:00:41.203 [Pipeline] } 00:00:41.211 [Pipeline] // stage 00:00:41.220 [Pipeline] stage 00:00:41.222 [Pipeline] { (Run VM) 00:00:41.232 [Pipeline] sh 00:00:41.525 + jbp/jenkins/jjb-config/jobs/scripts/prepare_nvme.sh 00:00:41.525 + echo 'Start stage prepare_nvme.sh' 00:00:41.525 Start stage prepare_nvme.sh 00:00:41.525 + [[ -n 0 ]] 00:00:41.525 + disk_prefix=ex0 00:00:41.525 + [[ -n /var/jenkins/workspace/raid-vg-autotest ]] 00:00:41.525 + [[ -e /var/jenkins/workspace/raid-vg-autotest/autorun-spdk.conf ]] 00:00:41.525 + source /var/jenkins/workspace/raid-vg-autotest/autorun-spdk.conf 00:00:41.525 ++ SPDK_RUN_FUNCTIONAL_TEST=1 00:00:41.525 ++ SPDK_RUN_ASAN=1 00:00:41.525 ++ SPDK_RUN_UBSAN=1 00:00:41.525 ++ SPDK_TEST_RAID=1 00:00:41.525 ++ SPDK_TEST_NATIVE_DPDK=v22.11.4 00:00:41.525 ++ SPDK_RUN_EXTERNAL_DPDK=/home/vagrant/spdk_repo/dpdk/build 00:00:41.525 ++ SPDK_ABI_DIR=/home/vagrant/spdk_repo/spdk-abi 00:00:41.525 ++ RUN_NIGHTLY=1 00:00:41.525 + cd /var/jenkins/workspace/raid-vg-autotest 00:00:41.525 + nvme_files=() 00:00:41.525 + declare -A nvme_files 00:00:41.525 + backend_dir=/var/lib/libvirt/images/backends 00:00:41.525 + nvme_files['nvme.img']=5G 00:00:41.525 + nvme_files['nvme-cmb.img']=5G 00:00:41.525 + nvme_files['nvme-multi0.img']=4G 00:00:41.525 + nvme_files['nvme-multi1.img']=4G 00:00:41.525 + nvme_files['nvme-multi2.img']=4G 00:00:41.525 + nvme_files['nvme-openstack.img']=8G 00:00:41.525 + nvme_files['nvme-zns.img']=5G 00:00:41.525 + (( SPDK_TEST_NVME_PMR == 1 )) 00:00:41.525 + (( SPDK_TEST_FTL == 1 )) 00:00:41.525 + (( SPDK_TEST_NVME_FDP == 1 )) 00:00:41.525 + [[ ! -d /var/lib/libvirt/images/backends ]] 00:00:41.525 + for nvme in "${!nvme_files[@]}" 00:00:41.525 + sudo -E spdk/scripts/vagrant/create_nvme_img.sh -n /var/lib/libvirt/images/backends/ex0-nvme-multi2.img -s 4G 00:00:41.525 Formatting '/var/lib/libvirt/images/backends/ex0-nvme-multi2.img', fmt=raw size=4294967296 preallocation=falloc 00:00:41.525 + for nvme in "${!nvme_files[@]}" 00:00:41.525 + sudo -E spdk/scripts/vagrant/create_nvme_img.sh -n /var/lib/libvirt/images/backends/ex0-nvme-cmb.img -s 5G 00:00:41.525 Formatting '/var/lib/libvirt/images/backends/ex0-nvme-cmb.img', fmt=raw size=5368709120 preallocation=falloc 00:00:41.525 + for nvme in "${!nvme_files[@]}" 00:00:41.525 + sudo -E spdk/scripts/vagrant/create_nvme_img.sh -n /var/lib/libvirt/images/backends/ex0-nvme-openstack.img -s 8G 00:00:41.525 Formatting '/var/lib/libvirt/images/backends/ex0-nvme-openstack.img', fmt=raw size=8589934592 preallocation=falloc 00:00:41.525 + for nvme in "${!nvme_files[@]}" 00:00:41.525 + sudo -E spdk/scripts/vagrant/create_nvme_img.sh -n /var/lib/libvirt/images/backends/ex0-nvme-zns.img -s 5G 00:00:41.525 Formatting '/var/lib/libvirt/images/backends/ex0-nvme-zns.img', fmt=raw size=5368709120 preallocation=falloc 00:00:41.525 + for nvme in "${!nvme_files[@]}" 00:00:41.525 + sudo -E spdk/scripts/vagrant/create_nvme_img.sh -n /var/lib/libvirt/images/backends/ex0-nvme-multi1.img -s 4G 00:00:41.525 Formatting '/var/lib/libvirt/images/backends/ex0-nvme-multi1.img', fmt=raw size=4294967296 preallocation=falloc 00:00:41.525 + for nvme in "${!nvme_files[@]}" 00:00:41.525 + sudo -E spdk/scripts/vagrant/create_nvme_img.sh -n /var/lib/libvirt/images/backends/ex0-nvme-multi0.img -s 4G 00:00:41.525 Formatting '/var/lib/libvirt/images/backends/ex0-nvme-multi0.img', fmt=raw size=4294967296 preallocation=falloc 00:00:41.525 + for nvme in "${!nvme_files[@]}" 00:00:41.525 + sudo -E spdk/scripts/vagrant/create_nvme_img.sh -n /var/lib/libvirt/images/backends/ex0-nvme.img -s 5G 00:00:41.802 Formatting '/var/lib/libvirt/images/backends/ex0-nvme.img', fmt=raw size=5368709120 preallocation=falloc 00:00:41.802 ++ sudo grep -rl ex0-nvme.img /etc/libvirt/qemu 00:00:41.802 + echo 'End stage prepare_nvme.sh' 00:00:41.802 End stage prepare_nvme.sh 00:00:41.813 [Pipeline] sh 00:00:42.097 + DISTRO=fedora39 CPUS=10 RAM=12288 jbp/jenkins/jjb-config/jobs/scripts/vagrant_create_vm.sh 00:00:42.098 Setup: -n 10 -s 12288 -x http://proxy-dmz.intel.com:911 -p libvirt --qemu-emulator=/usr/local/qemu/vanilla-v8.0.0/bin/qemu-system-x86_64 -b /var/lib/libvirt/images/backends/ex0-nvme.img -b /var/lib/libvirt/images/backends/ex0-nvme-multi0.img,nvme,/var/lib/libvirt/images/backends/ex0-nvme-multi1.img:/var/lib/libvirt/images/backends/ex0-nvme-multi2.img -H -a -v -f fedora39 00:00:42.098 00:00:42.098 DIR=/var/jenkins/workspace/raid-vg-autotest/spdk/scripts/vagrant 00:00:42.098 SPDK_DIR=/var/jenkins/workspace/raid-vg-autotest/spdk 00:00:42.098 VAGRANT_TARGET=/var/jenkins/workspace/raid-vg-autotest 00:00:42.098 HELP=0 00:00:42.098 DRY_RUN=0 00:00:42.098 NVME_FILE=/var/lib/libvirt/images/backends/ex0-nvme.img,/var/lib/libvirt/images/backends/ex0-nvme-multi0.img, 00:00:42.098 NVME_DISKS_TYPE=nvme,nvme, 00:00:42.098 NVME_AUTO_CREATE=0 00:00:42.098 NVME_DISKS_NAMESPACES=,/var/lib/libvirt/images/backends/ex0-nvme-multi1.img:/var/lib/libvirt/images/backends/ex0-nvme-multi2.img, 00:00:42.098 NVME_CMB=,, 00:00:42.098 NVME_PMR=,, 00:00:42.098 NVME_ZNS=,, 00:00:42.098 NVME_MS=,, 00:00:42.098 NVME_FDP=,, 00:00:42.098 SPDK_VAGRANT_DISTRO=fedora39 00:00:42.098 SPDK_VAGRANT_VMCPU=10 00:00:42.098 SPDK_VAGRANT_VMRAM=12288 00:00:42.098 SPDK_VAGRANT_PROVIDER=libvirt 00:00:42.098 SPDK_VAGRANT_HTTP_PROXY=http://proxy-dmz.intel.com:911 00:00:42.098 SPDK_QEMU_EMULATOR=/usr/local/qemu/vanilla-v8.0.0/bin/qemu-system-x86_64 00:00:42.098 SPDK_OPENSTACK_NETWORK=0 00:00:42.098 VAGRANT_PACKAGE_BOX=0 00:00:42.098 VAGRANTFILE=/var/jenkins/workspace/raid-vg-autotest/spdk/scripts/vagrant/Vagrantfile 00:00:42.098 FORCE_DISTRO=true 00:00:42.098 VAGRANT_BOX_VERSION= 00:00:42.098 EXTRA_VAGRANTFILES= 00:00:42.098 NIC_MODEL=virtio 00:00:42.098 00:00:42.098 mkdir: created directory '/var/jenkins/workspace/raid-vg-autotest/fedora39-libvirt' 00:00:42.098 /var/jenkins/workspace/raid-vg-autotest/fedora39-libvirt /var/jenkins/workspace/raid-vg-autotest 00:00:44.008 Bringing machine 'default' up with 'libvirt' provider... 00:00:44.269 ==> default: Creating image (snapshot of base box volume). 00:00:44.529 ==> default: Creating domain with the following settings... 00:00:44.529 ==> default: -- Name: fedora39-39-1.5-1721788873-2326_default_1732664072_765d9918233ef1ea60c6 00:00:44.529 ==> default: -- Domain type: kvm 00:00:44.529 ==> default: -- Cpus: 10 00:00:44.529 ==> default: -- Feature: acpi 00:00:44.529 ==> default: -- Feature: apic 00:00:44.529 ==> default: -- Feature: pae 00:00:44.529 ==> default: -- Memory: 12288M 00:00:44.529 ==> default: -- Memory Backing: hugepages: 00:00:44.529 ==> default: -- Management MAC: 00:00:44.529 ==> default: -- Loader: 00:00:44.529 ==> default: -- Nvram: 00:00:44.529 ==> default: -- Base box: spdk/fedora39 00:00:44.529 ==> default: -- Storage pool: default 00:00:44.529 ==> default: -- Image: /var/lib/libvirt/images/fedora39-39-1.5-1721788873-2326_default_1732664072_765d9918233ef1ea60c6.img (20G) 00:00:44.529 ==> default: -- Volume Cache: default 00:00:44.529 ==> default: -- Kernel: 00:00:44.529 ==> default: -- Initrd: 00:00:44.529 ==> default: -- Graphics Type: vnc 00:00:44.529 ==> default: -- Graphics Port: -1 00:00:44.529 ==> default: -- Graphics IP: 127.0.0.1 00:00:44.529 ==> default: -- Graphics Password: Not defined 00:00:44.529 ==> default: -- Video Type: cirrus 00:00:44.529 ==> default: -- Video VRAM: 9216 00:00:44.529 ==> default: -- Sound Type: 00:00:44.529 ==> default: -- Keymap: en-us 00:00:44.529 ==> default: -- TPM Path: 00:00:44.529 ==> default: -- INPUT: type=mouse, bus=ps2 00:00:44.529 ==> default: -- Command line args: 00:00:44.529 ==> default: -> value=-device, 00:00:44.529 ==> default: -> value=nvme,id=nvme-0,serial=12340,addr=0x10, 00:00:44.529 ==> default: -> value=-drive, 00:00:44.529 ==> default: -> value=format=raw,file=/var/lib/libvirt/images/backends/ex0-nvme.img,if=none,id=nvme-0-drive0, 00:00:44.529 ==> default: -> value=-device, 00:00:44.529 ==> default: -> value=nvme-ns,drive=nvme-0-drive0,bus=nvme-0,nsid=1,zoned=false,logical_block_size=4096,physical_block_size=4096, 00:00:44.529 ==> default: -> value=-device, 00:00:44.529 ==> default: -> value=nvme,id=nvme-1,serial=12341,addr=0x11, 00:00:44.529 ==> default: -> value=-drive, 00:00:44.529 ==> default: -> value=format=raw,file=/var/lib/libvirt/images/backends/ex0-nvme-multi0.img,if=none,id=nvme-1-drive0, 00:00:44.529 ==> default: -> value=-device, 00:00:44.529 ==> default: -> value=nvme-ns,drive=nvme-1-drive0,bus=nvme-1,nsid=1,zoned=false,logical_block_size=4096,physical_block_size=4096, 00:00:44.529 ==> default: -> value=-drive, 00:00:44.529 ==> default: -> value=format=raw,file=/var/lib/libvirt/images/backends/ex0-nvme-multi1.img,if=none,id=nvme-1-drive1, 00:00:44.529 ==> default: -> value=-device, 00:00:44.529 ==> default: -> value=nvme-ns,drive=nvme-1-drive1,bus=nvme-1,nsid=2,zoned=false,logical_block_size=4096,physical_block_size=4096, 00:00:44.529 ==> default: -> value=-drive, 00:00:44.529 ==> default: -> value=format=raw,file=/var/lib/libvirt/images/backends/ex0-nvme-multi2.img,if=none,id=nvme-1-drive2, 00:00:44.529 ==> default: -> value=-device, 00:00:44.530 ==> default: -> value=nvme-ns,drive=nvme-1-drive2,bus=nvme-1,nsid=3,zoned=false,logical_block_size=4096,physical_block_size=4096, 00:00:44.530 ==> default: Creating shared folders metadata... 00:00:44.530 ==> default: Starting domain. 00:00:46.441 ==> default: Waiting for domain to get an IP address... 00:01:04.564 ==> default: Waiting for SSH to become available... 00:01:04.564 ==> default: Configuring and enabling network interfaces... 00:01:09.847 default: SSH address: 192.168.121.108:22 00:01:09.847 default: SSH username: vagrant 00:01:09.847 default: SSH auth method: private key 00:01:12.398 ==> default: Rsyncing folder: /mnt/jenkins_nvme/jenkins/workspace/raid-vg-autotest/spdk/ => /home/vagrant/spdk_repo/spdk 00:01:20.550 ==> default: Rsyncing folder: /mnt/jenkins_nvme/jenkins/workspace/raid-vg-autotest/dpdk/ => /home/vagrant/spdk_repo/dpdk 00:01:25.835 ==> default: Mounting SSHFS shared folder... 00:01:28.412 ==> default: Mounting folder via SSHFS: /mnt/jenkins_nvme/jenkins/workspace/raid-vg-autotest/fedora39-libvirt/output => /home/vagrant/spdk_repo/output 00:01:28.412 ==> default: Checking Mount.. 00:01:29.799 ==> default: Folder Successfully Mounted! 00:01:29.799 ==> default: Running provisioner: file... 00:01:30.734 default: ~/.gitconfig => .gitconfig 00:01:31.301 00:01:31.301 SUCCESS! 00:01:31.301 00:01:31.301 cd to /var/jenkins/workspace/raid-vg-autotest/fedora39-libvirt and type "vagrant ssh" to use. 00:01:31.301 Use vagrant "suspend" and vagrant "resume" to stop and start. 00:01:31.301 Use vagrant "destroy" followed by "rm -rf /var/jenkins/workspace/raid-vg-autotest/fedora39-libvirt" to destroy all trace of vm. 00:01:31.301 00:01:31.308 [Pipeline] } 00:01:31.321 [Pipeline] // stage 00:01:31.328 [Pipeline] dir 00:01:31.329 Running in /var/jenkins/workspace/raid-vg-autotest/fedora39-libvirt 00:01:31.330 [Pipeline] { 00:01:31.339 [Pipeline] catchError 00:01:31.340 [Pipeline] { 00:01:31.349 [Pipeline] sh 00:01:31.625 + vagrant ssh-config --host vagrant+ 00:01:31.625 sed -ne /^Host/,$p 00:01:31.625 + tee ssh_conf 00:01:34.156 Host vagrant 00:01:34.156 HostName 192.168.121.108 00:01:34.156 User vagrant 00:01:34.156 Port 22 00:01:34.156 UserKnownHostsFile /dev/null 00:01:34.156 StrictHostKeyChecking no 00:01:34.156 PasswordAuthentication no 00:01:34.156 IdentityFile /var/lib/libvirt/images/.vagrant.d/boxes/spdk-VAGRANTSLASH-fedora39/39-1.5-1721788873-2326/libvirt/fedora39 00:01:34.156 IdentitiesOnly yes 00:01:34.156 LogLevel FATAL 00:01:34.156 ForwardAgent yes 00:01:34.156 ForwardX11 yes 00:01:34.156 00:01:34.170 [Pipeline] withEnv 00:01:34.172 [Pipeline] { 00:01:34.185 [Pipeline] sh 00:01:34.464 + /usr/local/bin/ssh -t -F ssh_conf vagrant@vagrant #!/bin/bash 00:01:34.465 source /etc/os-release 00:01:34.465 [[ -e /image.version ]] && img=$(< /image.version) 00:01:34.465 # Minimal, systemd-like check. 00:01:34.465 if [[ -e /.dockerenv ]]; then 00:01:34.465 # Clear garbage from the node's name: 00:01:34.465 # agt-er_autotest_547-896 -> autotest_547-896 00:01:34.465 # $HOSTNAME is the actual container id 00:01:34.465 agent=$HOSTNAME@${DOCKER_SWARM_PLUGIN_JENKINS_AGENT_NAME#*_} 00:01:34.465 if grep -q "/etc/hostname" /proc/self/mountinfo; then 00:01:34.465 # We can assume this is a mount from a host where container is running, 00:01:34.465 # so fetch its hostname to easily identify the target swarm worker. 00:01:34.465 container="$(< /etc/hostname) ($agent)" 00:01:34.465 else 00:01:34.465 # Fallback 00:01:34.465 container=$agent 00:01:34.465 fi 00:01:34.465 fi 00:01:34.465 echo "${NAME} ${VERSION_ID}|$(uname -r)|${img:-N/A}|${container:-N/A}" 00:01:34.465 00:01:34.735 [Pipeline] } 00:01:34.751 [Pipeline] // withEnv 00:01:34.759 [Pipeline] setCustomBuildProperty 00:01:34.774 [Pipeline] stage 00:01:34.776 [Pipeline] { (Tests) 00:01:34.794 [Pipeline] sh 00:01:35.077 + scp -F ssh_conf -r /var/jenkins/workspace/raid-vg-autotest/jbp/jenkins/jjb-config/jobs/scripts/autoruner.sh vagrant@vagrant:./ 00:01:35.351 [Pipeline] sh 00:01:35.636 + scp -F ssh_conf -r /var/jenkins/workspace/raid-vg-autotest/jbp/jenkins/jjb-config/jobs/scripts/pkgdep-autoruner.sh vagrant@vagrant:./ 00:01:35.915 [Pipeline] timeout 00:01:35.915 Timeout set to expire in 1 hr 30 min 00:01:35.917 [Pipeline] { 00:01:35.931 [Pipeline] sh 00:01:36.215 + /usr/local/bin/ssh -t -F ssh_conf vagrant@vagrant git -C spdk_repo/spdk reset --hard 00:01:36.787 HEAD is now at 2f2acf4eb doc: move nvmf_tracing.md to tracing.md 00:01:36.802 [Pipeline] sh 00:01:37.085 + /usr/local/bin/ssh -t -F ssh_conf vagrant@vagrant sudo chown vagrant:vagrant spdk_repo 00:01:37.359 [Pipeline] sh 00:01:37.643 + scp -F ssh_conf -r /var/jenkins/workspace/raid-vg-autotest/autorun-spdk.conf vagrant@vagrant:spdk_repo 00:01:37.932 [Pipeline] sh 00:01:38.211 + /usr/local/bin/ssh -t -F ssh_conf vagrant@vagrant JOB_BASE_NAME=raid-vg-autotest ./autoruner.sh spdk_repo 00:01:38.471 ++ readlink -f spdk_repo 00:01:38.472 + DIR_ROOT=/home/vagrant/spdk_repo 00:01:38.472 + [[ -n /home/vagrant/spdk_repo ]] 00:01:38.472 + DIR_SPDK=/home/vagrant/spdk_repo/spdk 00:01:38.472 + DIR_OUTPUT=/home/vagrant/spdk_repo/output 00:01:38.472 + [[ -d /home/vagrant/spdk_repo/spdk ]] 00:01:38.472 + [[ ! -d /home/vagrant/spdk_repo/output ]] 00:01:38.472 + [[ -d /home/vagrant/spdk_repo/output ]] 00:01:38.472 + [[ raid-vg-autotest == pkgdep-* ]] 00:01:38.472 + cd /home/vagrant/spdk_repo 00:01:38.472 + source /etc/os-release 00:01:38.472 ++ NAME='Fedora Linux' 00:01:38.472 ++ VERSION='39 (Cloud Edition)' 00:01:38.472 ++ ID=fedora 00:01:38.472 ++ VERSION_ID=39 00:01:38.472 ++ VERSION_CODENAME= 00:01:38.472 ++ PLATFORM_ID=platform:f39 00:01:38.472 ++ PRETTY_NAME='Fedora Linux 39 (Cloud Edition)' 00:01:38.472 ++ ANSI_COLOR='0;38;2;60;110;180' 00:01:38.472 ++ LOGO=fedora-logo-icon 00:01:38.472 ++ CPE_NAME=cpe:/o:fedoraproject:fedora:39 00:01:38.472 ++ HOME_URL=https://fedoraproject.org/ 00:01:38.472 ++ DOCUMENTATION_URL=https://docs.fedoraproject.org/en-US/fedora/f39/system-administrators-guide/ 00:01:38.472 ++ SUPPORT_URL=https://ask.fedoraproject.org/ 00:01:38.472 ++ BUG_REPORT_URL=https://bugzilla.redhat.com/ 00:01:38.472 ++ REDHAT_BUGZILLA_PRODUCT=Fedora 00:01:38.472 ++ REDHAT_BUGZILLA_PRODUCT_VERSION=39 00:01:38.472 ++ REDHAT_SUPPORT_PRODUCT=Fedora 00:01:38.472 ++ REDHAT_SUPPORT_PRODUCT_VERSION=39 00:01:38.472 ++ SUPPORT_END=2024-11-12 00:01:38.472 ++ VARIANT='Cloud Edition' 00:01:38.472 ++ VARIANT_ID=cloud 00:01:38.472 + uname -a 00:01:38.472 Linux fedora39-cloud-1721788873-2326 6.8.9-200.fc39.x86_64 #1 SMP PREEMPT_DYNAMIC Wed Jul 24 03:04:40 UTC 2024 x86_64 GNU/Linux 00:01:38.472 + sudo /home/vagrant/spdk_repo/spdk/scripts/setup.sh status 00:01:39.040 0000:00:03.0 (1af4 1001): Active devices: mount@vda:vda2,mount@vda:vda3,mount@vda:vda5, so not binding PCI dev 00:01:39.040 Hugepages 00:01:39.040 node hugesize free / total 00:01:39.040 node0 1048576kB 0 / 0 00:01:39.040 node0 2048kB 0 / 0 00:01:39.040 00:01:39.040 Type BDF Vendor Device NUMA Driver Device Block devices 00:01:39.040 virtio 0000:00:03.0 1af4 1001 unknown virtio-pci - vda 00:01:39.040 NVMe 0000:00:10.0 1b36 0010 unknown nvme nvme0 nvme0n1 00:01:39.040 NVMe 0000:00:11.0 1b36 0010 unknown nvme nvme1 nvme1n1 nvme1n2 nvme1n3 00:01:39.040 + rm -f /tmp/spdk-ld-path 00:01:39.040 + source autorun-spdk.conf 00:01:39.040 ++ SPDK_RUN_FUNCTIONAL_TEST=1 00:01:39.040 ++ SPDK_RUN_ASAN=1 00:01:39.040 ++ SPDK_RUN_UBSAN=1 00:01:39.040 ++ SPDK_TEST_RAID=1 00:01:39.040 ++ SPDK_TEST_NATIVE_DPDK=v22.11.4 00:01:39.040 ++ SPDK_RUN_EXTERNAL_DPDK=/home/vagrant/spdk_repo/dpdk/build 00:01:39.040 ++ SPDK_ABI_DIR=/home/vagrant/spdk_repo/spdk-abi 00:01:39.040 ++ RUN_NIGHTLY=1 00:01:39.040 + (( SPDK_TEST_NVME_CMB == 1 || SPDK_TEST_NVME_PMR == 1 )) 00:01:39.040 + [[ -n '' ]] 00:01:39.040 + sudo git config --global --add safe.directory /home/vagrant/spdk_repo/spdk 00:01:39.040 + for M in /var/spdk/build-*-manifest.txt 00:01:39.040 + [[ -f /var/spdk/build-kernel-manifest.txt ]] 00:01:39.040 + cp /var/spdk/build-kernel-manifest.txt /home/vagrant/spdk_repo/output/ 00:01:39.040 + for M in /var/spdk/build-*-manifest.txt 00:01:39.040 + [[ -f /var/spdk/build-pkg-manifest.txt ]] 00:01:39.040 + cp /var/spdk/build-pkg-manifest.txt /home/vagrant/spdk_repo/output/ 00:01:39.040 + for M in /var/spdk/build-*-manifest.txt 00:01:39.040 + [[ -f /var/spdk/build-repo-manifest.txt ]] 00:01:39.040 + cp /var/spdk/build-repo-manifest.txt /home/vagrant/spdk_repo/output/ 00:01:39.040 ++ uname 00:01:39.040 + [[ Linux == \L\i\n\u\x ]] 00:01:39.040 + sudo dmesg -T 00:01:39.300 + sudo dmesg --clear 00:01:39.300 + dmesg_pid=6157 00:01:39.300 + sudo dmesg -Tw 00:01:39.300 + [[ Fedora Linux == FreeBSD ]] 00:01:39.300 + export UNBIND_ENTIRE_IOMMU_GROUP=yes 00:01:39.300 + UNBIND_ENTIRE_IOMMU_GROUP=yes 00:01:39.300 + [[ -e /var/spdk/dependencies/vhost/spdk_test_image.qcow2 ]] 00:01:39.300 + [[ -x /usr/src/fio-static/fio ]] 00:01:39.300 + export FIO_BIN=/usr/src/fio-static/fio 00:01:39.300 + FIO_BIN=/usr/src/fio-static/fio 00:01:39.300 + [[ '' == \/\q\e\m\u\_\v\f\i\o\/* ]] 00:01:39.300 + [[ ! -v VFIO_QEMU_BIN ]] 00:01:39.300 + [[ -e /usr/local/qemu/vfio-user-latest ]] 00:01:39.300 + export VFIO_QEMU_BIN=/usr/local/qemu/vfio-user-latest/bin/qemu-system-x86_64 00:01:39.300 + VFIO_QEMU_BIN=/usr/local/qemu/vfio-user-latest/bin/qemu-system-x86_64 00:01:39.300 + [[ -e /usr/local/qemu/vanilla-latest ]] 00:01:39.300 + export QEMU_BIN=/usr/local/qemu/vanilla-latest/bin/qemu-system-x86_64 00:01:39.300 + QEMU_BIN=/usr/local/qemu/vanilla-latest/bin/qemu-system-x86_64 00:01:39.300 + spdk/autorun.sh /home/vagrant/spdk_repo/autorun-spdk.conf 00:01:39.300 23:35:27 -- common/autotest_common.sh@1692 -- $ [[ n == y ]] 00:01:39.300 23:35:27 -- spdk/autorun.sh@20 -- $ source /home/vagrant/spdk_repo/autorun-spdk.conf 00:01:39.300 23:35:27 -- spdk_repo/autorun-spdk.conf@1 -- $ SPDK_RUN_FUNCTIONAL_TEST=1 00:01:39.300 23:35:27 -- spdk_repo/autorun-spdk.conf@2 -- $ SPDK_RUN_ASAN=1 00:01:39.300 23:35:27 -- spdk_repo/autorun-spdk.conf@3 -- $ SPDK_RUN_UBSAN=1 00:01:39.300 23:35:27 -- spdk_repo/autorun-spdk.conf@4 -- $ SPDK_TEST_RAID=1 00:01:39.300 23:35:27 -- spdk_repo/autorun-spdk.conf@5 -- $ SPDK_TEST_NATIVE_DPDK=v22.11.4 00:01:39.300 23:35:27 -- spdk_repo/autorun-spdk.conf@6 -- $ SPDK_RUN_EXTERNAL_DPDK=/home/vagrant/spdk_repo/dpdk/build 00:01:39.300 23:35:27 -- spdk_repo/autorun-spdk.conf@7 -- $ SPDK_ABI_DIR=/home/vagrant/spdk_repo/spdk-abi 00:01:39.300 23:35:27 -- spdk_repo/autorun-spdk.conf@8 -- $ RUN_NIGHTLY=1 00:01:39.300 23:35:27 -- spdk/autorun.sh@22 -- $ trap 'timing_finish || exit 1' EXIT 00:01:39.300 23:35:27 -- spdk/autorun.sh@25 -- $ /home/vagrant/spdk_repo/spdk/autobuild.sh /home/vagrant/spdk_repo/autorun-spdk.conf 00:01:39.300 23:35:27 -- common/autotest_common.sh@1692 -- $ [[ n == y ]] 00:01:39.300 23:35:27 -- common/autobuild_common.sh@15 -- $ source /home/vagrant/spdk_repo/spdk/scripts/common.sh 00:01:39.300 23:35:27 -- scripts/common.sh@15 -- $ shopt -s extglob 00:01:39.300 23:35:27 -- scripts/common.sh@544 -- $ [[ -e /bin/wpdk_common.sh ]] 00:01:39.300 23:35:27 -- scripts/common.sh@552 -- $ [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:01:39.300 23:35:27 -- scripts/common.sh@553 -- $ source /etc/opt/spdk-pkgdep/paths/export.sh 00:01:39.300 23:35:27 -- paths/export.sh@2 -- $ PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/vagrant/.local/bin:/home/vagrant/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:01:39.300 23:35:27 -- paths/export.sh@3 -- $ PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/vagrant/.local/bin:/home/vagrant/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:01:39.300 23:35:27 -- paths/export.sh@4 -- $ PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/vagrant/.local/bin:/home/vagrant/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:01:39.300 23:35:27 -- paths/export.sh@5 -- $ export PATH 00:01:39.300 23:35:27 -- paths/export.sh@6 -- $ echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/vagrant/.local/bin:/home/vagrant/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:01:39.300 23:35:27 -- common/autobuild_common.sh@492 -- $ out=/home/vagrant/spdk_repo/spdk/../output 00:01:39.560 23:35:27 -- common/autobuild_common.sh@493 -- $ date +%s 00:01:39.560 23:35:27 -- common/autobuild_common.sh@493 -- $ mktemp -dt spdk_1732664127.XXXXXX 00:01:39.560 23:35:27 -- common/autobuild_common.sh@493 -- $ SPDK_WORKSPACE=/tmp/spdk_1732664127.PFxg5R 00:01:39.560 23:35:27 -- common/autobuild_common.sh@495 -- $ [[ -n '' ]] 00:01:39.560 23:35:27 -- common/autobuild_common.sh@499 -- $ '[' -n v22.11.4 ']' 00:01:39.560 23:35:27 -- common/autobuild_common.sh@500 -- $ dirname /home/vagrant/spdk_repo/dpdk/build 00:01:39.560 23:35:27 -- common/autobuild_common.sh@500 -- $ scanbuild_exclude=' --exclude /home/vagrant/spdk_repo/dpdk' 00:01:39.560 23:35:27 -- common/autobuild_common.sh@506 -- $ scanbuild_exclude+=' --exclude /home/vagrant/spdk_repo/spdk/xnvme --exclude /tmp' 00:01:39.560 23:35:27 -- common/autobuild_common.sh@508 -- $ scanbuild='scan-build -o /home/vagrant/spdk_repo/spdk/../output/scan-build-tmp --exclude /home/vagrant/spdk_repo/dpdk --exclude /home/vagrant/spdk_repo/spdk/xnvme --exclude /tmp --status-bugs' 00:01:39.560 23:35:27 -- common/autobuild_common.sh@509 -- $ get_config_params 00:01:39.560 23:35:27 -- common/autotest_common.sh@409 -- $ xtrace_disable 00:01:39.560 23:35:27 -- common/autotest_common.sh@10 -- $ set +x 00:01:39.560 23:35:27 -- common/autobuild_common.sh@509 -- $ config_params='--enable-debug --enable-werror --with-rdma --with-idxd --with-fio=/usr/src/fio --with-iscsi-initiator --disable-unit-tests --enable-ubsan --enable-asan --enable-coverage --with-ublk --with-raid5f --with-dpdk=/home/vagrant/spdk_repo/dpdk/build' 00:01:39.560 23:35:27 -- common/autobuild_common.sh@511 -- $ start_monitor_resources 00:01:39.560 23:35:27 -- pm/common@17 -- $ local monitor 00:01:39.560 23:35:27 -- pm/common@19 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:01:39.560 23:35:27 -- pm/common@19 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:01:39.560 23:35:27 -- pm/common@25 -- $ sleep 1 00:01:39.560 23:35:27 -- pm/common@21 -- $ date +%s 00:01:39.560 23:35:27 -- pm/common@21 -- $ date +%s 00:01:39.560 23:35:27 -- pm/common@21 -- $ /home/vagrant/spdk_repo/spdk/scripts/perf/pm/collect-vmstat -d /home/vagrant/spdk_repo/spdk/../output/power -l -p monitor.autobuild.sh.1732664127 00:01:39.560 23:35:27 -- pm/common@21 -- $ /home/vagrant/spdk_repo/spdk/scripts/perf/pm/collect-cpu-load -d /home/vagrant/spdk_repo/spdk/../output/power -l -p monitor.autobuild.sh.1732664127 00:01:39.560 Redirecting to /home/vagrant/spdk_repo/spdk/../output/power/monitor.autobuild.sh.1732664127_collect-cpu-load.pm.log 00:01:39.560 Redirecting to /home/vagrant/spdk_repo/spdk/../output/power/monitor.autobuild.sh.1732664127_collect-vmstat.pm.log 00:01:40.498 23:35:28 -- common/autobuild_common.sh@512 -- $ trap stop_monitor_resources EXIT 00:01:40.498 23:35:28 -- spdk/autobuild.sh@11 -- $ SPDK_TEST_AUTOBUILD= 00:01:40.498 23:35:28 -- spdk/autobuild.sh@12 -- $ umask 022 00:01:40.498 23:35:28 -- spdk/autobuild.sh@13 -- $ cd /home/vagrant/spdk_repo/spdk 00:01:40.498 23:35:28 -- spdk/autobuild.sh@16 -- $ date -u 00:01:40.498 Tue Nov 26 11:35:28 PM UTC 2024 00:01:40.498 23:35:28 -- spdk/autobuild.sh@17 -- $ git describe --tags 00:01:40.498 v25.01-pre-271-g2f2acf4eb 00:01:40.498 23:35:28 -- spdk/autobuild.sh@19 -- $ '[' 1 -eq 1 ']' 00:01:40.498 23:35:28 -- spdk/autobuild.sh@20 -- $ run_test asan echo 'using asan' 00:01:40.498 23:35:28 -- common/autotest_common.sh@1105 -- $ '[' 3 -le 1 ']' 00:01:40.498 23:35:28 -- common/autotest_common.sh@1111 -- $ xtrace_disable 00:01:40.498 23:35:28 -- common/autotest_common.sh@10 -- $ set +x 00:01:40.498 ************************************ 00:01:40.498 START TEST asan 00:01:40.498 ************************************ 00:01:40.498 using asan 00:01:40.498 23:35:28 asan -- common/autotest_common.sh@1129 -- $ echo 'using asan' 00:01:40.498 00:01:40.498 real 0m0.000s 00:01:40.498 user 0m0.000s 00:01:40.498 sys 0m0.000s 00:01:40.498 23:35:28 asan -- common/autotest_common.sh@1130 -- $ xtrace_disable 00:01:40.498 23:35:28 asan -- common/autotest_common.sh@10 -- $ set +x 00:01:40.498 ************************************ 00:01:40.498 END TEST asan 00:01:40.498 ************************************ 00:01:40.498 23:35:28 -- spdk/autobuild.sh@23 -- $ '[' 1 -eq 1 ']' 00:01:40.498 23:35:28 -- spdk/autobuild.sh@24 -- $ run_test ubsan echo 'using ubsan' 00:01:40.498 23:35:28 -- common/autotest_common.sh@1105 -- $ '[' 3 -le 1 ']' 00:01:40.498 23:35:28 -- common/autotest_common.sh@1111 -- $ xtrace_disable 00:01:40.498 23:35:28 -- common/autotest_common.sh@10 -- $ set +x 00:01:40.498 ************************************ 00:01:40.498 START TEST ubsan 00:01:40.498 ************************************ 00:01:40.498 using ubsan 00:01:40.498 23:35:28 ubsan -- common/autotest_common.sh@1129 -- $ echo 'using ubsan' 00:01:40.498 00:01:40.498 real 0m0.001s 00:01:40.498 user 0m0.000s 00:01:40.498 sys 0m0.000s 00:01:40.498 23:35:28 ubsan -- common/autotest_common.sh@1130 -- $ xtrace_disable 00:01:40.498 23:35:28 ubsan -- common/autotest_common.sh@10 -- $ set +x 00:01:40.498 ************************************ 00:01:40.498 END TEST ubsan 00:01:40.498 ************************************ 00:01:40.498 23:35:28 -- spdk/autobuild.sh@27 -- $ '[' -n v22.11.4 ']' 00:01:40.498 23:35:28 -- spdk/autobuild.sh@28 -- $ build_native_dpdk 00:01:40.498 23:35:28 -- common/autobuild_common.sh@449 -- $ run_test build_native_dpdk _build_native_dpdk 00:01:40.498 23:35:28 -- common/autotest_common.sh@1105 -- $ '[' 2 -le 1 ']' 00:01:40.498 23:35:28 -- common/autotest_common.sh@1111 -- $ xtrace_disable 00:01:40.498 23:35:28 -- common/autotest_common.sh@10 -- $ set +x 00:01:40.759 ************************************ 00:01:40.759 START TEST build_native_dpdk 00:01:40.759 ************************************ 00:01:40.759 23:35:28 build_native_dpdk -- common/autotest_common.sh@1129 -- $ _build_native_dpdk 00:01:40.759 23:35:28 build_native_dpdk -- common/autobuild_common.sh@48 -- $ local external_dpdk_dir 00:01:40.759 23:35:28 build_native_dpdk -- common/autobuild_common.sh@49 -- $ local external_dpdk_base_dir 00:01:40.759 23:35:28 build_native_dpdk -- common/autobuild_common.sh@50 -- $ local compiler_version 00:01:40.759 23:35:28 build_native_dpdk -- common/autobuild_common.sh@51 -- $ local compiler 00:01:40.759 23:35:28 build_native_dpdk -- common/autobuild_common.sh@52 -- $ local dpdk_kmods 00:01:40.759 23:35:28 build_native_dpdk -- common/autobuild_common.sh@53 -- $ local repo=dpdk 00:01:40.759 23:35:28 build_native_dpdk -- common/autobuild_common.sh@55 -- $ compiler=gcc 00:01:40.759 23:35:28 build_native_dpdk -- common/autobuild_common.sh@61 -- $ export CC=gcc 00:01:40.759 23:35:28 build_native_dpdk -- common/autobuild_common.sh@61 -- $ CC=gcc 00:01:40.759 23:35:28 build_native_dpdk -- common/autobuild_common.sh@63 -- $ [[ gcc != *clang* ]] 00:01:40.759 23:35:28 build_native_dpdk -- common/autobuild_common.sh@63 -- $ [[ gcc != *gcc* ]] 00:01:40.759 23:35:28 build_native_dpdk -- common/autobuild_common.sh@68 -- $ gcc -dumpversion 00:01:40.759 23:35:28 build_native_dpdk -- common/autobuild_common.sh@68 -- $ compiler_version=13 00:01:40.759 23:35:28 build_native_dpdk -- common/autobuild_common.sh@69 -- $ compiler_version=13 00:01:40.759 23:35:28 build_native_dpdk -- common/autobuild_common.sh@70 -- $ external_dpdk_dir=/home/vagrant/spdk_repo/dpdk/build 00:01:40.759 23:35:28 build_native_dpdk -- common/autobuild_common.sh@71 -- $ dirname /home/vagrant/spdk_repo/dpdk/build 00:01:40.759 23:35:28 build_native_dpdk -- common/autobuild_common.sh@71 -- $ external_dpdk_base_dir=/home/vagrant/spdk_repo/dpdk 00:01:40.759 23:35:28 build_native_dpdk -- common/autobuild_common.sh@73 -- $ [[ ! -d /home/vagrant/spdk_repo/dpdk ]] 00:01:40.759 23:35:28 build_native_dpdk -- common/autobuild_common.sh@82 -- $ orgdir=/home/vagrant/spdk_repo/spdk 00:01:40.759 23:35:28 build_native_dpdk -- common/autobuild_common.sh@83 -- $ git -C /home/vagrant/spdk_repo/dpdk log --oneline -n 5 00:01:40.759 caf0f5d395 version: 22.11.4 00:01:40.759 7d6f1cc05f Revert "net/iavf: fix abnormal disable HW interrupt" 00:01:40.759 dc9c799c7d vhost: fix missing spinlock unlock 00:01:40.759 4307659a90 net/mlx5: fix LACP redirection in Rx domain 00:01:40.759 6ef77f2a5e net/gve: fix RX buffer size alignment 00:01:40.759 23:35:28 build_native_dpdk -- common/autobuild_common.sh@85 -- $ dpdk_cflags='-fPIC -g -fcommon' 00:01:40.759 23:35:28 build_native_dpdk -- common/autobuild_common.sh@86 -- $ dpdk_ldflags= 00:01:40.759 23:35:28 build_native_dpdk -- common/autobuild_common.sh@87 -- $ dpdk_ver=22.11.4 00:01:40.759 23:35:28 build_native_dpdk -- common/autobuild_common.sh@89 -- $ [[ gcc == *gcc* ]] 00:01:40.759 23:35:28 build_native_dpdk -- common/autobuild_common.sh@89 -- $ [[ 13 -ge 5 ]] 00:01:40.759 23:35:28 build_native_dpdk -- common/autobuild_common.sh@90 -- $ dpdk_cflags+=' -Werror' 00:01:40.759 23:35:28 build_native_dpdk -- common/autobuild_common.sh@93 -- $ [[ gcc == *gcc* ]] 00:01:40.759 23:35:28 build_native_dpdk -- common/autobuild_common.sh@93 -- $ [[ 13 -ge 10 ]] 00:01:40.759 23:35:28 build_native_dpdk -- common/autobuild_common.sh@94 -- $ dpdk_cflags+=' -Wno-stringop-overflow' 00:01:40.759 23:35:28 build_native_dpdk -- common/autobuild_common.sh@102 -- $ DPDK_DRIVERS=("bus" "bus/pci" "bus/vdev" "mempool/ring" "net/i40e" "net/i40e/base" "power/acpi" "power/amd_pstate" "power/cppc" "power/intel_pstate" "power/intel_uncore" "power/kvm_vm") 00:01:40.759 23:35:28 build_native_dpdk -- common/autobuild_common.sh@103 -- $ local mlx5_libs_added=n 00:01:40.759 23:35:28 build_native_dpdk -- common/autobuild_common.sh@104 -- $ [[ 0 -eq 1 ]] 00:01:40.759 23:35:28 build_native_dpdk -- common/autobuild_common.sh@104 -- $ [[ 0 -eq 1 ]] 00:01:40.759 23:35:28 build_native_dpdk -- common/autobuild_common.sh@146 -- $ [[ 0 -eq 1 ]] 00:01:40.759 23:35:28 build_native_dpdk -- common/autobuild_common.sh@174 -- $ cd /home/vagrant/spdk_repo/dpdk 00:01:40.759 23:35:28 build_native_dpdk -- common/autobuild_common.sh@175 -- $ uname -s 00:01:40.759 23:35:28 build_native_dpdk -- common/autobuild_common.sh@175 -- $ '[' Linux = Linux ']' 00:01:40.759 23:35:28 build_native_dpdk -- common/autobuild_common.sh@176 -- $ lt 22.11.4 21.11.0 00:01:40.759 23:35:28 build_native_dpdk -- scripts/common.sh@373 -- $ cmp_versions 22.11.4 '<' 21.11.0 00:01:40.759 23:35:28 build_native_dpdk -- scripts/common.sh@333 -- $ local ver1 ver1_l 00:01:40.759 23:35:28 build_native_dpdk -- scripts/common.sh@334 -- $ local ver2 ver2_l 00:01:40.759 23:35:28 build_native_dpdk -- scripts/common.sh@336 -- $ IFS=.-: 00:01:40.759 23:35:28 build_native_dpdk -- scripts/common.sh@336 -- $ read -ra ver1 00:01:40.759 23:35:28 build_native_dpdk -- scripts/common.sh@337 -- $ IFS=.-: 00:01:40.759 23:35:28 build_native_dpdk -- scripts/common.sh@337 -- $ read -ra ver2 00:01:40.759 23:35:28 build_native_dpdk -- scripts/common.sh@338 -- $ local 'op=<' 00:01:40.759 23:35:28 build_native_dpdk -- scripts/common.sh@340 -- $ ver1_l=3 00:01:40.759 23:35:28 build_native_dpdk -- scripts/common.sh@341 -- $ ver2_l=3 00:01:40.759 23:35:28 build_native_dpdk -- scripts/common.sh@343 -- $ local lt=0 gt=0 eq=0 v 00:01:40.759 23:35:28 build_native_dpdk -- scripts/common.sh@344 -- $ case "$op" in 00:01:40.759 23:35:28 build_native_dpdk -- scripts/common.sh@345 -- $ : 1 00:01:40.759 23:35:28 build_native_dpdk -- scripts/common.sh@364 -- $ (( v = 0 )) 00:01:40.759 23:35:28 build_native_dpdk -- scripts/common.sh@364 -- $ (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:01:40.759 23:35:28 build_native_dpdk -- scripts/common.sh@365 -- $ decimal 22 00:01:40.759 23:35:28 build_native_dpdk -- scripts/common.sh@353 -- $ local d=22 00:01:40.759 23:35:28 build_native_dpdk -- scripts/common.sh@354 -- $ [[ 22 =~ ^[0-9]+$ ]] 00:01:40.759 23:35:28 build_native_dpdk -- scripts/common.sh@355 -- $ echo 22 00:01:40.759 23:35:28 build_native_dpdk -- scripts/common.sh@365 -- $ ver1[v]=22 00:01:40.759 23:35:28 build_native_dpdk -- scripts/common.sh@366 -- $ decimal 21 00:01:40.759 23:35:28 build_native_dpdk -- scripts/common.sh@353 -- $ local d=21 00:01:40.759 23:35:28 build_native_dpdk -- scripts/common.sh@354 -- $ [[ 21 =~ ^[0-9]+$ ]] 00:01:40.759 23:35:28 build_native_dpdk -- scripts/common.sh@355 -- $ echo 21 00:01:40.759 23:35:28 build_native_dpdk -- scripts/common.sh@366 -- $ ver2[v]=21 00:01:40.759 23:35:28 build_native_dpdk -- scripts/common.sh@367 -- $ (( ver1[v] > ver2[v] )) 00:01:40.759 23:35:28 build_native_dpdk -- scripts/common.sh@367 -- $ return 1 00:01:40.759 23:35:28 build_native_dpdk -- common/autobuild_common.sh@180 -- $ patch -p1 00:01:40.759 patching file config/rte_config.h 00:01:40.759 Hunk #1 succeeded at 60 (offset 1 line). 00:01:40.759 23:35:28 build_native_dpdk -- common/autobuild_common.sh@183 -- $ lt 22.11.4 24.07.0 00:01:40.759 23:35:28 build_native_dpdk -- scripts/common.sh@373 -- $ cmp_versions 22.11.4 '<' 24.07.0 00:01:40.759 23:35:28 build_native_dpdk -- scripts/common.sh@333 -- $ local ver1 ver1_l 00:01:40.759 23:35:28 build_native_dpdk -- scripts/common.sh@334 -- $ local ver2 ver2_l 00:01:40.759 23:35:28 build_native_dpdk -- scripts/common.sh@336 -- $ IFS=.-: 00:01:40.759 23:35:28 build_native_dpdk -- scripts/common.sh@336 -- $ read -ra ver1 00:01:40.759 23:35:28 build_native_dpdk -- scripts/common.sh@337 -- $ IFS=.-: 00:01:40.759 23:35:28 build_native_dpdk -- scripts/common.sh@337 -- $ read -ra ver2 00:01:40.759 23:35:28 build_native_dpdk -- scripts/common.sh@338 -- $ local 'op=<' 00:01:40.759 23:35:28 build_native_dpdk -- scripts/common.sh@340 -- $ ver1_l=3 00:01:40.759 23:35:28 build_native_dpdk -- scripts/common.sh@341 -- $ ver2_l=3 00:01:40.759 23:35:28 build_native_dpdk -- scripts/common.sh@343 -- $ local lt=0 gt=0 eq=0 v 00:01:40.759 23:35:28 build_native_dpdk -- scripts/common.sh@344 -- $ case "$op" in 00:01:40.759 23:35:28 build_native_dpdk -- scripts/common.sh@345 -- $ : 1 00:01:40.759 23:35:28 build_native_dpdk -- scripts/common.sh@364 -- $ (( v = 0 )) 00:01:40.759 23:35:28 build_native_dpdk -- scripts/common.sh@364 -- $ (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:01:40.759 23:35:28 build_native_dpdk -- scripts/common.sh@365 -- $ decimal 22 00:01:40.759 23:35:28 build_native_dpdk -- scripts/common.sh@353 -- $ local d=22 00:01:40.759 23:35:28 build_native_dpdk -- scripts/common.sh@354 -- $ [[ 22 =~ ^[0-9]+$ ]] 00:01:40.759 23:35:28 build_native_dpdk -- scripts/common.sh@355 -- $ echo 22 00:01:40.759 23:35:28 build_native_dpdk -- scripts/common.sh@365 -- $ ver1[v]=22 00:01:40.759 23:35:28 build_native_dpdk -- scripts/common.sh@366 -- $ decimal 24 00:01:40.759 23:35:28 build_native_dpdk -- scripts/common.sh@353 -- $ local d=24 00:01:40.759 23:35:28 build_native_dpdk -- scripts/common.sh@354 -- $ [[ 24 =~ ^[0-9]+$ ]] 00:01:40.759 23:35:28 build_native_dpdk -- scripts/common.sh@355 -- $ echo 24 00:01:40.759 23:35:28 build_native_dpdk -- scripts/common.sh@366 -- $ ver2[v]=24 00:01:40.759 23:35:28 build_native_dpdk -- scripts/common.sh@367 -- $ (( ver1[v] > ver2[v] )) 00:01:40.759 23:35:28 build_native_dpdk -- scripts/common.sh@368 -- $ (( ver1[v] < ver2[v] )) 00:01:40.759 23:35:28 build_native_dpdk -- scripts/common.sh@368 -- $ return 0 00:01:40.759 23:35:28 build_native_dpdk -- common/autobuild_common.sh@184 -- $ patch -p1 00:01:40.759 patching file lib/pcapng/rte_pcapng.c 00:01:40.759 Hunk #1 succeeded at 110 (offset -18 lines). 00:01:40.759 23:35:28 build_native_dpdk -- common/autobuild_common.sh@186 -- $ ge 22.11.4 24.07.0 00:01:40.759 23:35:28 build_native_dpdk -- scripts/common.sh@376 -- $ cmp_versions 22.11.4 '>=' 24.07.0 00:01:40.759 23:35:28 build_native_dpdk -- scripts/common.sh@333 -- $ local ver1 ver1_l 00:01:40.759 23:35:28 build_native_dpdk -- scripts/common.sh@334 -- $ local ver2 ver2_l 00:01:40.759 23:35:28 build_native_dpdk -- scripts/common.sh@336 -- $ IFS=.-: 00:01:40.759 23:35:28 build_native_dpdk -- scripts/common.sh@336 -- $ read -ra ver1 00:01:40.759 23:35:28 build_native_dpdk -- scripts/common.sh@337 -- $ IFS=.-: 00:01:40.759 23:35:28 build_native_dpdk -- scripts/common.sh@337 -- $ read -ra ver2 00:01:40.759 23:35:28 build_native_dpdk -- scripts/common.sh@338 -- $ local 'op=>=' 00:01:40.759 23:35:28 build_native_dpdk -- scripts/common.sh@340 -- $ ver1_l=3 00:01:40.759 23:35:28 build_native_dpdk -- scripts/common.sh@341 -- $ ver2_l=3 00:01:40.759 23:35:28 build_native_dpdk -- scripts/common.sh@343 -- $ local lt=0 gt=0 eq=0 v 00:01:40.759 23:35:28 build_native_dpdk -- scripts/common.sh@344 -- $ case "$op" in 00:01:40.759 23:35:28 build_native_dpdk -- scripts/common.sh@348 -- $ : 1 00:01:40.759 23:35:28 build_native_dpdk -- scripts/common.sh@364 -- $ (( v = 0 )) 00:01:40.760 23:35:28 build_native_dpdk -- scripts/common.sh@364 -- $ (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:01:40.760 23:35:28 build_native_dpdk -- scripts/common.sh@365 -- $ decimal 22 00:01:40.760 23:35:28 build_native_dpdk -- scripts/common.sh@353 -- $ local d=22 00:01:40.760 23:35:28 build_native_dpdk -- scripts/common.sh@354 -- $ [[ 22 =~ ^[0-9]+$ ]] 00:01:40.760 23:35:28 build_native_dpdk -- scripts/common.sh@355 -- $ echo 22 00:01:40.760 23:35:28 build_native_dpdk -- scripts/common.sh@365 -- $ ver1[v]=22 00:01:40.760 23:35:28 build_native_dpdk -- scripts/common.sh@366 -- $ decimal 24 00:01:40.760 23:35:28 build_native_dpdk -- scripts/common.sh@353 -- $ local d=24 00:01:40.760 23:35:28 build_native_dpdk -- scripts/common.sh@354 -- $ [[ 24 =~ ^[0-9]+$ ]] 00:01:40.760 23:35:28 build_native_dpdk -- scripts/common.sh@355 -- $ echo 24 00:01:40.760 23:35:28 build_native_dpdk -- scripts/common.sh@366 -- $ ver2[v]=24 00:01:40.760 23:35:28 build_native_dpdk -- scripts/common.sh@367 -- $ (( ver1[v] > ver2[v] )) 00:01:40.760 23:35:28 build_native_dpdk -- scripts/common.sh@368 -- $ (( ver1[v] < ver2[v] )) 00:01:40.760 23:35:28 build_native_dpdk -- scripts/common.sh@368 -- $ return 1 00:01:40.760 23:35:28 build_native_dpdk -- common/autobuild_common.sh@190 -- $ dpdk_kmods=false 00:01:40.760 23:35:28 build_native_dpdk -- common/autobuild_common.sh@191 -- $ uname -s 00:01:40.760 23:35:28 build_native_dpdk -- common/autobuild_common.sh@191 -- $ '[' Linux = FreeBSD ']' 00:01:40.760 23:35:28 build_native_dpdk -- common/autobuild_common.sh@195 -- $ printf %s, bus bus/pci bus/vdev mempool/ring net/i40e net/i40e/base power/acpi power/amd_pstate power/cppc power/intel_pstate power/intel_uncore power/kvm_vm 00:01:40.760 23:35:28 build_native_dpdk -- common/autobuild_common.sh@195 -- $ meson build-tmp --prefix=/home/vagrant/spdk_repo/dpdk/build --libdir lib -Denable_docs=false -Denable_kmods=false -Dtests=false -Dc_link_args= '-Dc_args=-fPIC -g -fcommon -Werror -Wno-stringop-overflow' -Dmachine=native -Denable_drivers=bus,bus/pci,bus/vdev,mempool/ring,net/i40e,net/i40e/base,power/acpi,power/amd_pstate,power/cppc,power/intel_pstate,power/intel_uncore,power/kvm_vm, 00:01:46.086 The Meson build system 00:01:46.086 Version: 1.5.0 00:01:46.086 Source dir: /home/vagrant/spdk_repo/dpdk 00:01:46.086 Build dir: /home/vagrant/spdk_repo/dpdk/build-tmp 00:01:46.086 Build type: native build 00:01:46.086 Program cat found: YES (/usr/bin/cat) 00:01:46.086 Project name: DPDK 00:01:46.086 Project version: 22.11.4 00:01:46.086 C compiler for the host machine: gcc (gcc 13.3.1 "gcc (GCC) 13.3.1 20240522 (Red Hat 13.3.1-1)") 00:01:46.086 C linker for the host machine: gcc ld.bfd 2.40-14 00:01:46.086 Host machine cpu family: x86_64 00:01:46.086 Host machine cpu: x86_64 00:01:46.086 Message: ## Building in Developer Mode ## 00:01:46.086 Program pkg-config found: YES (/usr/bin/pkg-config) 00:01:46.086 Program check-symbols.sh found: YES (/home/vagrant/spdk_repo/dpdk/buildtools/check-symbols.sh) 00:01:46.086 Program options-ibverbs-static.sh found: YES (/home/vagrant/spdk_repo/dpdk/buildtools/options-ibverbs-static.sh) 00:01:46.086 Program objdump found: YES (/usr/bin/objdump) 00:01:46.086 Program python3 found: YES (/usr/bin/python3) 00:01:46.086 Program cat found: YES (/usr/bin/cat) 00:01:46.086 config/meson.build:83: WARNING: The "machine" option is deprecated. Please use "cpu_instruction_set" instead. 00:01:46.086 Checking for size of "void *" : 8 00:01:46.086 Checking for size of "void *" : 8 (cached) 00:01:46.086 Library m found: YES 00:01:46.086 Library numa found: YES 00:01:46.086 Has header "numaif.h" : YES 00:01:46.086 Library fdt found: NO 00:01:46.086 Library execinfo found: NO 00:01:46.086 Has header "execinfo.h" : YES 00:01:46.086 Found pkg-config: YES (/usr/bin/pkg-config) 1.9.5 00:01:46.086 Run-time dependency libarchive found: NO (tried pkgconfig) 00:01:46.086 Run-time dependency libbsd found: NO (tried pkgconfig) 00:01:46.086 Run-time dependency jansson found: NO (tried pkgconfig) 00:01:46.086 Run-time dependency openssl found: YES 3.1.1 00:01:46.086 Run-time dependency libpcap found: YES 1.10.4 00:01:46.086 Has header "pcap.h" with dependency libpcap: YES 00:01:46.086 Compiler for C supports arguments -Wcast-qual: YES 00:01:46.086 Compiler for C supports arguments -Wdeprecated: YES 00:01:46.086 Compiler for C supports arguments -Wformat: YES 00:01:46.086 Compiler for C supports arguments -Wformat-nonliteral: NO 00:01:46.086 Compiler for C supports arguments -Wformat-security: NO 00:01:46.086 Compiler for C supports arguments -Wmissing-declarations: YES 00:01:46.086 Compiler for C supports arguments -Wmissing-prototypes: YES 00:01:46.086 Compiler for C supports arguments -Wnested-externs: YES 00:01:46.086 Compiler for C supports arguments -Wold-style-definition: YES 00:01:46.086 Compiler for C supports arguments -Wpointer-arith: YES 00:01:46.086 Compiler for C supports arguments -Wsign-compare: YES 00:01:46.086 Compiler for C supports arguments -Wstrict-prototypes: YES 00:01:46.086 Compiler for C supports arguments -Wundef: YES 00:01:46.086 Compiler for C supports arguments -Wwrite-strings: YES 00:01:46.086 Compiler for C supports arguments -Wno-address-of-packed-member: YES 00:01:46.086 Compiler for C supports arguments -Wno-packed-not-aligned: YES 00:01:46.086 Compiler for C supports arguments -Wno-missing-field-initializers: YES 00:01:46.086 Compiler for C supports arguments -Wno-zero-length-bounds: YES 00:01:46.086 Compiler for C supports arguments -mavx512f: YES 00:01:46.086 Checking if "AVX512 checking" compiles: YES 00:01:46.086 Fetching value of define "__SSE4_2__" : 1 00:01:46.086 Fetching value of define "__AES__" : 1 00:01:46.086 Fetching value of define "__AVX__" : 1 00:01:46.086 Fetching value of define "__AVX2__" : 1 00:01:46.086 Fetching value of define "__AVX512BW__" : 1 00:01:46.086 Fetching value of define "__AVX512CD__" : 1 00:01:46.086 Fetching value of define "__AVX512DQ__" : 1 00:01:46.086 Fetching value of define "__AVX512F__" : 1 00:01:46.086 Fetching value of define "__AVX512VL__" : 1 00:01:46.086 Fetching value of define "__PCLMUL__" : 1 00:01:46.086 Fetching value of define "__RDRND__" : 1 00:01:46.086 Fetching value of define "__RDSEED__" : 1 00:01:46.086 Fetching value of define "__VPCLMULQDQ__" : (undefined) 00:01:46.086 Compiler for C supports arguments -Wno-format-truncation: YES 00:01:46.086 Message: lib/kvargs: Defining dependency "kvargs" 00:01:46.086 Message: lib/telemetry: Defining dependency "telemetry" 00:01:46.086 Checking for function "getentropy" : YES 00:01:46.086 Message: lib/eal: Defining dependency "eal" 00:01:46.086 Message: lib/ring: Defining dependency "ring" 00:01:46.086 Message: lib/rcu: Defining dependency "rcu" 00:01:46.086 Message: lib/mempool: Defining dependency "mempool" 00:01:46.086 Message: lib/mbuf: Defining dependency "mbuf" 00:01:46.086 Fetching value of define "__PCLMUL__" : 1 (cached) 00:01:46.086 Fetching value of define "__AVX512F__" : 1 (cached) 00:01:46.086 Fetching value of define "__AVX512BW__" : 1 (cached) 00:01:46.086 Fetching value of define "__AVX512DQ__" : 1 (cached) 00:01:46.086 Fetching value of define "__AVX512VL__" : 1 (cached) 00:01:46.086 Fetching value of define "__VPCLMULQDQ__" : (undefined) (cached) 00:01:46.086 Compiler for C supports arguments -mpclmul: YES 00:01:46.086 Compiler for C supports arguments -maes: YES 00:01:46.086 Compiler for C supports arguments -mavx512f: YES (cached) 00:01:46.086 Compiler for C supports arguments -mavx512bw: YES 00:01:46.086 Compiler for C supports arguments -mavx512dq: YES 00:01:46.086 Compiler for C supports arguments -mavx512vl: YES 00:01:46.086 Compiler for C supports arguments -mvpclmulqdq: YES 00:01:46.086 Compiler for C supports arguments -mavx2: YES 00:01:46.086 Compiler for C supports arguments -mavx: YES 00:01:46.086 Message: lib/net: Defining dependency "net" 00:01:46.086 Message: lib/meter: Defining dependency "meter" 00:01:46.086 Message: lib/ethdev: Defining dependency "ethdev" 00:01:46.086 Message: lib/pci: Defining dependency "pci" 00:01:46.086 Message: lib/cmdline: Defining dependency "cmdline" 00:01:46.086 Message: lib/metrics: Defining dependency "metrics" 00:01:46.086 Message: lib/hash: Defining dependency "hash" 00:01:46.086 Message: lib/timer: Defining dependency "timer" 00:01:46.086 Fetching value of define "__AVX2__" : 1 (cached) 00:01:46.086 Fetching value of define "__AVX512F__" : 1 (cached) 00:01:46.086 Fetching value of define "__AVX512VL__" : 1 (cached) 00:01:46.086 Fetching value of define "__AVX512CD__" : 1 (cached) 00:01:46.086 Fetching value of define "__AVX512BW__" : 1 (cached) 00:01:46.086 Message: lib/acl: Defining dependency "acl" 00:01:46.086 Message: lib/bbdev: Defining dependency "bbdev" 00:01:46.086 Message: lib/bitratestats: Defining dependency "bitratestats" 00:01:46.086 Run-time dependency libelf found: YES 0.191 00:01:46.086 Message: lib/bpf: Defining dependency "bpf" 00:01:46.086 Message: lib/cfgfile: Defining dependency "cfgfile" 00:01:46.086 Message: lib/compressdev: Defining dependency "compressdev" 00:01:46.087 Message: lib/cryptodev: Defining dependency "cryptodev" 00:01:46.087 Message: lib/distributor: Defining dependency "distributor" 00:01:46.087 Message: lib/efd: Defining dependency "efd" 00:01:46.087 Message: lib/eventdev: Defining dependency "eventdev" 00:01:46.087 Message: lib/gpudev: Defining dependency "gpudev" 00:01:46.087 Message: lib/gro: Defining dependency "gro" 00:01:46.087 Message: lib/gso: Defining dependency "gso" 00:01:46.087 Message: lib/ip_frag: Defining dependency "ip_frag" 00:01:46.087 Message: lib/jobstats: Defining dependency "jobstats" 00:01:46.087 Message: lib/latencystats: Defining dependency "latencystats" 00:01:46.087 Message: lib/lpm: Defining dependency "lpm" 00:01:46.087 Fetching value of define "__AVX512F__" : 1 (cached) 00:01:46.087 Fetching value of define "__AVX512DQ__" : 1 (cached) 00:01:46.087 Fetching value of define "__AVX512IFMA__" : (undefined) 00:01:46.087 Compiler for C supports arguments -mavx512f -mavx512dq -mavx512ifma: YES 00:01:46.087 Message: lib/member: Defining dependency "member" 00:01:46.087 Message: lib/pcapng: Defining dependency "pcapng" 00:01:46.087 Compiler for C supports arguments -Wno-cast-qual: YES 00:01:46.087 Message: lib/power: Defining dependency "power" 00:01:46.087 Message: lib/rawdev: Defining dependency "rawdev" 00:01:46.087 Message: lib/regexdev: Defining dependency "regexdev" 00:01:46.087 Message: lib/dmadev: Defining dependency "dmadev" 00:01:46.087 Message: lib/rib: Defining dependency "rib" 00:01:46.087 Message: lib/reorder: Defining dependency "reorder" 00:01:46.087 Message: lib/sched: Defining dependency "sched" 00:01:46.087 Message: lib/security: Defining dependency "security" 00:01:46.087 Message: lib/stack: Defining dependency "stack" 00:01:46.087 Has header "linux/userfaultfd.h" : YES 00:01:46.087 Message: lib/vhost: Defining dependency "vhost" 00:01:46.087 Message: lib/ipsec: Defining dependency "ipsec" 00:01:46.087 Fetching value of define "__AVX512F__" : 1 (cached) 00:01:46.087 Fetching value of define "__AVX512DQ__" : 1 (cached) 00:01:46.087 Fetching value of define "__AVX512BW__" : 1 (cached) 00:01:46.087 Message: lib/fib: Defining dependency "fib" 00:01:46.087 Message: lib/port: Defining dependency "port" 00:01:46.087 Message: lib/pdump: Defining dependency "pdump" 00:01:46.087 Message: lib/table: Defining dependency "table" 00:01:46.087 Message: lib/pipeline: Defining dependency "pipeline" 00:01:46.087 Message: lib/graph: Defining dependency "graph" 00:01:46.087 Message: lib/node: Defining dependency "node" 00:01:46.087 Compiler for C supports arguments -Wno-format-truncation: YES (cached) 00:01:46.087 Message: drivers/bus/pci: Defining dependency "bus_pci" 00:01:46.087 Message: drivers/bus/vdev: Defining dependency "bus_vdev" 00:01:46.087 Message: drivers/mempool/ring: Defining dependency "mempool_ring" 00:01:46.087 Compiler for C supports arguments -Wno-sign-compare: YES 00:01:46.087 Compiler for C supports arguments -Wno-unused-value: YES 00:01:46.087 Compiler for C supports arguments -Wno-format: YES 00:01:46.087 Compiler for C supports arguments -Wno-format-security: YES 00:01:46.087 Compiler for C supports arguments -Wno-format-nonliteral: YES 00:01:46.087 Compiler for C supports arguments -Wno-strict-aliasing: YES 00:01:47.470 Compiler for C supports arguments -Wno-unused-but-set-variable: YES 00:01:47.470 Compiler for C supports arguments -Wno-unused-parameter: YES 00:01:47.470 Fetching value of define "__AVX2__" : 1 (cached) 00:01:47.470 Fetching value of define "__AVX512F__" : 1 (cached) 00:01:47.470 Fetching value of define "__AVX512BW__" : 1 (cached) 00:01:47.470 Compiler for C supports arguments -mavx512f: YES (cached) 00:01:47.470 Compiler for C supports arguments -mavx512bw: YES (cached) 00:01:47.470 Compiler for C supports arguments -march=skylake-avx512: YES 00:01:47.470 Message: drivers/net/i40e: Defining dependency "net_i40e" 00:01:47.470 Program doxygen found: YES (/usr/local/bin/doxygen) 00:01:47.470 Configuring doxy-api.conf using configuration 00:01:47.470 Program sphinx-build found: NO 00:01:47.470 Configuring rte_build_config.h using configuration 00:01:47.470 Message: 00:01:47.470 ================= 00:01:47.470 Applications Enabled 00:01:47.470 ================= 00:01:47.470 00:01:47.470 apps: 00:01:47.470 dumpcap, pdump, proc-info, test-acl, test-bbdev, test-cmdline, test-compress-perf, test-crypto-perf, 00:01:47.470 test-eventdev, test-fib, test-flow-perf, test-gpudev, test-pipeline, test-pmd, test-regex, test-sad, 00:01:47.470 test-security-perf, 00:01:47.470 00:01:47.470 Message: 00:01:47.470 ================= 00:01:47.470 Libraries Enabled 00:01:47.470 ================= 00:01:47.470 00:01:47.470 libs: 00:01:47.470 kvargs, telemetry, eal, ring, rcu, mempool, mbuf, net, 00:01:47.470 meter, ethdev, pci, cmdline, metrics, hash, timer, acl, 00:01:47.470 bbdev, bitratestats, bpf, cfgfile, compressdev, cryptodev, distributor, efd, 00:01:47.470 eventdev, gpudev, gro, gso, ip_frag, jobstats, latencystats, lpm, 00:01:47.470 member, pcapng, power, rawdev, regexdev, dmadev, rib, reorder, 00:01:47.470 sched, security, stack, vhost, ipsec, fib, port, pdump, 00:01:47.470 table, pipeline, graph, node, 00:01:47.470 00:01:47.470 Message: 00:01:47.470 =============== 00:01:47.470 Drivers Enabled 00:01:47.470 =============== 00:01:47.470 00:01:47.470 common: 00:01:47.470 00:01:47.470 bus: 00:01:47.470 pci, vdev, 00:01:47.470 mempool: 00:01:47.470 ring, 00:01:47.470 dma: 00:01:47.470 00:01:47.470 net: 00:01:47.470 i40e, 00:01:47.470 raw: 00:01:47.470 00:01:47.470 crypto: 00:01:47.470 00:01:47.470 compress: 00:01:47.470 00:01:47.470 regex: 00:01:47.470 00:01:47.470 vdpa: 00:01:47.470 00:01:47.470 event: 00:01:47.470 00:01:47.470 baseband: 00:01:47.470 00:01:47.470 gpu: 00:01:47.470 00:01:47.470 00:01:47.470 Message: 00:01:47.470 ================= 00:01:47.470 Content Skipped 00:01:47.470 ================= 00:01:47.470 00:01:47.470 apps: 00:01:47.470 00:01:47.470 libs: 00:01:47.470 kni: explicitly disabled via build config (deprecated lib) 00:01:47.470 flow_classify: explicitly disabled via build config (deprecated lib) 00:01:47.470 00:01:47.470 drivers: 00:01:47.470 common/cpt: not in enabled drivers build config 00:01:47.470 common/dpaax: not in enabled drivers build config 00:01:47.470 common/iavf: not in enabled drivers build config 00:01:47.470 common/idpf: not in enabled drivers build config 00:01:47.470 common/mvep: not in enabled drivers build config 00:01:47.470 common/octeontx: not in enabled drivers build config 00:01:47.470 bus/auxiliary: not in enabled drivers build config 00:01:47.470 bus/dpaa: not in enabled drivers build config 00:01:47.470 bus/fslmc: not in enabled drivers build config 00:01:47.470 bus/ifpga: not in enabled drivers build config 00:01:47.470 bus/vmbus: not in enabled drivers build config 00:01:47.470 common/cnxk: not in enabled drivers build config 00:01:47.470 common/mlx5: not in enabled drivers build config 00:01:47.470 common/qat: not in enabled drivers build config 00:01:47.470 common/sfc_efx: not in enabled drivers build config 00:01:47.470 mempool/bucket: not in enabled drivers build config 00:01:47.470 mempool/cnxk: not in enabled drivers build config 00:01:47.470 mempool/dpaa: not in enabled drivers build config 00:01:47.470 mempool/dpaa2: not in enabled drivers build config 00:01:47.470 mempool/octeontx: not in enabled drivers build config 00:01:47.470 mempool/stack: not in enabled drivers build config 00:01:47.470 dma/cnxk: not in enabled drivers build config 00:01:47.470 dma/dpaa: not in enabled drivers build config 00:01:47.470 dma/dpaa2: not in enabled drivers build config 00:01:47.470 dma/hisilicon: not in enabled drivers build config 00:01:47.470 dma/idxd: not in enabled drivers build config 00:01:47.471 dma/ioat: not in enabled drivers build config 00:01:47.471 dma/skeleton: not in enabled drivers build config 00:01:47.471 net/af_packet: not in enabled drivers build config 00:01:47.471 net/af_xdp: not in enabled drivers build config 00:01:47.471 net/ark: not in enabled drivers build config 00:01:47.471 net/atlantic: not in enabled drivers build config 00:01:47.471 net/avp: not in enabled drivers build config 00:01:47.471 net/axgbe: not in enabled drivers build config 00:01:47.471 net/bnx2x: not in enabled drivers build config 00:01:47.471 net/bnxt: not in enabled drivers build config 00:01:47.471 net/bonding: not in enabled drivers build config 00:01:47.471 net/cnxk: not in enabled drivers build config 00:01:47.471 net/cxgbe: not in enabled drivers build config 00:01:47.471 net/dpaa: not in enabled drivers build config 00:01:47.471 net/dpaa2: not in enabled drivers build config 00:01:47.471 net/e1000: not in enabled drivers build config 00:01:47.471 net/ena: not in enabled drivers build config 00:01:47.471 net/enetc: not in enabled drivers build config 00:01:47.471 net/enetfec: not in enabled drivers build config 00:01:47.471 net/enic: not in enabled drivers build config 00:01:47.471 net/failsafe: not in enabled drivers build config 00:01:47.471 net/fm10k: not in enabled drivers build config 00:01:47.471 net/gve: not in enabled drivers build config 00:01:47.471 net/hinic: not in enabled drivers build config 00:01:47.471 net/hns3: not in enabled drivers build config 00:01:47.471 net/iavf: not in enabled drivers build config 00:01:47.471 net/ice: not in enabled drivers build config 00:01:47.471 net/idpf: not in enabled drivers build config 00:01:47.471 net/igc: not in enabled drivers build config 00:01:47.471 net/ionic: not in enabled drivers build config 00:01:47.471 net/ipn3ke: not in enabled drivers build config 00:01:47.471 net/ixgbe: not in enabled drivers build config 00:01:47.471 net/kni: not in enabled drivers build config 00:01:47.471 net/liquidio: not in enabled drivers build config 00:01:47.471 net/mana: not in enabled drivers build config 00:01:47.471 net/memif: not in enabled drivers build config 00:01:47.471 net/mlx4: not in enabled drivers build config 00:01:47.471 net/mlx5: not in enabled drivers build config 00:01:47.471 net/mvneta: not in enabled drivers build config 00:01:47.471 net/mvpp2: not in enabled drivers build config 00:01:47.471 net/netvsc: not in enabled drivers build config 00:01:47.471 net/nfb: not in enabled drivers build config 00:01:47.471 net/nfp: not in enabled drivers build config 00:01:47.471 net/ngbe: not in enabled drivers build config 00:01:47.471 net/null: not in enabled drivers build config 00:01:47.471 net/octeontx: not in enabled drivers build config 00:01:47.471 net/octeon_ep: not in enabled drivers build config 00:01:47.471 net/pcap: not in enabled drivers build config 00:01:47.471 net/pfe: not in enabled drivers build config 00:01:47.471 net/qede: not in enabled drivers build config 00:01:47.471 net/ring: not in enabled drivers build config 00:01:47.471 net/sfc: not in enabled drivers build config 00:01:47.471 net/softnic: not in enabled drivers build config 00:01:47.471 net/tap: not in enabled drivers build config 00:01:47.471 net/thunderx: not in enabled drivers build config 00:01:47.471 net/txgbe: not in enabled drivers build config 00:01:47.471 net/vdev_netvsc: not in enabled drivers build config 00:01:47.471 net/vhost: not in enabled drivers build config 00:01:47.471 net/virtio: not in enabled drivers build config 00:01:47.471 net/vmxnet3: not in enabled drivers build config 00:01:47.471 raw/cnxk_bphy: not in enabled drivers build config 00:01:47.471 raw/cnxk_gpio: not in enabled drivers build config 00:01:47.471 raw/dpaa2_cmdif: not in enabled drivers build config 00:01:47.471 raw/ifpga: not in enabled drivers build config 00:01:47.471 raw/ntb: not in enabled drivers build config 00:01:47.471 raw/skeleton: not in enabled drivers build config 00:01:47.471 crypto/armv8: not in enabled drivers build config 00:01:47.471 crypto/bcmfs: not in enabled drivers build config 00:01:47.471 crypto/caam_jr: not in enabled drivers build config 00:01:47.471 crypto/ccp: not in enabled drivers build config 00:01:47.471 crypto/cnxk: not in enabled drivers build config 00:01:47.471 crypto/dpaa_sec: not in enabled drivers build config 00:01:47.471 crypto/dpaa2_sec: not in enabled drivers build config 00:01:47.471 crypto/ipsec_mb: not in enabled drivers build config 00:01:47.471 crypto/mlx5: not in enabled drivers build config 00:01:47.471 crypto/mvsam: not in enabled drivers build config 00:01:47.471 crypto/nitrox: not in enabled drivers build config 00:01:47.471 crypto/null: not in enabled drivers build config 00:01:47.471 crypto/octeontx: not in enabled drivers build config 00:01:47.471 crypto/openssl: not in enabled drivers build config 00:01:47.471 crypto/scheduler: not in enabled drivers build config 00:01:47.471 crypto/uadk: not in enabled drivers build config 00:01:47.471 crypto/virtio: not in enabled drivers build config 00:01:47.471 compress/isal: not in enabled drivers build config 00:01:47.471 compress/mlx5: not in enabled drivers build config 00:01:47.471 compress/octeontx: not in enabled drivers build config 00:01:47.471 compress/zlib: not in enabled drivers build config 00:01:47.471 regex/mlx5: not in enabled drivers build config 00:01:47.471 regex/cn9k: not in enabled drivers build config 00:01:47.471 vdpa/ifc: not in enabled drivers build config 00:01:47.471 vdpa/mlx5: not in enabled drivers build config 00:01:47.471 vdpa/sfc: not in enabled drivers build config 00:01:47.471 event/cnxk: not in enabled drivers build config 00:01:47.471 event/dlb2: not in enabled drivers build config 00:01:47.471 event/dpaa: not in enabled drivers build config 00:01:47.471 event/dpaa2: not in enabled drivers build config 00:01:47.471 event/dsw: not in enabled drivers build config 00:01:47.471 event/opdl: not in enabled drivers build config 00:01:47.471 event/skeleton: not in enabled drivers build config 00:01:47.471 event/sw: not in enabled drivers build config 00:01:47.471 event/octeontx: not in enabled drivers build config 00:01:47.471 baseband/acc: not in enabled drivers build config 00:01:47.471 baseband/fpga_5gnr_fec: not in enabled drivers build config 00:01:47.471 baseband/fpga_lte_fec: not in enabled drivers build config 00:01:47.471 baseband/la12xx: not in enabled drivers build config 00:01:47.471 baseband/null: not in enabled drivers build config 00:01:47.471 baseband/turbo_sw: not in enabled drivers build config 00:01:47.471 gpu/cuda: not in enabled drivers build config 00:01:47.471 00:01:47.471 00:01:47.471 Build targets in project: 311 00:01:47.471 00:01:47.471 DPDK 22.11.4 00:01:47.471 00:01:47.471 User defined options 00:01:47.471 libdir : lib 00:01:47.471 prefix : /home/vagrant/spdk_repo/dpdk/build 00:01:47.471 c_args : -fPIC -g -fcommon -Werror -Wno-stringop-overflow 00:01:47.471 c_link_args : 00:01:47.471 enable_docs : false 00:01:47.471 enable_drivers: bus,bus/pci,bus/vdev,mempool/ring,net/i40e,net/i40e/base,power/acpi,power/amd_pstate,power/cppc,power/intel_pstate,power/intel_uncore,power/kvm_vm, 00:01:47.471 enable_kmods : false 00:01:47.471 machine : native 00:01:47.471 tests : false 00:01:47.471 00:01:47.471 Found ninja-1.11.1.git.kitware.jobserver-1 at /usr/local/bin/ninja 00:01:47.471 WARNING: Running the setup command as `meson [options]` instead of `meson setup [options]` is ambiguous and deprecated. 00:01:47.471 23:35:35 build_native_dpdk -- common/autobuild_common.sh@199 -- $ ninja -C /home/vagrant/spdk_repo/dpdk/build-tmp -j10 00:01:47.471 ninja: Entering directory `/home/vagrant/spdk_repo/dpdk/build-tmp' 00:01:47.471 [1/740] Generating lib/rte_kvargs_def with a custom command 00:01:47.471 [2/740] Generating lib/rte_kvargs_mingw with a custom command 00:01:47.471 [3/740] Generating lib/rte_telemetry_mingw with a custom command 00:01:47.471 [4/740] Generating lib/rte_telemetry_def with a custom command 00:01:47.471 [5/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_class.c.o 00:01:47.471 [6/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_debug.c.o 00:01:47.471 [7/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_errno.c.o 00:01:47.471 [8/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_config.c.o 00:01:47.471 [9/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_bus.c.o 00:01:47.471 [10/740] Compiling C object lib/librte_kvargs.a.p/kvargs_rte_kvargs.c.o 00:01:47.472 [11/740] Compiling C object lib/librte_telemetry.a.p/telemetry_telemetry_legacy.c.o 00:01:47.472 [12/740] Linking static target lib/librte_kvargs.a 00:01:47.472 [13/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_hexdump.c.o 00:01:47.472 [14/740] Compiling C object lib/librte_telemetry.a.p/telemetry_telemetry_data.c.o 00:01:47.731 [15/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_launch.c.o 00:01:47.731 [16/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_dev.c.o 00:01:47.731 [17/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_devargs.c.o 00:01:47.731 [18/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_interrupts.c.o 00:01:47.731 [19/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_mcfg.c.o 00:01:47.731 [20/740] Generating lib/kvargs.sym_chk with a custom command (wrapped by meson to capture output) 00:01:47.731 [21/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_lcore.c.o 00:01:47.731 [22/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_log.c.o 00:01:47.731 [23/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_string_fns.c.o 00:01:47.731 [24/740] Linking target lib/librte_kvargs.so.23.0 00:01:47.731 [25/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_memalloc.c.o 00:01:47.731 [26/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_timer.c.o 00:01:47.731 [27/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_tailqs.c.o 00:01:47.998 [28/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_memzone.c.o 00:01:47.998 [29/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_fbarray.c.o 00:01:47.998 [30/740] Compiling C object lib/librte_telemetry.a.p/telemetry_telemetry.c.o 00:01:47.998 [31/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_thread.c.o 00:01:47.998 [32/740] Linking static target lib/librte_telemetry.a 00:01:47.998 [33/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_uuid.c.o 00:01:47.998 [34/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_trace_points.c.o 00:01:47.998 [35/740] Compiling C object lib/librte_eal.a.p/eal_common_rte_reciprocal.c.o 00:01:47.998 [36/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_memory.c.o 00:01:47.998 [37/740] Compiling C object lib/librte_eal.a.p/eal_common_rte_version.c.o 00:01:47.998 [38/740] Generating symbol file lib/librte_kvargs.so.23.0.p/librte_kvargs.so.23.0.symbols 00:01:47.998 [39/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_cpuflags.c.o 00:01:47.998 [40/740] Compiling C object lib/librte_eal.a.p/eal_common_malloc_elem.c.o 00:01:47.998 [41/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_hypervisor.c.o 00:01:48.264 [42/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_dynmem.c.o 00:01:48.264 [43/740] Generating lib/telemetry.sym_chk with a custom command (wrapped by meson to capture output) 00:01:48.264 [44/740] Linking target lib/librte_telemetry.so.23.0 00:01:48.264 [45/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_trace_ctf.c.o 00:01:48.264 [46/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_trace.c.o 00:01:48.264 [47/740] Compiling C object lib/librte_eal.a.p/eal_common_rte_random.c.o 00:01:48.264 [48/740] Compiling C object lib/librte_eal.a.p/eal_common_rte_service.c.o 00:01:48.264 [49/740] Compiling C object lib/librte_eal.a.p/eal_common_malloc_heap.c.o 00:01:48.264 [50/740] Generating symbol file lib/librte_telemetry.so.23.0.p/librte_telemetry.so.23.0.symbols 00:01:48.264 [51/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_trace_utils.c.o 00:01:48.264 [52/740] Compiling C object lib/librte_eal.a.p/eal_common_rte_keepalive.c.o 00:01:48.264 [53/740] Compiling C object lib/librte_eal.a.p/eal_unix_eal_debug.c.o 00:01:48.264 [54/740] Compiling C object lib/librte_eal.a.p/eal_unix_eal_file.c.o 00:01:48.264 [55/740] Compiling C object lib/librte_eal.a.p/eal_common_hotplug_mp.c.o 00:01:48.524 [56/740] Compiling C object lib/librte_eal.a.p/eal_unix_eal_firmware.c.o 00:01:48.524 [57/740] Compiling C object lib/librte_eal.a.p/eal_unix_eal_filesystem.c.o 00:01:48.524 [58/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_proc.c.o 00:01:48.524 [59/740] Compiling C object lib/librte_eal.a.p/eal_unix_eal_unix_timer.c.o 00:01:48.524 [60/740] Compiling C object lib/librte_eal.a.p/eal_common_malloc_mp.c.o 00:01:48.524 [61/740] Compiling C object lib/librte_eal.a.p/eal_unix_eal_unix_thread.c.o 00:01:48.524 [62/740] Compiling C object lib/librte_eal.a.p/eal_unix_eal_unix_memory.c.o 00:01:48.524 [63/740] Compiling C object lib/librte_eal.a.p/eal_common_rte_malloc.c.o 00:01:48.524 [64/740] Compiling C object lib/librte_eal.a.p/eal_linux_eal_cpuflags.c.o 00:01:48.524 [65/740] Compiling C object lib/librte_eal.a.p/eal_unix_rte_thread.c.o 00:01:48.524 [66/740] Compiling C object lib/librte_eal.a.p/eal_linux_eal_log.c.o 00:01:48.524 [67/740] Compiling C object lib/librte_eal.a.p/eal_linux_eal_lcore.c.o 00:01:48.524 [68/740] Compiling C object lib/librte_eal.a.p/eal_linux_eal_thread.c.o 00:01:48.524 [69/740] Compiling C object lib/librte_eal.a.p/eal_linux_eal_alarm.c.o 00:01:48.524 [70/740] Compiling C object lib/librte_eal.a.p/eal_linux_eal_dev.c.o 00:01:48.524 [71/740] Compiling C object lib/librte_eal.a.p/eal_linux_eal_timer.c.o 00:01:48.524 [72/740] Compiling C object lib/librte_eal.a.p/eal_linux_eal_vfio_mp_sync.c.o 00:01:48.524 [73/740] Compiling C object lib/librte_eal.a.p/eal_x86_rte_cpuflags.c.o 00:01:48.524 [74/740] Compiling C object lib/librte_eal.a.p/eal_linux_eal_hugepage_info.c.o 00:01:48.524 [75/740] Compiling C object lib/librte_eal.a.p/eal_x86_rte_hypervisor.c.o 00:01:48.524 [76/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_options.c.o 00:01:48.791 [77/740] Compiling C object lib/librte_eal.a.p/eal_x86_rte_spinlock.c.o 00:01:48.791 [78/740] Generating lib/rte_eal_def with a custom command 00:01:48.791 [79/740] Generating lib/rte_eal_mingw with a custom command 00:01:48.791 [80/740] Compiling C object lib/librte_eal.a.p/eal_x86_rte_cycles.c.o 00:01:48.791 [81/740] Generating lib/rte_ring_def with a custom command 00:01:48.791 [82/740] Generating lib/rte_ring_mingw with a custom command 00:01:48.791 [83/740] Generating lib/rte_rcu_def with a custom command 00:01:48.791 [84/740] Generating lib/rte_rcu_mingw with a custom command 00:01:48.791 [85/740] Compiling C object lib/librte_eal.a.p/eal_linux_eal.c.o 00:01:48.791 [86/740] Compiling C object lib/librte_eal.a.p/eal_x86_rte_power_intrinsics.c.o 00:01:48.791 [87/740] Compiling C object lib/librte_ring.a.p/ring_rte_ring.c.o 00:01:48.791 [88/740] Linking static target lib/librte_ring.a 00:01:48.791 [89/740] Compiling C object lib/librte_eal.a.p/eal_linux_eal_interrupts.c.o 00:01:48.791 [90/740] Generating lib/rte_mempool_def with a custom command 00:01:48.791 [91/740] Generating lib/rte_mempool_mingw with a custom command 00:01:49.057 [92/740] Compiling C object lib/librte_eal.a.p/eal_linux_eal_memalloc.c.o 00:01:49.058 [93/740] Compiling C object lib/librte_eal.a.p/eal_linux_eal_memory.c.o 00:01:49.058 [94/740] Generating lib/ring.sym_chk with a custom command (wrapped by meson to capture output) 00:01:49.058 [95/740] Compiling C object lib/librte_mempool.a.p/mempool_rte_mempool_ops_default.c.o 00:01:49.058 [96/740] Generating lib/rte_mbuf_def with a custom command 00:01:49.058 [97/740] Generating lib/rte_mbuf_mingw with a custom command 00:01:49.058 [98/740] Compiling C object lib/librte_mempool.a.p/mempool_rte_mempool_ops.c.o 00:01:49.058 [99/740] Compiling C object lib/librte_eal.a.p/eal_linux_eal_vfio.c.o 00:01:49.058 [100/740] Compiling C object lib/librte_mempool.a.p/mempool_mempool_trace_points.c.o 00:01:49.318 [101/740] Linking static target lib/librte_eal.a 00:01:49.318 [102/740] Compiling C object lib/librte_mbuf.a.p/mbuf_rte_mbuf_pool_ops.c.o 00:01:49.318 [103/740] Compiling C object lib/librte_mbuf.a.p/mbuf_rte_mbuf_ptype.c.o 00:01:49.318 [104/740] Compiling C object lib/librte_rcu.a.p/rcu_rte_rcu_qsbr.c.o 00:01:49.318 [105/740] Linking static target lib/librte_rcu.a 00:01:49.579 [106/740] Compiling C object lib/librte_mempool.a.p/mempool_rte_mempool.c.o 00:01:49.579 [107/740] Linking static target lib/librte_mempool.a 00:01:49.579 [108/740] Compiling C object lib/librte_mbuf.a.p/mbuf_rte_mbuf_dyn.c.o 00:01:49.579 [109/740] Compiling C object lib/librte_net.a.p/net_rte_ether.c.o 00:01:49.579 [110/740] Generating lib/rte_net_def with a custom command 00:01:49.579 [111/740] Compiling C object lib/net/libnet_crc_avx512_lib.a.p/net_crc_avx512.c.o 00:01:49.579 [112/740] Linking static target lib/net/libnet_crc_avx512_lib.a 00:01:49.579 [113/740] Generating lib/rte_net_mingw with a custom command 00:01:49.579 [114/740] Compiling C object lib/librte_net.a.p/net_rte_arp.c.o 00:01:49.579 [115/740] Generating lib/rte_meter_def with a custom command 00:01:49.579 [116/740] Generating lib/rte_meter_mingw with a custom command 00:01:49.579 [117/740] Generating lib/rcu.sym_chk with a custom command (wrapped by meson to capture output) 00:01:49.579 [118/740] Compiling C object lib/librte_meter.a.p/meter_rte_meter.c.o 00:01:49.579 [119/740] Linking static target lib/librte_meter.a 00:01:49.840 [120/740] Compiling C object lib/librte_net.a.p/net_rte_net.c.o 00:01:49.840 [121/740] Compiling C object lib/librte_net.a.p/net_rte_net_crc.c.o 00:01:49.840 [122/740] Generating lib/meter.sym_chk with a custom command (wrapped by meson to capture output) 00:01:49.840 [123/740] Compiling C object lib/librte_net.a.p/net_net_crc_sse.c.o 00:01:49.840 [124/740] Linking static target lib/librte_net.a 00:01:49.840 [125/740] Compiling C object lib/librte_mbuf.a.p/mbuf_rte_mbuf.c.o 00:01:49.840 [126/740] Linking static target lib/librte_mbuf.a 00:01:50.100 [127/740] Compiling C object lib/librte_ethdev.a.p/ethdev_ethdev_profile.c.o 00:01:50.100 [128/740] Compiling C object lib/librte_ethdev.a.p/ethdev_ethdev_private.c.o 00:01:50.100 [129/740] Compiling C object lib/librte_ethdev.a.p/ethdev_ethdev_trace_points.c.o 00:01:50.100 [130/740] Compiling C object lib/librte_ethdev.a.p/ethdev_ethdev_driver.c.o 00:01:50.100 [131/740] Generating lib/net.sym_chk with a custom command (wrapped by meson to capture output) 00:01:50.100 [132/740] Generating lib/mempool.sym_chk with a custom command (wrapped by meson to capture output) 00:01:50.100 [133/740] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_class_eth.c.o 00:01:50.361 [134/740] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_ethdev_cman.c.o 00:01:50.361 [135/740] Generating lib/mbuf.sym_chk with a custom command (wrapped by meson to capture output) 00:01:50.622 [136/740] Compiling C object lib/librte_ethdev.a.p/ethdev_sff_telemetry.c.o 00:01:50.622 [137/740] Generating lib/rte_ethdev_def with a custom command 00:01:50.622 [138/740] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_mtr.c.o 00:01:50.622 [139/740] Generating lib/rte_ethdev_mingw with a custom command 00:01:50.622 [140/740] Generating lib/rte_pci_def with a custom command 00:01:50.622 [141/740] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_tm.c.o 00:01:50.622 [142/740] Generating lib/rte_pci_mingw with a custom command 00:01:50.622 [143/740] Compiling C object lib/librte_pci.a.p/pci_rte_pci.c.o 00:01:50.622 [144/740] Linking static target lib/librte_pci.a 00:01:50.622 [145/740] Compiling C object lib/librte_ethdev.a.p/ethdev_sff_common.c.o 00:01:50.622 [146/740] Compiling C object lib/librte_ethdev.a.p/ethdev_sff_8079.c.o 00:01:50.622 [147/740] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline.c.o 00:01:50.622 [148/740] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_cirbuf.c.o 00:01:50.882 [149/740] Compiling C object lib/librte_ethdev.a.p/ethdev_sff_8472.c.o 00:01:50.882 [150/740] Generating lib/pci.sym_chk with a custom command (wrapped by meson to capture output) 00:01:50.882 [151/740] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse_num.c.o 00:01:50.882 [152/740] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse.c.o 00:01:50.882 [153/740] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse_portlist.c.o 00:01:50.882 [154/740] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_socket.c.o 00:01:50.882 [155/740] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse_string.c.o 00:01:50.882 [156/740] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_vt100.c.o 00:01:50.882 [157/740] Generating lib/rte_cmdline_def with a custom command 00:01:50.882 [158/740] Generating lib/rte_cmdline_mingw with a custom command 00:01:50.882 [159/740] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_os_unix.c.o 00:01:50.882 [160/740] Generating lib/rte_metrics_def with a custom command 00:01:50.882 [161/740] Compiling C object lib/librte_ethdev.a.p/ethdev_sff_8636.c.o 00:01:50.882 [162/740] Generating lib/rte_metrics_mingw with a custom command 00:01:51.142 [163/740] Compiling C object lib/librte_metrics.a.p/metrics_rte_metrics.c.o 00:01:51.142 [164/740] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse_etheraddr.c.o 00:01:51.142 [165/740] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_rdline.c.o 00:01:51.142 [166/740] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse_ipaddr.c.o 00:01:51.142 [167/740] Generating lib/rte_hash_def with a custom command 00:01:51.142 [168/740] Linking static target lib/librte_cmdline.a 00:01:51.142 [169/740] Generating lib/rte_hash_mingw with a custom command 00:01:51.142 [170/740] Generating lib/rte_timer_def with a custom command 00:01:51.142 [171/740] Compiling C object lib/librte_hash.a.p/hash_rte_fbk_hash.c.o 00:01:51.142 [172/740] Generating lib/rte_timer_mingw with a custom command 00:01:51.142 [173/740] Compiling C object lib/librte_metrics.a.p/metrics_rte_metrics_telemetry.c.o 00:01:51.142 [174/740] Linking static target lib/librte_metrics.a 00:01:51.142 [175/740] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_flow.c.o 00:01:51.406 [176/740] Compiling C object lib/librte_timer.a.p/timer_rte_timer.c.o 00:01:51.406 [177/740] Linking static target lib/librte_timer.a 00:01:51.406 [178/740] Generating lib/metrics.sym_chk with a custom command (wrapped by meson to capture output) 00:01:51.678 [179/740] Compiling C object lib/librte_acl.a.p/acl_acl_gen.c.o 00:01:51.678 [180/740] Generating lib/timer.sym_chk with a custom command (wrapped by meson to capture output) 00:01:51.678 [181/740] Compiling C object lib/librte_hash.a.p/hash_rte_thash.c.o 00:01:51.678 [182/740] Generating lib/cmdline.sym_chk with a custom command (wrapped by meson to capture output) 00:01:51.945 [183/740] Compiling C object lib/librte_acl.a.p/acl_acl_run_scalar.c.o 00:01:51.945 [184/740] Generating lib/rte_acl_def with a custom command 00:01:51.945 [185/740] Generating lib/rte_acl_mingw with a custom command 00:01:51.946 [186/740] Compiling C object lib/librte_acl.a.p/acl_rte_acl.c.o 00:01:51.946 [187/740] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_ethdev.c.o 00:01:51.946 [188/740] Generating lib/rte_bbdev_def with a custom command 00:01:51.946 [189/740] Compiling C object lib/librte_acl.a.p/acl_tb_mem.c.o 00:01:51.946 [190/740] Linking static target lib/librte_ethdev.a 00:01:51.946 [191/740] Generating lib/rte_bbdev_mingw with a custom command 00:01:51.946 [192/740] Generating lib/rte_bitratestats_def with a custom command 00:01:51.946 [193/740] Generating lib/rte_bitratestats_mingw with a custom command 00:01:52.204 [194/740] Compiling C object lib/librte_bitratestats.a.p/bitratestats_rte_bitrate.c.o 00:01:52.204 [195/740] Linking static target lib/librte_bitratestats.a 00:01:52.462 [196/740] Compiling C object lib/librte_bpf.a.p/bpf_bpf.c.o 00:01:52.462 [197/740] Compiling C object lib/librte_acl.a.p/acl_acl_bld.c.o 00:01:52.462 [198/740] Generating lib/bitratestats.sym_chk with a custom command (wrapped by meson to capture output) 00:01:52.462 [199/740] Compiling C object lib/librte_bbdev.a.p/bbdev_rte_bbdev.c.o 00:01:52.462 [200/740] Linking static target lib/librte_bbdev.a 00:01:52.721 [201/740] Compiling C object lib/librte_bpf.a.p/bpf_bpf_dump.c.o 00:01:52.979 [202/740] Compiling C object lib/librte_bpf.a.p/bpf_bpf_load.c.o 00:01:52.979 [203/740] Generating lib/bbdev.sym_chk with a custom command (wrapped by meson to capture output) 00:01:52.979 [204/740] Compiling C object lib/librte_bpf.a.p/bpf_bpf_exec.c.o 00:01:53.238 [205/740] Compiling C object lib/librte_hash.a.p/hash_rte_cuckoo_hash.c.o 00:01:53.238 [206/740] Linking static target lib/librte_hash.a 00:01:53.238 [207/740] Compiling C object lib/librte_acl.a.p/acl_acl_run_sse.c.o 00:01:53.238 [208/740] Compiling C object lib/librte_bpf.a.p/bpf_bpf_stub.c.o 00:01:53.497 [209/740] Compiling C object lib/librte_bpf.a.p/bpf_bpf_pkt.c.o 00:01:53.497 [210/740] Generating lib/rte_bpf_def with a custom command 00:01:53.497 [211/740] Generating lib/rte_bpf_mingw with a custom command 00:01:53.497 [212/740] Compiling C object lib/librte_bpf.a.p/bpf_bpf_load_elf.c.o 00:01:53.497 [213/740] Generating lib/rte_cfgfile_def with a custom command 00:01:53.758 [214/740] Generating lib/rte_cfgfile_mingw with a custom command 00:01:53.758 [215/740] Compiling C object lib/librte_cfgfile.a.p/cfgfile_rte_cfgfile.c.o 00:01:53.758 [216/740] Compiling C object lib/librte_bpf.a.p/bpf_bpf_convert.c.o 00:01:53.758 [217/740] Linking static target lib/librte_cfgfile.a 00:01:53.758 [218/740] Generating lib/hash.sym_chk with a custom command (wrapped by meson to capture output) 00:01:53.758 [219/740] Compiling C object lib/librte_bpf.a.p/bpf_bpf_validate.c.o 00:01:53.758 [220/740] Generating lib/rte_compressdev_def with a custom command 00:01:53.758 [221/740] Generating lib/rte_compressdev_mingw with a custom command 00:01:53.758 [222/740] Compiling C object lib/librte_acl.a.p/acl_acl_run_avx2.c.o 00:01:54.018 [223/740] Generating lib/cfgfile.sym_chk with a custom command (wrapped by meson to capture output) 00:01:54.018 [224/740] Compiling C object lib/librte_bpf.a.p/bpf_bpf_jit_x86.c.o 00:01:54.018 [225/740] Linking static target lib/librte_bpf.a 00:01:54.018 [226/740] Compiling C object lib/librte_compressdev.a.p/compressdev_rte_compressdev_pmd.c.o 00:01:54.018 [227/740] Generating lib/rte_cryptodev_def with a custom command 00:01:54.018 [228/740] Generating lib/rte_cryptodev_mingw with a custom command 00:01:54.018 [229/740] Compiling C object lib/librte_acl.a.p/acl_acl_run_avx512.c.o 00:01:54.018 [230/740] Compiling C object lib/librte_compressdev.a.p/compressdev_rte_comp.c.o 00:01:54.018 [231/740] Compiling C object lib/librte_compressdev.a.p/compressdev_rte_compressdev.c.o 00:01:54.018 [232/740] Compiling C object lib/librte_cryptodev.a.p/cryptodev_cryptodev_pmd.c.o 00:01:54.018 [233/740] Linking static target lib/librte_acl.a 00:01:54.018 [234/740] Linking static target lib/librte_compressdev.a 00:01:54.277 [235/740] Generating lib/bpf.sym_chk with a custom command (wrapped by meson to capture output) 00:01:54.277 [236/740] Generating lib/rte_distributor_def with a custom command 00:01:54.277 [237/740] Generating lib/rte_distributor_mingw with a custom command 00:01:54.277 [238/740] Compiling C object lib/librte_cryptodev.a.p/cryptodev_cryptodev_trace_points.c.o 00:01:54.277 [239/740] Generating lib/rte_efd_def with a custom command 00:01:54.277 [240/740] Generating lib/rte_efd_mingw with a custom command 00:01:54.277 [241/740] Generating lib/acl.sym_chk with a custom command (wrapped by meson to capture output) 00:01:54.536 [242/740] Compiling C object lib/librte_distributor.a.p/distributor_rte_distributor_match_sse.c.o 00:01:54.536 [243/740] Compiling C object lib/librte_distributor.a.p/distributor_rte_distributor_single.c.o 00:01:54.797 [244/740] Compiling C object lib/librte_eventdev.a.p/eventdev_eventdev_private.c.o 00:01:54.797 [245/740] Generating lib/eal.sym_chk with a custom command (wrapped by meson to capture output) 00:01:54.797 [246/740] Compiling C object lib/librte_distributor.a.p/distributor_rte_distributor.c.o 00:01:54.797 [247/740] Linking static target lib/librte_distributor.a 00:01:54.797 [248/740] Linking target lib/librte_eal.so.23.0 00:01:54.797 [249/740] Generating lib/compressdev.sym_chk with a custom command (wrapped by meson to capture output) 00:01:54.797 [250/740] Compiling C object lib/librte_eventdev.a.p/eventdev_eventdev_trace_points.c.o 00:01:54.797 [251/740] Generating symbol file lib/librte_eal.so.23.0.p/librte_eal.so.23.0.symbols 00:01:55.057 [252/740] Linking target lib/librte_ring.so.23.0 00:01:55.057 [253/740] Generating lib/distributor.sym_chk with a custom command (wrapped by meson to capture output) 00:01:55.057 [254/740] Linking target lib/librte_meter.so.23.0 00:01:55.057 [255/740] Generating symbol file lib/librte_ring.so.23.0.p/librte_ring.so.23.0.symbols 00:01:55.057 [256/740] Linking target lib/librte_rcu.so.23.0 00:01:55.057 [257/740] Generating symbol file lib/librte_meter.so.23.0.p/librte_meter.so.23.0.symbols 00:01:55.057 [258/740] Compiling C object lib/librte_eventdev.a.p/eventdev_rte_event_ring.c.o 00:01:55.057 [259/740] Linking target lib/librte_pci.so.23.0 00:01:55.057 [260/740] Linking target lib/librte_mempool.so.23.0 00:01:55.057 [261/740] Generating symbol file lib/librte_rcu.so.23.0.p/librte_rcu.so.23.0.symbols 00:01:55.315 [262/740] Linking target lib/librte_timer.so.23.0 00:01:55.315 [263/740] Generating symbol file lib/librte_pci.so.23.0.p/librte_pci.so.23.0.symbols 00:01:55.315 [264/740] Generating symbol file lib/librte_mempool.so.23.0.p/librte_mempool.so.23.0.symbols 00:01:55.315 [265/740] Linking target lib/librte_mbuf.so.23.0 00:01:55.315 [266/740] Linking target lib/librte_acl.so.23.0 00:01:55.315 [267/740] Generating symbol file lib/librte_timer.so.23.0.p/librte_timer.so.23.0.symbols 00:01:55.315 [268/740] Compiling C object lib/librte_efd.a.p/efd_rte_efd.c.o 00:01:55.315 [269/740] Linking static target lib/librte_efd.a 00:01:55.315 [270/740] Linking target lib/librte_cfgfile.so.23.0 00:01:55.315 [271/740] Generating symbol file lib/librte_acl.so.23.0.p/librte_acl.so.23.0.symbols 00:01:55.315 [272/740] Generating symbol file lib/librte_mbuf.so.23.0.p/librte_mbuf.so.23.0.symbols 00:01:55.315 [273/740] Generating lib/rte_eventdev_def with a custom command 00:01:55.315 [274/740] Linking target lib/librte_bbdev.so.23.0 00:01:55.315 [275/740] Linking target lib/librte_net.so.23.0 00:01:55.574 [276/740] Compiling C object lib/librte_eventdev.a.p/eventdev_rte_event_crypto_adapter.c.o 00:01:55.574 [277/740] Linking target lib/librte_compressdev.so.23.0 00:01:55.574 [278/740] Linking target lib/librte_distributor.so.23.0 00:01:55.574 [279/740] Generating lib/rte_eventdev_mingw with a custom command 00:01:55.574 [280/740] Generating lib/rte_gpudev_def with a custom command 00:01:55.574 [281/740] Generating lib/efd.sym_chk with a custom command (wrapped by meson to capture output) 00:01:55.574 [282/740] Generating lib/rte_gpudev_mingw with a custom command 00:01:55.574 [283/740] Generating symbol file lib/librte_net.so.23.0.p/librte_net.so.23.0.symbols 00:01:55.574 [284/740] Generating lib/ethdev.sym_chk with a custom command (wrapped by meson to capture output) 00:01:55.574 [285/740] Linking target lib/librte_cmdline.so.23.0 00:01:55.574 [286/740] Linking target lib/librte_hash.so.23.0 00:01:55.574 [287/740] Linking target lib/librte_ethdev.so.23.0 00:01:55.575 [288/740] Compiling C object lib/librte_eventdev.a.p/eventdev_rte_event_eth_tx_adapter.c.o 00:01:55.834 [289/740] Generating symbol file lib/librte_hash.so.23.0.p/librte_hash.so.23.0.symbols 00:01:55.834 [290/740] Generating symbol file lib/librte_ethdev.so.23.0.p/librte_ethdev.so.23.0.symbols 00:01:55.834 [291/740] Linking target lib/librte_efd.so.23.0 00:01:55.834 [292/740] Compiling C object lib/librte_eventdev.a.p/eventdev_rte_event_timer_adapter.c.o 00:01:55.834 [293/740] Linking target lib/librte_metrics.so.23.0 00:01:55.834 [294/740] Linking target lib/librte_bpf.so.23.0 00:01:55.834 [295/740] Compiling C object lib/librte_cryptodev.a.p/cryptodev_rte_cryptodev.c.o 00:01:55.834 [296/740] Linking static target lib/librte_cryptodev.a 00:01:55.834 [297/740] Generating symbol file lib/librte_metrics.so.23.0.p/librte_metrics.so.23.0.symbols 00:01:55.834 [298/740] Generating symbol file lib/librte_bpf.so.23.0.p/librte_bpf.so.23.0.symbols 00:01:55.834 [299/740] Linking target lib/librte_bitratestats.so.23.0 00:01:55.834 [300/740] Generating lib/rte_gro_def with a custom command 00:01:56.095 [301/740] Generating lib/rte_gro_mingw with a custom command 00:01:56.095 [302/740] Compiling C object lib/librte_gro.a.p/gro_gro_tcp4.c.o 00:01:56.095 [303/740] Compiling C object lib/librte_gro.a.p/gro_rte_gro.c.o 00:01:56.095 [304/740] Compiling C object lib/librte_gpudev.a.p/gpudev_gpudev.c.o 00:01:56.095 [305/740] Linking static target lib/librte_gpudev.a 00:01:56.095 [306/740] Compiling C object lib/librte_gro.a.p/gro_gro_udp4.c.o 00:01:56.354 [307/740] Compiling C object lib/librte_gro.a.p/gro_gro_vxlan_tcp4.c.o 00:01:56.354 [308/740] Compiling C object lib/librte_gso.a.p/gso_gso_tcp4.c.o 00:01:56.354 [309/740] Compiling C object lib/librte_gso.a.p/gso_gso_udp4.c.o 00:01:56.354 [310/740] Compiling C object lib/librte_eventdev.a.p/eventdev_rte_eventdev.c.o 00:01:56.354 [311/740] Generating lib/rte_gso_def with a custom command 00:01:56.354 [312/740] Generating lib/rte_gso_mingw with a custom command 00:01:56.354 [313/740] Compiling C object lib/librte_gro.a.p/gro_gro_vxlan_udp4.c.o 00:01:56.354 [314/740] Linking static target lib/librte_gro.a 00:01:56.613 [315/740] Compiling C object lib/librte_gso.a.p/gso_gso_tunnel_tcp4.c.o 00:01:56.613 [316/740] Generating lib/gro.sym_chk with a custom command (wrapped by meson to capture output) 00:01:56.613 [317/740] Compiling C object lib/librte_gso.a.p/gso_gso_common.c.o 00:01:56.613 [318/740] Compiling C object lib/librte_eventdev.a.p/eventdev_rte_event_eth_rx_adapter.c.o 00:01:56.613 [319/740] Linking target lib/librte_gro.so.23.0 00:01:56.613 [320/740] Linking static target lib/librte_eventdev.a 00:01:56.613 [321/740] Compiling C object lib/librte_gso.a.p/gso_gso_tunnel_udp4.c.o 00:01:56.873 [322/740] Compiling C object lib/librte_gso.a.p/gso_rte_gso.c.o 00:01:56.873 [323/740] Linking static target lib/librte_gso.a 00:01:56.873 [324/740] Generating lib/gpudev.sym_chk with a custom command (wrapped by meson to capture output) 00:01:56.873 [325/740] Linking target lib/librte_gpudev.so.23.0 00:01:56.873 [326/740] Generating lib/rte_ip_frag_def with a custom command 00:01:56.873 [327/740] Compiling C object lib/librte_ip_frag.a.p/ip_frag_rte_ipv4_reassembly.c.o 00:01:56.873 [328/740] Generating lib/gso.sym_chk with a custom command (wrapped by meson to capture output) 00:01:56.873 [329/740] Generating lib/rte_ip_frag_mingw with a custom command 00:01:56.873 [330/740] Generating lib/rte_jobstats_def with a custom command 00:01:56.873 [331/740] Linking target lib/librte_gso.so.23.0 00:01:56.873 [332/740] Generating lib/rte_jobstats_mingw with a custom command 00:01:56.873 [333/740] Compiling C object lib/librte_ip_frag.a.p/ip_frag_rte_ipv6_fragmentation.c.o 00:01:56.873 [334/740] Generating lib/rte_latencystats_def with a custom command 00:01:56.873 [335/740] Generating lib/rte_latencystats_mingw with a custom command 00:01:56.873 [336/740] Compiling C object lib/librte_jobstats.a.p/jobstats_rte_jobstats.c.o 00:01:57.132 [337/740] Linking static target lib/librte_jobstats.a 00:01:57.133 [338/740] Compiling C object lib/librte_ip_frag.a.p/ip_frag_rte_ipv4_fragmentation.c.o 00:01:57.133 [339/740] Generating lib/rte_lpm_def with a custom command 00:01:57.133 [340/740] Compiling C object lib/librte_ip_frag.a.p/ip_frag_rte_ipv6_reassembly.c.o 00:01:57.133 [341/740] Generating lib/rte_lpm_mingw with a custom command 00:01:57.133 [342/740] Compiling C object lib/librte_ip_frag.a.p/ip_frag_rte_ip_frag_common.c.o 00:01:57.133 [343/740] Generating lib/jobstats.sym_chk with a custom command (wrapped by meson to capture output) 00:01:57.133 [344/740] Linking target lib/librte_jobstats.so.23.0 00:01:57.392 [345/740] Compiling C object lib/librte_ip_frag.a.p/ip_frag_ip_frag_internal.c.o 00:01:57.392 [346/740] Linking static target lib/librte_ip_frag.a 00:01:57.392 [347/740] Compiling C object lib/librte_member.a.p/member_rte_member.c.o 00:01:57.392 [348/740] Compiling C object lib/librte_latencystats.a.p/latencystats_rte_latencystats.c.o 00:01:57.392 [349/740] Linking static target lib/librte_latencystats.a 00:01:57.392 [350/740] Compiling C object lib/member/libsketch_avx512_tmp.a.p/rte_member_sketch_avx512.c.o 00:01:57.392 [351/740] Linking static target lib/member/libsketch_avx512_tmp.a 00:01:57.392 [352/740] Compiling C object lib/librte_lpm.a.p/lpm_rte_lpm.c.o 00:01:57.651 [353/740] Generating lib/rte_member_def with a custom command 00:01:57.651 [354/740] Generating lib/rte_member_mingw with a custom command 00:01:57.651 [355/740] Generating lib/rte_pcapng_def with a custom command 00:01:57.651 [356/740] Generating lib/ip_frag.sym_chk with a custom command (wrapped by meson to capture output) 00:01:57.651 [357/740] Generating lib/rte_pcapng_mingw with a custom command 00:01:57.652 [358/740] Linking target lib/librte_ip_frag.so.23.0 00:01:57.652 [359/740] Generating lib/cryptodev.sym_chk with a custom command (wrapped by meson to capture output) 00:01:57.652 [360/740] Generating lib/latencystats.sym_chk with a custom command (wrapped by meson to capture output) 00:01:57.652 [361/740] Compiling C object lib/librte_power.a.p/power_guest_channel.c.o 00:01:57.652 [362/740] Linking target lib/librte_cryptodev.so.23.0 00:01:57.652 [363/740] Linking target lib/librte_latencystats.so.23.0 00:01:57.652 [364/740] Generating symbol file lib/librte_ip_frag.so.23.0.p/librte_ip_frag.so.23.0.symbols 00:01:57.652 [365/740] Compiling C object lib/librte_power.a.p/power_power_common.c.o 00:01:57.652 [366/740] Generating symbol file lib/librte_cryptodev.so.23.0.p/librte_cryptodev.so.23.0.symbols 00:01:57.912 [367/740] Compiling C object lib/librte_lpm.a.p/lpm_rte_lpm6.c.o 00:01:57.912 [368/740] Linking static target lib/librte_lpm.a 00:01:57.912 [369/740] Compiling C object lib/librte_power.a.p/power_power_kvm_vm.c.o 00:01:57.912 [370/740] Compiling C object lib/librte_power.a.p/power_rte_power.c.o 00:01:58.172 [371/740] Compiling C object lib/librte_member.a.p/member_rte_member_vbf.c.o 00:01:58.172 [372/740] Compiling C object lib/librte_member.a.p/member_rte_member_ht.c.o 00:01:58.172 [373/740] Compiling C object lib/librte_power.a.p/power_power_acpi_cpufreq.c.o 00:01:58.172 [374/740] Generating lib/lpm.sym_chk with a custom command (wrapped by meson to capture output) 00:01:58.172 [375/740] Generating lib/rte_power_def with a custom command 00:01:58.172 [376/740] Compiling C object lib/librte_power.a.p/power_rte_power_empty_poll.c.o 00:01:58.172 [377/740] Compiling C object lib/librte_pcapng.a.p/pcapng_rte_pcapng.c.o 00:01:58.172 [378/740] Linking static target lib/librte_pcapng.a 00:01:58.172 [379/740] Linking target lib/librte_lpm.so.23.0 00:01:58.172 [380/740] Generating lib/rte_power_mingw with a custom command 00:01:58.172 [381/740] Compiling C object lib/librte_power.a.p/power_power_cppc_cpufreq.c.o 00:01:58.172 [382/740] Generating lib/rte_rawdev_def with a custom command 00:01:58.172 [383/740] Generating lib/rte_rawdev_mingw with a custom command 00:01:58.172 [384/740] Generating lib/rte_regexdev_def with a custom command 00:01:58.172 [385/740] Generating lib/rte_regexdev_mingw with a custom command 00:01:58.172 [386/740] Generating lib/eventdev.sym_chk with a custom command (wrapped by meson to capture output) 00:01:58.172 [387/740] Generating symbol file lib/librte_lpm.so.23.0.p/librte_lpm.so.23.0.symbols 00:01:58.172 [388/740] Generating lib/rte_dmadev_def with a custom command 00:01:58.433 [389/740] Linking target lib/librte_eventdev.so.23.0 00:01:58.433 [390/740] Generating lib/rte_dmadev_mingw with a custom command 00:01:58.433 [391/740] Compiling C object lib/librte_power.a.p/power_power_pstate_cpufreq.c.o 00:01:58.433 [392/740] Generating lib/pcapng.sym_chk with a custom command (wrapped by meson to capture output) 00:01:58.433 [393/740] Linking target lib/librte_pcapng.so.23.0 00:01:58.433 [394/740] Generating symbol file lib/librte_eventdev.so.23.0.p/librte_eventdev.so.23.0.symbols 00:01:58.433 [395/740] Compiling C object lib/librte_power.a.p/power_rte_power_intel_uncore.c.o 00:01:58.433 [396/740] Compiling C object lib/librte_rawdev.a.p/rawdev_rte_rawdev.c.o 00:01:58.433 [397/740] Linking static target lib/librte_rawdev.a 00:01:58.433 [398/740] Generating lib/rte_rib_def with a custom command 00:01:58.433 [399/740] Generating lib/rte_rib_mingw with a custom command 00:01:58.433 [400/740] Generating lib/rte_reorder_def with a custom command 00:01:58.433 [401/740] Generating lib/rte_reorder_mingw with a custom command 00:01:58.433 [402/740] Generating symbol file lib/librte_pcapng.so.23.0.p/librte_pcapng.so.23.0.symbols 00:01:58.433 [403/740] Compiling C object lib/librte_dmadev.a.p/dmadev_rte_dmadev.c.o 00:01:58.433 [404/740] Linking static target lib/librte_dmadev.a 00:01:58.433 [405/740] Compiling C object lib/librte_power.a.p/power_rte_power_pmd_mgmt.c.o 00:01:58.433 [406/740] Linking static target lib/librte_power.a 00:01:58.700 [407/740] Compiling C object lib/librte_sched.a.p/sched_rte_red.c.o 00:01:58.700 [408/740] Compiling C object lib/librte_regexdev.a.p/regexdev_rte_regexdev.c.o 00:01:58.700 [409/740] Linking static target lib/librte_regexdev.a 00:01:58.700 [410/740] Compiling C object lib/librte_sched.a.p/sched_rte_approx.c.o 00:01:58.700 [411/740] Compiling C object lib/librte_sched.a.p/sched_rte_pie.c.o 00:01:58.700 [412/740] Generating lib/rte_sched_def with a custom command 00:01:58.700 [413/740] Generating lib/rawdev.sym_chk with a custom command (wrapped by meson to capture output) 00:01:58.700 [414/740] Generating lib/rte_sched_mingw with a custom command 00:01:58.700 [415/740] Compiling C object lib/librte_rib.a.p/rib_rte_rib.c.o 00:01:58.959 [416/740] Generating lib/rte_security_def with a custom command 00:01:58.959 [417/740] Compiling C object lib/librte_member.a.p/member_rte_member_sketch.c.o 00:01:58.959 [418/740] Linking target lib/librte_rawdev.so.23.0 00:01:58.959 [419/740] Linking static target lib/librte_member.a 00:01:58.959 [420/740] Generating lib/rte_security_mingw with a custom command 00:01:58.959 [421/740] Compiling C object lib/librte_stack.a.p/stack_rte_stack_std.c.o 00:01:58.959 [422/740] Compiling C object lib/librte_stack.a.p/stack_rte_stack.c.o 00:01:58.959 [423/740] Generating lib/rte_stack_def with a custom command 00:01:58.959 [424/740] Compiling C object lib/librte_reorder.a.p/reorder_rte_reorder.c.o 00:01:58.959 [425/740] Generating lib/dmadev.sym_chk with a custom command (wrapped by meson to capture output) 00:01:58.959 [426/740] Linking static target lib/librte_reorder.a 00:01:58.959 [427/740] Generating lib/rte_stack_mingw with a custom command 00:01:58.959 [428/740] Compiling C object lib/librte_stack.a.p/stack_rte_stack_lf.c.o 00:01:58.959 [429/740] Linking target lib/librte_dmadev.so.23.0 00:01:58.959 [430/740] Linking static target lib/librte_stack.a 00:01:58.959 [431/740] Compiling C object lib/librte_rib.a.p/rib_rte_rib6.c.o 00:01:58.959 [432/740] Linking static target lib/librte_rib.a 00:01:59.219 [433/740] Compiling C object lib/librte_vhost.a.p/vhost_fd_man.c.o 00:01:59.219 [434/740] Generating symbol file lib/librte_dmadev.so.23.0.p/librte_dmadev.so.23.0.symbols 00:01:59.219 [435/740] Generating lib/member.sym_chk with a custom command (wrapped by meson to capture output) 00:01:59.219 [436/740] Generating lib/reorder.sym_chk with a custom command (wrapped by meson to capture output) 00:01:59.219 [437/740] Linking target lib/librte_member.so.23.0 00:01:59.219 [438/740] Linking target lib/librte_reorder.so.23.0 00:01:59.219 [439/740] Generating lib/stack.sym_chk with a custom command (wrapped by meson to capture output) 00:01:59.219 [440/740] Linking target lib/librte_stack.so.23.0 00:01:59.219 [441/740] Generating lib/regexdev.sym_chk with a custom command (wrapped by meson to capture output) 00:01:59.219 [442/740] Compiling C object lib/librte_security.a.p/security_rte_security.c.o 00:01:59.219 [443/740] Generating lib/power.sym_chk with a custom command (wrapped by meson to capture output) 00:01:59.478 [444/740] Linking target lib/librte_regexdev.so.23.0 00:01:59.478 [445/740] Linking static target lib/librte_security.a 00:01:59.478 [446/740] Linking target lib/librte_power.so.23.0 00:01:59.478 [447/740] Generating lib/rib.sym_chk with a custom command (wrapped by meson to capture output) 00:01:59.478 [448/740] Linking target lib/librte_rib.so.23.0 00:01:59.478 [449/740] Generating symbol file lib/librte_rib.so.23.0.p/librte_rib.so.23.0.symbols 00:01:59.478 [450/740] Compiling C object lib/librte_vhost.a.p/vhost_iotlb.c.o 00:01:59.478 [451/740] Generating lib/rte_vhost_def with a custom command 00:01:59.478 [452/740] Generating lib/rte_vhost_mingw with a custom command 00:01:59.737 [453/740] Compiling C object lib/librte_vhost.a.p/vhost_vdpa.c.o 00:01:59.737 [454/740] Compiling C object lib/librte_vhost.a.p/vhost_socket.c.o 00:01:59.737 [455/740] Generating lib/security.sym_chk with a custom command (wrapped by meson to capture output) 00:01:59.737 [456/740] Linking target lib/librte_security.so.23.0 00:01:59.737 [457/740] Compiling C object lib/librte_sched.a.p/sched_rte_sched.c.o 00:01:59.737 [458/740] Linking static target lib/librte_sched.a 00:01:59.737 [459/740] Generating symbol file lib/librte_security.so.23.0.p/librte_security.so.23.0.symbols 00:02:00.303 [460/740] Generating lib/sched.sym_chk with a custom command (wrapped by meson to capture output) 00:02:00.303 [461/740] Compiling C object lib/librte_ipsec.a.p/ipsec_ses.c.o 00:02:00.303 [462/740] Linking target lib/librte_sched.so.23.0 00:02:00.303 [463/740] Compiling C object lib/librte_ipsec.a.p/ipsec_sa.c.o 00:02:00.303 [464/740] Generating lib/rte_ipsec_def with a custom command 00:02:00.303 [465/740] Generating lib/rte_ipsec_mingw with a custom command 00:02:00.303 [466/740] Generating symbol file lib/librte_sched.so.23.0.p/librte_sched.so.23.0.symbols 00:02:00.303 [467/740] Compiling C object lib/librte_fib.a.p/fib_rte_fib.c.o 00:02:00.303 [468/740] Compiling C object lib/librte_vhost.a.p/vhost_vhost.c.o 00:02:00.303 [469/740] Compiling C object lib/librte_vhost.a.p/vhost_vhost_user.c.o 00:02:00.562 [470/740] Compiling C object lib/librte_ipsec.a.p/ipsec_ipsec_sad.c.o 00:02:00.562 [471/740] Compiling C object lib/librte_ipsec.a.p/ipsec_ipsec_telemetry.c.o 00:02:00.562 [472/740] Generating lib/rte_fib_def with a custom command 00:02:00.562 [473/740] Generating lib/rte_fib_mingw with a custom command 00:02:00.562 [474/740] Compiling C object lib/librte_fib.a.p/fib_rte_fib6.c.o 00:02:00.822 [475/740] Compiling C object lib/librte_fib.a.p/fib_dir24_8_avx512.c.o 00:02:00.822 [476/740] Compiling C object lib/librte_ipsec.a.p/ipsec_esp_outb.c.o 00:02:00.822 [477/740] Compiling C object lib/librte_fib.a.p/fib_trie_avx512.c.o 00:02:01.081 [478/740] Compiling C object lib/librte_fib.a.p/fib_trie.c.o 00:02:01.081 [479/740] Compiling C object lib/librte_port.a.p/port_rte_port_ethdev.c.o 00:02:01.340 [480/740] Compiling C object lib/librte_ipsec.a.p/ipsec_esp_inb.c.o 00:02:01.340 [481/740] Linking static target lib/librte_ipsec.a 00:02:01.340 [482/740] Compiling C object lib/librte_fib.a.p/fib_dir24_8.c.o 00:02:01.340 [483/740] Linking static target lib/librte_fib.a 00:02:01.340 [484/740] Compiling C object lib/librte_port.a.p/port_rte_port_frag.c.o 00:02:01.340 [485/740] Compiling C object lib/librte_port.a.p/port_rte_port_fd.c.o 00:02:01.340 [486/740] Compiling C object lib/librte_port.a.p/port_rte_port_ras.c.o 00:02:01.340 [487/740] Compiling C object lib/librte_port.a.p/port_rte_port_sched.c.o 00:02:01.599 [488/740] Generating lib/ipsec.sym_chk with a custom command (wrapped by meson to capture output) 00:02:01.599 [489/740] Generating lib/fib.sym_chk with a custom command (wrapped by meson to capture output) 00:02:01.599 [490/740] Linking target lib/librte_fib.so.23.0 00:02:01.599 [491/740] Linking target lib/librte_ipsec.so.23.0 00:02:01.858 [492/740] Compiling C object lib/librte_port.a.p/port_rte_port_source_sink.c.o 00:02:01.858 [493/740] Generating lib/rte_port_def with a custom command 00:02:01.858 [494/740] Generating lib/rte_port_mingw with a custom command 00:02:01.858 [495/740] Compiling C object lib/librte_port.a.p/port_rte_port_sym_crypto.c.o 00:02:01.858 [496/740] Generating lib/rte_pdump_def with a custom command 00:02:02.117 [497/740] Generating lib/rte_pdump_mingw with a custom command 00:02:02.117 [498/740] Compiling C object lib/librte_port.a.p/port_rte_swx_port_fd.c.o 00:02:02.117 [499/740] Compiling C object lib/librte_port.a.p/port_rte_swx_port_ethdev.c.o 00:02:02.117 [500/740] Compiling C object lib/librte_port.a.p/port_rte_port_eventdev.c.o 00:02:02.117 [501/740] Compiling C object lib/librte_table.a.p/table_rte_swx_keycmp.c.o 00:02:02.375 [502/740] Compiling C object lib/librte_table.a.p/table_rte_swx_table_em.c.o 00:02:02.375 [503/740] Compiling C object lib/librte_table.a.p/table_rte_swx_table_learner.c.o 00:02:02.375 [504/740] Compiling C object lib/librte_port.a.p/port_rte_swx_port_source_sink.c.o 00:02:02.375 [505/740] Compiling C object lib/librte_port.a.p/port_rte_swx_port_ring.c.o 00:02:02.375 [506/740] Compiling C object lib/librte_port.a.p/port_rte_port_ring.c.o 00:02:02.375 [507/740] Linking static target lib/librte_port.a 00:02:02.633 [508/740] Compiling C object lib/librte_table.a.p/table_rte_swx_table_wm.c.o 00:02:02.633 [509/740] Compiling C object lib/librte_table.a.p/table_rte_swx_table_selector.c.o 00:02:02.633 [510/740] Compiling C object lib/librte_table.a.p/table_rte_table_array.c.o 00:02:02.633 [511/740] Compiling C object lib/librte_pdump.a.p/pdump_rte_pdump.c.o 00:02:02.633 [512/740] Compiling C object lib/librte_table.a.p/table_rte_table_hash_cuckoo.c.o 00:02:02.633 [513/740] Linking static target lib/librte_pdump.a 00:02:02.633 [514/740] Compiling C object lib/librte_table.a.p/table_rte_table_acl.c.o 00:02:02.891 [515/740] Generating lib/port.sym_chk with a custom command (wrapped by meson to capture output) 00:02:02.892 [516/740] Linking target lib/librte_port.so.23.0 00:02:02.892 [517/740] Generating symbol file lib/librte_port.so.23.0.p/librte_port.so.23.0.symbols 00:02:03.151 [518/740] Generating lib/pdump.sym_chk with a custom command (wrapped by meson to capture output) 00:02:03.151 [519/740] Linking target lib/librte_pdump.so.23.0 00:02:03.151 [520/740] Compiling C object lib/librte_table.a.p/table_rte_table_hash_ext.c.o 00:02:03.151 [521/740] Compiling C object lib/librte_table.a.p/table_rte_table_lpm.c.o 00:02:03.151 [522/740] Generating lib/rte_table_def with a custom command 00:02:03.151 [523/740] Generating lib/rte_table_mingw with a custom command 00:02:03.151 [524/740] Compiling C object lib/librte_table.a.p/table_rte_table_hash_key8.c.o 00:02:03.409 [525/740] Compiling C object lib/librte_table.a.p/table_rte_table_lpm_ipv6.c.o 00:02:03.409 [526/740] Compiling C object lib/librte_table.a.p/table_rte_table_stub.c.o 00:02:03.409 [527/740] Compiling C object lib/librte_table.a.p/table_rte_table_hash_key16.c.o 00:02:03.668 [528/740] Compiling C object lib/librte_table.a.p/table_rte_table_hash_key32.c.o 00:02:03.668 [529/740] Generating lib/rte_pipeline_def with a custom command 00:02:03.668 [530/740] Compiling C object lib/librte_table.a.p/table_rte_table_hash_lru.c.o 00:02:03.668 [531/740] Generating lib/rte_pipeline_mingw with a custom command 00:02:03.668 [532/740] Linking static target lib/librte_table.a 00:02:03.668 [533/740] Compiling C object lib/librte_pipeline.a.p/pipeline_rte_port_in_action.c.o 00:02:03.945 [534/740] Compiling C object lib/librte_pipeline.a.p/pipeline_rte_pipeline.c.o 00:02:03.945 [535/740] Compiling C object lib/librte_graph.a.p/graph_node.c.o 00:02:03.945 [536/740] Generating lib/table.sym_chk with a custom command (wrapped by meson to capture output) 00:02:04.203 [537/740] Linking target lib/librte_table.so.23.0 00:02:04.203 [538/740] Compiling C object lib/librte_graph.a.p/graph_graph.c.o 00:02:04.203 [539/740] Compiling C object lib/librte_graph.a.p/graph_graph_ops.c.o 00:02:04.203 [540/740] Generating symbol file lib/librte_table.so.23.0.p/librte_table.so.23.0.symbols 00:02:04.203 [541/740] Generating lib/rte_graph_def with a custom command 00:02:04.203 [542/740] Generating lib/rte_graph_mingw with a custom command 00:02:04.203 [543/740] Compiling C object lib/librte_graph.a.p/graph_graph_debug.c.o 00:02:04.460 [544/740] Compiling C object lib/librte_pipeline.a.p/pipeline_rte_swx_ctl.c.o 00:02:04.460 [545/740] Compiling C object lib/librte_graph.a.p/graph_graph_populate.c.o 00:02:04.861 [546/740] Compiling C object lib/librte_node.a.p/node_ethdev_ctrl.c.o 00:02:04.861 [547/740] Compiling C object lib/librte_graph.a.p/graph_graph_stats.c.o 00:02:04.861 [548/740] Linking static target lib/librte_graph.a 00:02:04.861 [549/740] Compiling C object lib/librte_vhost.a.p/vhost_vhost_crypto.c.o 00:02:04.861 [550/740] Compiling C object lib/librte_node.a.p/node_ethdev_rx.c.o 00:02:04.861 [551/740] Compiling C object lib/librte_node.a.p/node_ethdev_tx.c.o 00:02:04.861 [552/740] Compiling C object lib/librte_node.a.p/node_null.c.o 00:02:05.147 [553/740] Compiling C object lib/librte_node.a.p/node_log.c.o 00:02:05.147 [554/740] Generating lib/rte_node_def with a custom command 00:02:05.147 [555/740] Generating lib/rte_node_mingw with a custom command 00:02:05.147 [556/740] Compiling C object lib/librte_node.a.p/node_pkt_drop.c.o 00:02:05.147 [557/740] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_pci_params.c.o 00:02:05.408 [558/740] Generating lib/graph.sym_chk with a custom command (wrapped by meson to capture output) 00:02:05.408 [559/740] Compiling C object lib/librte_pipeline.a.p/pipeline_rte_swx_pipeline_spec.c.o 00:02:05.408 [560/740] Linking target lib/librte_graph.so.23.0 00:02:05.408 [561/740] Compiling C object lib/librte_node.a.p/node_ip4_lookup.c.o 00:02:05.408 [562/740] Compiling C object lib/librte_node.a.p/node_pkt_cls.c.o 00:02:05.408 [563/740] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_pci_common_uio.c.o 00:02:05.408 [564/740] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_pci_common.c.o 00:02:05.408 [565/740] Generating drivers/rte_bus_pci_def with a custom command 00:02:05.408 [566/740] Generating symbol file lib/librte_graph.so.23.0.p/librte_graph.so.23.0.symbols 00:02:05.408 [567/740] Generating drivers/rte_bus_pci_mingw with a custom command 00:02:05.408 [568/740] Generating drivers/rte_bus_vdev_def with a custom command 00:02:05.408 [569/740] Compiling C object drivers/libtmp_rte_bus_vdev.a.p/bus_vdev_vdev_params.c.o 00:02:05.408 [570/740] Generating drivers/rte_bus_vdev_mingw with a custom command 00:02:05.408 [571/740] Generating drivers/rte_mempool_ring_def with a custom command 00:02:05.408 [572/740] Generating drivers/rte_mempool_ring_mingw with a custom command 00:02:05.408 [573/740] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_linux_pci_uio.c.o 00:02:05.667 [574/740] Compiling C object lib/librte_node.a.p/node_ip4_rewrite.c.o 00:02:05.667 [575/740] Linking static target lib/librte_node.a 00:02:05.667 [576/740] Compiling C object drivers/libtmp_rte_bus_vdev.a.p/bus_vdev_vdev.c.o 00:02:05.667 [577/740] Linking static target drivers/libtmp_rte_bus_vdev.a 00:02:05.667 [578/740] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_linux_pci_vfio.c.o 00:02:05.667 [579/740] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_linux_pci.c.o 00:02:05.667 [580/740] Linking static target drivers/libtmp_rte_bus_pci.a 00:02:05.667 [581/740] Generating lib/node.sym_chk with a custom command (wrapped by meson to capture output) 00:02:05.927 [582/740] Linking target lib/librte_node.so.23.0 00:02:05.927 [583/740] Generating drivers/rte_bus_vdev.pmd.c with a custom command 00:02:05.927 [584/740] Compiling C object drivers/librte_bus_vdev.a.p/meson-generated_.._rte_bus_vdev.pmd.c.o 00:02:05.927 [585/740] Linking static target drivers/librte_bus_vdev.a 00:02:05.927 [586/740] Generating drivers/rte_bus_pci.pmd.c with a custom command 00:02:05.927 [587/740] Compiling C object drivers/librte_bus_pci.a.p/meson-generated_.._rte_bus_pci.pmd.c.o 00:02:05.927 [588/740] Linking static target drivers/librte_bus_pci.a 00:02:05.927 [589/740] Generating drivers/rte_bus_vdev.sym_chk with a custom command (wrapped by meson to capture output) 00:02:06.185 [590/740] Compiling C object drivers/librte_bus_pci.so.23.0.p/meson-generated_.._rte_bus_pci.pmd.c.o 00:02:06.185 [591/740] Compiling C object drivers/librte_bus_vdev.so.23.0.p/meson-generated_.._rte_bus_vdev.pmd.c.o 00:02:06.185 [592/740] Linking target drivers/librte_bus_vdev.so.23.0 00:02:06.185 [593/740] Compiling C object drivers/net/i40e/base/libi40e_base.a.p/i40e_diag.c.o 00:02:06.185 [594/740] Compiling C object drivers/net/i40e/base/libi40e_base.a.p/i40e_adminq.c.o 00:02:06.185 [595/740] Generating symbol file drivers/librte_bus_vdev.so.23.0.p/librte_bus_vdev.so.23.0.symbols 00:02:06.185 [596/740] Generating drivers/rte_bus_pci.sym_chk with a custom command (wrapped by meson to capture output) 00:02:06.185 [597/740] Linking target drivers/librte_bus_pci.so.23.0 00:02:06.185 [598/740] Compiling C object drivers/net/i40e/base/libi40e_base.a.p/i40e_dcb.c.o 00:02:06.443 [599/740] Compiling C object drivers/libtmp_rte_mempool_ring.a.p/mempool_ring_rte_mempool_ring.c.o 00:02:06.443 [600/740] Linking static target drivers/libtmp_rte_mempool_ring.a 00:02:06.443 [601/740] Generating symbol file drivers/librte_bus_pci.so.23.0.p/librte_bus_pci.so.23.0.symbols 00:02:06.443 [602/740] Generating drivers/rte_mempool_ring.pmd.c with a custom command 00:02:06.443 [603/740] Compiling C object drivers/net/i40e/base/libi40e_base.a.p/i40e_hmc.c.o 00:02:06.443 [604/740] Compiling C object drivers/librte_mempool_ring.a.p/meson-generated_.._rte_mempool_ring.pmd.c.o 00:02:06.443 [605/740] Linking static target drivers/librte_mempool_ring.a 00:02:06.443 [606/740] Compiling C object drivers/librte_mempool_ring.so.23.0.p/meson-generated_.._rte_mempool_ring.pmd.c.o 00:02:06.702 [607/740] Linking target drivers/librte_mempool_ring.so.23.0 00:02:06.702 [608/740] Compiling C object drivers/net/i40e/base/libi40e_base.a.p/i40e_lan_hmc.c.o 00:02:06.960 [609/740] Compiling C object drivers/net/i40e/base/libi40e_base.a.p/i40e_nvm.c.o 00:02:07.219 [610/740] Compiling C object drivers/net/i40e/base/libi40e_base.a.p/i40e_common.c.o 00:02:07.219 [611/740] Linking static target drivers/net/i40e/base/libi40e_base.a 00:02:07.478 [612/740] Compiling C object drivers/libtmp_rte_net_i40e.a.p/net_i40e_i40e_pf.c.o 00:02:07.737 [613/740] Compiling C object drivers/net/i40e/libi40e_avx512_lib.a.p/i40e_rxtx_vec_avx512.c.o 00:02:07.737 [614/740] Linking static target drivers/net/i40e/libi40e_avx512_lib.a 00:02:07.996 [615/740] Compiling C object drivers/libtmp_rte_net_i40e.a.p/net_i40e_i40e_fdir.c.o 00:02:07.996 [616/740] Compiling C object drivers/libtmp_rte_net_i40e.a.p/net_i40e_i40e_tm.c.o 00:02:07.996 [617/740] Compiling C object drivers/libtmp_rte_net_i40e.a.p/net_i40e_i40e_hash.c.o 00:02:08.255 [618/740] Compiling C object drivers/libtmp_rte_net_i40e.a.p/net_i40e_i40e_flow.c.o 00:02:08.255 [619/740] Generating drivers/rte_net_i40e_def with a custom command 00:02:08.255 [620/740] Compiling C object drivers/libtmp_rte_net_i40e.a.p/net_i40e_i40e_vf_representor.c.o 00:02:08.255 [621/740] Generating drivers/rte_net_i40e_mingw with a custom command 00:02:08.822 [622/740] Compiling C object app/dpdk-dumpcap.p/dumpcap_main.c.o 00:02:09.078 [623/740] Compiling C object drivers/libtmp_rte_net_i40e.a.p/net_i40e_i40e_rxtx_vec_sse.c.o 00:02:09.078 [624/740] Compiling C object lib/librte_pipeline.a.p/pipeline_rte_swx_pipeline.c.o 00:02:09.336 [625/740] Compiling C object drivers/libtmp_rte_net_i40e.a.p/net_i40e_rte_pmd_i40e.c.o 00:02:09.337 [626/740] Compiling C object app/dpdk-test-bbdev.p/test-bbdev_main.c.o 00:02:09.337 [627/740] Compiling C object app/dpdk-pdump.p/pdump_main.c.o 00:02:09.595 [628/740] Compiling C object drivers/libtmp_rte_net_i40e.a.p/net_i40e_i40e_rxtx_vec_avx2.c.o 00:02:09.595 [629/740] Compiling C object app/dpdk-test-cmdline.p/test-cmdline_commands.c.o 00:02:09.595 [630/740] Compiling C object app/dpdk-test-cmdline.p/test-cmdline_cmdline_test.c.o 00:02:09.595 [631/740] Compiling C object app/dpdk-proc-info.p/proc-info_main.c.o 00:02:09.596 [632/740] Compiling C object app/dpdk-test-acl.p/test-acl_main.c.o 00:02:09.596 [633/740] Compiling C object drivers/libtmp_rte_net_i40e.a.p/net_i40e_i40e_rxtx.c.o 00:02:10.163 [634/740] Compiling C object app/dpdk-test-compress-perf.p/test-compress-perf_comp_perf_options_parse.c.o 00:02:10.163 [635/740] Compiling C object drivers/libtmp_rte_net_i40e.a.p/net_i40e_i40e_ethdev.c.o 00:02:10.163 [636/740] Linking static target drivers/libtmp_rte_net_i40e.a 00:02:10.163 [637/740] Compiling C object app/dpdk-test-bbdev.p/test-bbdev_test_bbdev.c.o 00:02:10.163 [638/740] Compiling C object app/dpdk-test-bbdev.p/test-bbdev_test_bbdev_vector.c.o 00:02:10.422 [639/740] Compiling C object app/dpdk-test-compress-perf.p/test-compress-perf_comp_perf_test_throughput.c.o 00:02:10.422 [640/740] Generating drivers/rte_net_i40e.pmd.c with a custom command 00:02:10.422 [641/740] Compiling C object drivers/librte_net_i40e.a.p/meson-generated_.._rte_net_i40e.pmd.c.o 00:02:10.422 [642/740] Linking static target drivers/librte_net_i40e.a 00:02:10.422 [643/740] Compiling C object app/dpdk-test-compress-perf.p/test-compress-perf_comp_perf_test_common.c.o 00:02:10.422 [644/740] Compiling C object drivers/librte_net_i40e.so.23.0.p/meson-generated_.._rte_net_i40e.pmd.c.o 00:02:10.422 [645/740] Compiling C object app/dpdk-test-compress-perf.p/test-compress-perf_comp_perf_test_cyclecount.c.o 00:02:10.679 [646/740] Compiling C object app/dpdk-test-compress-perf.p/test-compress-perf_main.c.o 00:02:10.679 [647/740] Compiling C object app/dpdk-test-compress-perf.p/test-compress-perf_comp_perf_test_verify.c.o 00:02:10.937 [648/740] Compiling C object app/dpdk-test-crypto-perf.p/test-crypto-perf_cperf_ops.c.o 00:02:10.937 [649/740] Generating drivers/rte_net_i40e.sym_chk with a custom command (wrapped by meson to capture output) 00:02:10.937 [650/740] Compiling C object app/dpdk-test-crypto-perf.p/test-crypto-perf_cperf_test_common.c.o 00:02:10.937 [651/740] Linking target drivers/librte_net_i40e.so.23.0 00:02:10.937 [652/740] Compiling C object app/dpdk-test-crypto-perf.p/test-crypto-perf_cperf_options_parsing.c.o 00:02:11.199 [653/740] Compiling C object app/dpdk-test-crypto-perf.p/test-crypto-perf_cperf_test_latency.c.o 00:02:11.200 [654/740] Compiling C object app/dpdk-test-crypto-perf.p/test-crypto-perf_cperf_test_vectors.c.o 00:02:11.457 [655/740] Compiling C object app/dpdk-test-crypto-perf.p/test-crypto-perf_cperf_test_throughput.c.o 00:02:11.457 [656/740] Compiling C object app/dpdk-test-crypto-perf.p/test-crypto-perf_cperf_test_pmd_cyclecount.c.o 00:02:11.457 [657/740] Compiling C object app/dpdk-test-crypto-perf.p/test-crypto-perf_cperf_test_vector_parsing.c.o 00:02:11.457 [658/740] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_evt_test.c.o 00:02:11.457 [659/740] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_parser.c.o 00:02:11.457 [660/740] Compiling C object app/dpdk-test-crypto-perf.p/test-crypto-perf_cperf_test_verify.c.o 00:02:11.716 [661/740] Compiling C object app/dpdk-test-crypto-perf.p/test-crypto-perf_main.c.o 00:02:11.716 [662/740] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_evt_main.c.o 00:02:11.716 [663/740] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_evt_options.c.o 00:02:11.716 [664/740] Compiling C object lib/librte_vhost.a.p/vhost_virtio_net.c.o 00:02:11.975 [665/740] Linking static target lib/librte_vhost.a 00:02:11.975 [666/740] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_test_order_common.c.o 00:02:12.241 [667/740] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_test_order_atq.c.o 00:02:12.241 [668/740] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_test_order_queue.c.o 00:02:12.502 [669/740] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_test_perf_atq.c.o 00:02:12.502 [670/740] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_test_pipeline_atq.c.o 00:02:12.760 [671/740] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_test_perf_queue.c.o 00:02:12.760 [672/740] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_test_pipeline_common.c.o 00:02:12.760 [673/740] Generating lib/vhost.sym_chk with a custom command (wrapped by meson to capture output) 00:02:13.020 [674/740] Linking target lib/librte_vhost.so.23.0 00:02:13.020 [675/740] Compiling C object app/dpdk-test-flow-perf.p/test-flow-perf_actions_gen.c.o 00:02:13.020 [676/740] Compiling C object app/dpdk-test-flow-perf.p/test-flow-perf_flow_gen.c.o 00:02:13.020 [677/740] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_test_pipeline_queue.c.o 00:02:13.020 [678/740] Compiling C object app/dpdk-test-flow-perf.p/test-flow-perf_items_gen.c.o 00:02:13.279 [679/740] Compiling C object app/dpdk-test-fib.p/test-fib_main.c.o 00:02:13.279 [680/740] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_test_perf_common.c.o 00:02:13.279 [681/740] Compiling C object app/dpdk-test-gpudev.p/test-gpudev_main.c.o 00:02:13.279 [682/740] Compiling C object app/dpdk-test-pipeline.p/test-pipeline_config.c.o 00:02:13.279 [683/740] Compiling C object app/dpdk-test-pipeline.p/test-pipeline_init.c.o 00:02:13.538 [684/740] Compiling C object app/dpdk-test-pipeline.p/test-pipeline_main.c.o 00:02:13.538 [685/740] Compiling C object app/dpdk-test-pipeline.p/test-pipeline_pipeline_acl.c.o 00:02:13.796 [686/740] Compiling C object app/dpdk-test-pipeline.p/test-pipeline_pipeline_lpm.c.o 00:02:13.796 [687/740] Compiling C object app/dpdk-test-pipeline.p/test-pipeline_pipeline_lpm_ipv6.c.o 00:02:13.796 [688/740] Compiling C object app/dpdk-test-pipeline.p/test-pipeline_pipeline_stub.c.o 00:02:13.796 [689/740] Compiling C object app/dpdk-test-pipeline.p/test-pipeline_pipeline_hash.c.o 00:02:13.796 [690/740] Compiling C object app/dpdk-test-flow-perf.p/test-flow-perf_main.c.o 00:02:14.055 [691/740] Compiling C object app/dpdk-testpmd.p/test-pmd_5tswap.c.o 00:02:14.055 [692/740] Compiling C object app/dpdk-test-pipeline.p/test-pipeline_runtime.c.o 00:02:14.055 [693/740] Compiling C object app/dpdk-testpmd.p/test-pmd_cmd_flex_item.c.o 00:02:14.055 [694/740] Compiling C object app/dpdk-test-bbdev.p/test-bbdev_test_bbdev_perf.c.o 00:02:14.314 [695/740] Compiling C object app/dpdk-testpmd.p/test-pmd_cmdline_mtr.c.o 00:02:14.314 [696/740] Compiling C object app/dpdk-testpmd.p/test-pmd_cmdline_tm.c.o 00:02:14.900 [697/740] Compiling C object app/dpdk-testpmd.p/test-pmd_ieee1588fwd.c.o 00:02:14.900 [698/740] Compiling C object app/dpdk-testpmd.p/test-pmd_flowgen.c.o 00:02:14.900 [699/740] Compiling C object app/dpdk-testpmd.p/test-pmd_icmpecho.c.o 00:02:14.900 [700/740] Compiling C object app/dpdk-testpmd.p/test-pmd_iofwd.c.o 00:02:14.900 [701/740] Compiling C object app/dpdk-testpmd.p/test-pmd_macfwd.c.o 00:02:15.468 [702/740] Compiling C object app/dpdk-testpmd.p/test-pmd_macswap.c.o 00:02:15.468 [703/740] Compiling C object app/dpdk-testpmd.p/test-pmd_csumonly.c.o 00:02:15.468 [704/740] Compiling C object app/dpdk-testpmd.p/test-pmd_rxonly.c.o 00:02:15.468 [705/740] Compiling C object app/dpdk-testpmd.p/test-pmd_cmdline.c.o 00:02:15.468 [706/740] Compiling C object app/dpdk-testpmd.p/test-pmd_shared_rxq_fwd.c.o 00:02:15.468 [707/740] Compiling C object app/dpdk-testpmd.p/test-pmd_parameters.c.o 00:02:16.037 [708/740] Compiling C object app/dpdk-testpmd.p/test-pmd_bpf_cmd.c.o 00:02:16.037 [709/740] Compiling C object app/dpdk-testpmd.p/test-pmd_util.c.o 00:02:16.296 [710/740] Compiling C object app/dpdk-testpmd.p/.._drivers_net_i40e_i40e_testpmd.c.o 00:02:16.296 [711/740] Compiling C object app/dpdk-testpmd.p/test-pmd_config.c.o 00:02:16.296 [712/740] Compiling C object app/dpdk-test-regex.p/test-regex_main.c.o 00:02:16.297 [713/740] Compiling C object app/dpdk-test-sad.p/test-sad_main.c.o 00:02:16.297 [714/740] Compiling C object app/dpdk-testpmd.p/test-pmd_cmdline_flow.c.o 00:02:16.556 [715/740] Compiling C object app/dpdk-testpmd.p/test-pmd_txonly.c.o 00:02:16.556 [716/740] Compiling C object app/dpdk-testpmd.p/test-pmd_noisy_vnf.c.o 00:02:16.556 [717/740] Compiling C object app/dpdk-test-security-perf.p/test-security-perf_test_security_perf.c.o 00:02:16.815 [718/740] Compiling C object lib/librte_pipeline.a.p/pipeline_rte_table_action.c.o 00:02:16.815 [719/740] Linking static target lib/librte_pipeline.a 00:02:17.074 [720/740] Compiling C object app/dpdk-testpmd.p/test-pmd_testpmd.c.o 00:02:17.074 [721/740] Linking target app/dpdk-test-cmdline 00:02:17.074 [722/740] Linking target app/dpdk-test-acl 00:02:17.074 [723/740] Linking target app/dpdk-pdump 00:02:17.074 [724/740] Linking target app/dpdk-dumpcap 00:02:17.074 [725/740] Linking target app/dpdk-test-compress-perf 00:02:17.074 [726/740] Linking target app/dpdk-proc-info 00:02:17.074 [727/740] Linking target app/dpdk-test-bbdev 00:02:17.074 [728/740] Compiling C object app/dpdk-test-security-perf.p/test_test_cryptodev_security_ipsec.c.o 00:02:17.332 [729/740] Linking target app/dpdk-test-crypto-perf 00:02:17.591 [730/740] Linking target app/dpdk-test-pipeline 00:02:17.591 [731/740] Linking target app/dpdk-test-gpudev 00:02:17.591 [732/740] Linking target app/dpdk-test-flow-perf 00:02:17.591 [733/740] Linking target app/dpdk-test-regex 00:02:17.591 [734/740] Linking target app/dpdk-test-fib 00:02:17.591 [735/740] Linking target app/dpdk-test-eventdev 00:02:17.591 [736/740] Linking target app/dpdk-test-sad 00:02:17.591 [737/740] Linking target app/dpdk-testpmd 00:02:17.851 [738/740] Linking target app/dpdk-test-security-perf 00:02:22.062 [739/740] Generating lib/pipeline.sym_chk with a custom command (wrapped by meson to capture output) 00:02:22.062 [740/740] Linking target lib/librte_pipeline.so.23.0 00:02:22.326 23:36:10 build_native_dpdk -- common/autobuild_common.sh@201 -- $ uname -s 00:02:22.326 23:36:10 build_native_dpdk -- common/autobuild_common.sh@201 -- $ [[ Linux == \F\r\e\e\B\S\D ]] 00:02:22.326 23:36:10 build_native_dpdk -- common/autobuild_common.sh@214 -- $ ninja -C /home/vagrant/spdk_repo/dpdk/build-tmp -j10 install 00:02:22.326 ninja: Entering directory `/home/vagrant/spdk_repo/dpdk/build-tmp' 00:02:22.326 [0/1] Installing files. 00:02:22.588 Installing subdir /home/vagrant/spdk_repo/dpdk/examples to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples 00:02:22.588 Installing /home/vagrant/spdk_repo/dpdk/examples/bbdev_app/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/bbdev_app 00:02:22.588 Installing /home/vagrant/spdk_repo/dpdk/examples/bbdev_app/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/bbdev_app 00:02:22.588 Installing /home/vagrant/spdk_repo/dpdk/examples/bond/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/bond 00:02:22.588 Installing /home/vagrant/spdk_repo/dpdk/examples/bond/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/bond 00:02:22.588 Installing /home/vagrant/spdk_repo/dpdk/examples/bond/main.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/bond 00:02:22.588 Installing /home/vagrant/spdk_repo/dpdk/examples/bpf/README to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/bpf 00:02:22.588 Installing /home/vagrant/spdk_repo/dpdk/examples/bpf/dummy.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/bpf 00:02:22.588 Installing /home/vagrant/spdk_repo/dpdk/examples/bpf/t1.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/bpf 00:02:22.588 Installing /home/vagrant/spdk_repo/dpdk/examples/bpf/t2.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/bpf 00:02:22.588 Installing /home/vagrant/spdk_repo/dpdk/examples/bpf/t3.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/bpf 00:02:22.588 Installing /home/vagrant/spdk_repo/dpdk/examples/cmdline/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/cmdline 00:02:22.588 Installing /home/vagrant/spdk_repo/dpdk/examples/cmdline/commands.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/cmdline 00:02:22.588 Installing /home/vagrant/spdk_repo/dpdk/examples/cmdline/commands.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/cmdline 00:02:22.588 Installing /home/vagrant/spdk_repo/dpdk/examples/cmdline/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/cmdline 00:02:22.588 Installing /home/vagrant/spdk_repo/dpdk/examples/cmdline/parse_obj_list.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/cmdline 00:02:22.588 Installing /home/vagrant/spdk_repo/dpdk/examples/cmdline/parse_obj_list.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/cmdline 00:02:22.588 Installing /home/vagrant/spdk_repo/dpdk/examples/common/pkt_group.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/common 00:02:22.588 Installing /home/vagrant/spdk_repo/dpdk/examples/common/altivec/port_group.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/common/altivec 00:02:22.589 Installing /home/vagrant/spdk_repo/dpdk/examples/common/neon/port_group.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/common/neon 00:02:22.589 Installing /home/vagrant/spdk_repo/dpdk/examples/common/sse/port_group.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/common/sse 00:02:22.589 Installing /home/vagrant/spdk_repo/dpdk/examples/distributor/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/distributor 00:02:22.589 Installing /home/vagrant/spdk_repo/dpdk/examples/distributor/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/distributor 00:02:22.589 Installing /home/vagrant/spdk_repo/dpdk/examples/dma/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/dma 00:02:22.589 Installing /home/vagrant/spdk_repo/dpdk/examples/dma/dmafwd.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/dma 00:02:22.589 Installing /home/vagrant/spdk_repo/dpdk/examples/ethtool/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ethtool 00:02:22.589 Installing /home/vagrant/spdk_repo/dpdk/examples/ethtool/ethtool-app/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ethtool/ethtool-app 00:02:22.589 Installing /home/vagrant/spdk_repo/dpdk/examples/ethtool/ethtool-app/ethapp.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ethtool/ethtool-app 00:02:22.589 Installing /home/vagrant/spdk_repo/dpdk/examples/ethtool/ethtool-app/ethapp.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ethtool/ethtool-app 00:02:22.589 Installing /home/vagrant/spdk_repo/dpdk/examples/ethtool/ethtool-app/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ethtool/ethtool-app 00:02:22.589 Installing /home/vagrant/spdk_repo/dpdk/examples/ethtool/lib/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ethtool/lib 00:02:22.589 Installing /home/vagrant/spdk_repo/dpdk/examples/ethtool/lib/rte_ethtool.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ethtool/lib 00:02:22.589 Installing /home/vagrant/spdk_repo/dpdk/examples/ethtool/lib/rte_ethtool.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ethtool/lib 00:02:22.589 Installing /home/vagrant/spdk_repo/dpdk/examples/eventdev_pipeline/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/eventdev_pipeline 00:02:22.589 Installing /home/vagrant/spdk_repo/dpdk/examples/eventdev_pipeline/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/eventdev_pipeline 00:02:22.589 Installing /home/vagrant/spdk_repo/dpdk/examples/eventdev_pipeline/pipeline_common.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/eventdev_pipeline 00:02:22.589 Installing /home/vagrant/spdk_repo/dpdk/examples/eventdev_pipeline/pipeline_worker_generic.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/eventdev_pipeline 00:02:22.589 Installing /home/vagrant/spdk_repo/dpdk/examples/eventdev_pipeline/pipeline_worker_tx.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/eventdev_pipeline 00:02:22.589 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:02:22.589 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_dev_self_test.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:02:22.589 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_dev_self_test.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:02:22.589 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_validation.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:02:22.589 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_validation.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:02:22.589 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_validation_aes.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:02:22.589 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_validation_ccm.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:02:22.589 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_validation_cmac.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:02:22.589 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_validation_ecdsa.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:02:22.589 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_validation_gcm.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:02:22.589 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_validation_hmac.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:02:22.589 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_validation_rsa.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:02:22.589 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_validation_sha.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:02:22.589 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_validation_tdes.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:02:22.589 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_validation_xts.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:02:22.589 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:02:22.589 Installing /home/vagrant/spdk_repo/dpdk/examples/flow_classify/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/flow_classify 00:02:22.589 Installing /home/vagrant/spdk_repo/dpdk/examples/flow_classify/flow_classify.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/flow_classify 00:02:22.589 Installing /home/vagrant/spdk_repo/dpdk/examples/flow_classify/ipv4_rules_file.txt to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/flow_classify 00:02:22.589 Installing /home/vagrant/spdk_repo/dpdk/examples/flow_filtering/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/flow_filtering 00:02:22.589 Installing /home/vagrant/spdk_repo/dpdk/examples/flow_filtering/flow_blocks.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/flow_filtering 00:02:22.589 Installing /home/vagrant/spdk_repo/dpdk/examples/flow_filtering/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/flow_filtering 00:02:22.589 Installing /home/vagrant/spdk_repo/dpdk/examples/helloworld/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/helloworld 00:02:22.589 Installing /home/vagrant/spdk_repo/dpdk/examples/helloworld/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/helloworld 00:02:22.589 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_fragmentation/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_fragmentation 00:02:22.589 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_fragmentation/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_fragmentation 00:02:22.589 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:22.589 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/action.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:22.589 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/action.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:22.589 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/cli.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:22.589 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/cli.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:22.589 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/common.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:22.589 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/conn.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:22.589 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/conn.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:22.589 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/cryptodev.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:22.589 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/cryptodev.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:22.589 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/kni.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:22.589 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/kni.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:22.589 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/link.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:22.589 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/link.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:22.589 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:22.589 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/mempool.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:22.589 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/mempool.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:22.589 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/parser.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:22.589 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/parser.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:22.589 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/pipeline.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:22.589 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/pipeline.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:22.589 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/swq.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:22.589 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/swq.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:22.589 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/tap.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:22.589 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/tap.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:22.589 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/thread.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:22.589 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/thread.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:22.589 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/tmgr.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:22.589 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/tmgr.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:22.589 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/examples/firewall.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline/examples 00:02:22.589 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/examples/flow.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline/examples 00:02:22.589 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/examples/flow_crypto.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline/examples 00:02:22.589 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/examples/kni.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline/examples 00:02:22.589 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/examples/l2fwd.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline/examples 00:02:22.589 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/examples/route.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline/examples 00:02:22.589 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/examples/route_ecmp.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline/examples 00:02:22.589 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/examples/rss.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline/examples 00:02:22.589 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/examples/tap.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline/examples 00:02:22.589 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_reassembly/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_reassembly 00:02:22.589 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_reassembly/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_reassembly 00:02:22.590 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:02:22.590 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/ep0.cfg to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:02:22.590 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/ep1.cfg to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:02:22.590 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/esp.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:02:22.590 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/esp.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:02:22.590 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/event_helper.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:02:22.590 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/event_helper.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:02:22.590 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/flow.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:02:22.590 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/flow.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:02:22.590 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/ipip.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:02:22.590 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/ipsec-secgw.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:02:22.590 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/ipsec-secgw.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:02:22.590 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/ipsec.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:02:22.590 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/ipsec.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:02:22.590 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/ipsec_lpm_neon.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:02:22.590 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/ipsec_neon.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:02:22.590 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/ipsec_process.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:02:22.590 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/ipsec_worker.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:02:22.590 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/ipsec_worker.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:02:22.590 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/parser.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:02:22.590 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/parser.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:02:22.590 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/rt.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:02:22.590 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/sa.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:02:22.590 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/sad.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:02:22.590 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/sad.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:02:22.590 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/sp4.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:02:22.590 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/sp6.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:02:22.590 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/bypass_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:02:22.590 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/common_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:02:22.590 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/common_defs_secgw.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:02:22.590 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/data_rxtx.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:02:22.590 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/linux_test.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:02:22.590 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/load_env.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:02:22.590 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/pkttest.py to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:02:22.590 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/pkttest.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:02:22.590 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/run_test.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:02:22.590 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/trs_3descbc_sha1_common_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:02:22.590 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/trs_3descbc_sha1_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:02:22.590 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/trs_aescbc_sha1_common_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:02:22.590 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/trs_aescbc_sha1_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:02:22.590 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/trs_aesctr_sha1_common_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:02:22.590 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/trs_aesctr_sha1_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:02:22.590 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/trs_aesgcm_common_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:02:22.590 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/trs_aesgcm_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:02:22.590 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/trs_ipv6opts.py to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:02:22.590 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/tun_3descbc_sha1_common_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:02:22.590 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/tun_3descbc_sha1_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:02:22.590 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/tun_aescbc_sha1_common_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:02:22.590 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/tun_aescbc_sha1_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:02:22.590 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/tun_aesctr_sha1_common_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:02:22.590 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/tun_aesctr_sha1_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:02:22.590 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/tun_aesgcm_common_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:02:22.590 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/tun_aesgcm_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:02:22.590 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/tun_null_header_reconstruct.py to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:02:22.590 Installing /home/vagrant/spdk_repo/dpdk/examples/ipv4_multicast/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipv4_multicast 00:02:22.590 Installing /home/vagrant/spdk_repo/dpdk/examples/ipv4_multicast/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipv4_multicast 00:02:22.590 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-cat/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-cat 00:02:22.590 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-cat/cat.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-cat 00:02:22.590 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-cat/cat.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-cat 00:02:22.590 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-cat/l2fwd-cat.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-cat 00:02:22.590 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-crypto/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-crypto 00:02:22.590 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-crypto/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-crypto 00:02:22.590 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-event/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-event 00:02:22.590 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-event/l2fwd_common.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-event 00:02:22.590 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-event/l2fwd_common.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-event 00:02:22.590 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-event/l2fwd_event.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-event 00:02:22.590 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-event/l2fwd_event.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-event 00:02:22.590 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-event/l2fwd_event_generic.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-event 00:02:22.590 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-event/l2fwd_event_internal_port.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-event 00:02:22.590 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-event/l2fwd_poll.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-event 00:02:22.590 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-event/l2fwd_poll.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-event 00:02:22.590 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-event/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-event 00:02:22.590 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-jobstats/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-jobstats 00:02:22.590 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-jobstats/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-jobstats 00:02:22.590 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-keepalive/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-keepalive 00:02:22.590 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-keepalive/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-keepalive 00:02:22.590 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-keepalive/shm.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-keepalive 00:02:22.590 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-keepalive/shm.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-keepalive 00:02:22.590 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-keepalive/ka-agent/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-keepalive/ka-agent 00:02:22.590 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-keepalive/ka-agent/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-keepalive/ka-agent 00:02:22.590 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd 00:02:22.590 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd 00:02:22.590 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd-graph/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd-graph 00:02:22.590 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd-graph/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd-graph 00:02:22.590 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd-power/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd-power 00:02:22.590 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd-power/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd-power 00:02:22.591 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd-power/main.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd-power 00:02:22.591 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd-power/perf_core.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd-power 00:02:22.591 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd-power/perf_core.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd-power 00:02:22.591 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:22.591 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/em_default_v4.cfg to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:22.591 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/em_default_v6.cfg to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:22.591 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/em_route_parse.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:22.591 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:22.591 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_acl.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:22.591 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_acl.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:22.591 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_acl_scalar.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:22.591 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_altivec.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:22.591 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_common.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:22.591 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_em.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:22.591 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_em.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:22.591 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_em_hlm.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:22.591 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_em_hlm_neon.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:22.591 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_em_hlm_sse.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:22.591 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_em_sequential.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:22.591 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_event.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:22.591 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_event.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:22.591 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_event_generic.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:22.591 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_event_internal_port.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:22.591 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_fib.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:22.591 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_lpm.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:22.591 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_lpm.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:22.591 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_lpm_altivec.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:22.591 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_lpm_neon.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:22.591 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_lpm_sse.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:22.591 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_neon.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:22.591 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_route.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:22.591 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_sse.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:22.591 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/lpm_default_v4.cfg to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:22.591 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/lpm_default_v6.cfg to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:22.591 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/lpm_route_parse.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:22.591 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:22.591 Installing /home/vagrant/spdk_repo/dpdk/examples/link_status_interrupt/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/link_status_interrupt 00:02:22.591 Installing /home/vagrant/spdk_repo/dpdk/examples/link_status_interrupt/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/link_status_interrupt 00:02:22.591 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process 00:02:22.591 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/client_server_mp/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/client_server_mp 00:02:22.591 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/client_server_mp/mp_client/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/client_server_mp/mp_client 00:02:22.591 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/client_server_mp/mp_client/client.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/client_server_mp/mp_client 00:02:22.591 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/client_server_mp/mp_server/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/client_server_mp/mp_server 00:02:22.591 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/client_server_mp/mp_server/args.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/client_server_mp/mp_server 00:02:22.591 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/client_server_mp/mp_server/args.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/client_server_mp/mp_server 00:02:22.591 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/client_server_mp/mp_server/init.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/client_server_mp/mp_server 00:02:22.591 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/client_server_mp/mp_server/init.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/client_server_mp/mp_server 00:02:22.591 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/client_server_mp/mp_server/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/client_server_mp/mp_server 00:02:22.591 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/client_server_mp/shared/common.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/client_server_mp/shared 00:02:22.591 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/hotplug_mp/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/hotplug_mp 00:02:22.591 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/hotplug_mp/commands.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/hotplug_mp 00:02:22.591 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/hotplug_mp/commands.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/hotplug_mp 00:02:22.591 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/hotplug_mp/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/hotplug_mp 00:02:22.591 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/simple_mp/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/simple_mp 00:02:22.591 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/simple_mp/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/simple_mp 00:02:22.591 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/simple_mp/mp_commands.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/simple_mp 00:02:22.591 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/simple_mp/mp_commands.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/simple_mp 00:02:22.591 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/symmetric_mp/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/symmetric_mp 00:02:22.591 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/symmetric_mp/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/symmetric_mp 00:02:22.591 Installing /home/vagrant/spdk_repo/dpdk/examples/ntb/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ntb 00:02:22.591 Installing /home/vagrant/spdk_repo/dpdk/examples/ntb/ntb_fwd.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ntb 00:02:22.591 Installing /home/vagrant/spdk_repo/dpdk/examples/packet_ordering/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/packet_ordering 00:02:22.591 Installing /home/vagrant/spdk_repo/dpdk/examples/packet_ordering/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/packet_ordering 00:02:22.591 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline 00:02:22.591 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/cli.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline 00:02:22.591 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/cli.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline 00:02:22.591 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/conn.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline 00:02:22.591 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/conn.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline 00:02:22.591 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline 00:02:22.591 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/obj.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline 00:02:22.591 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/obj.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline 00:02:22.591 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/thread.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline 00:02:22.591 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/thread.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline 00:02:22.591 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/ethdev.io to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:22.591 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/fib.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:22.591 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/fib.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:22.591 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/fib_nexthop_group_table.txt to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:22.591 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/fib_nexthop_table.txt to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:22.591 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/fib_routing_table.txt to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:22.591 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/hash_func.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:22.591 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/hash_func.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:22.591 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/l2fwd.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:22.591 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/l2fwd.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:22.591 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/l2fwd_macswp.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:22.591 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/l2fwd_macswp.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:22.592 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/l2fwd_macswp_pcap.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:22.592 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/l2fwd_pcap.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:22.592 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/learner.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:22.592 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/learner.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:22.592 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/meter.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:22.592 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/meter.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:22.592 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/mirroring.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:22.592 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/mirroring.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:22.592 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/packet.txt to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:22.592 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/pcap.io to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:22.592 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/recirculation.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:22.592 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/recirculation.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:22.592 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/registers.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:22.592 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/registers.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:22.592 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/selector.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:22.592 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/selector.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:22.592 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/selector.txt to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:22.592 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/varbit.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:22.592 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/varbit.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:22.592 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/vxlan.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:22.592 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/vxlan.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:22.592 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/vxlan_pcap.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:22.592 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/vxlan_table.py to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:22.592 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/vxlan_table.txt to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:22.592 Installing /home/vagrant/spdk_repo/dpdk/examples/ptpclient/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ptpclient 00:02:22.592 Installing /home/vagrant/spdk_repo/dpdk/examples/ptpclient/ptpclient.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ptpclient 00:02:22.592 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_meter/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_meter 00:02:22.592 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_meter/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_meter 00:02:22.592 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_meter/main.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_meter 00:02:22.592 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_meter/rte_policer.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_meter 00:02:22.592 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_meter/rte_policer.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_meter 00:02:22.592 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:02:22.592 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/app_thread.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:02:22.592 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/args.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:02:22.592 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/cfg_file.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:02:22.592 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/cfg_file.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:02:22.592 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/cmdline.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:02:22.592 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/init.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:02:22.592 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:02:22.592 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/main.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:02:22.592 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/profile.cfg to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:02:22.592 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/profile_ov.cfg to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:02:22.592 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/profile_pie.cfg to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:02:22.592 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/profile_red.cfg to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:02:22.592 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/stats.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:02:22.592 Installing /home/vagrant/spdk_repo/dpdk/examples/rxtx_callbacks/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/rxtx_callbacks 00:02:22.592 Installing /home/vagrant/spdk_repo/dpdk/examples/rxtx_callbacks/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/rxtx_callbacks 00:02:22.592 Installing /home/vagrant/spdk_repo/dpdk/examples/server_node_efd/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/server_node_efd 00:02:22.592 Installing /home/vagrant/spdk_repo/dpdk/examples/server_node_efd/node/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/server_node_efd/node 00:02:22.592 Installing /home/vagrant/spdk_repo/dpdk/examples/server_node_efd/node/node.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/server_node_efd/node 00:02:22.592 Installing /home/vagrant/spdk_repo/dpdk/examples/server_node_efd/server/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/server_node_efd/server 00:02:22.592 Installing /home/vagrant/spdk_repo/dpdk/examples/server_node_efd/server/args.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/server_node_efd/server 00:02:22.592 Installing /home/vagrant/spdk_repo/dpdk/examples/server_node_efd/server/args.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/server_node_efd/server 00:02:22.592 Installing /home/vagrant/spdk_repo/dpdk/examples/server_node_efd/server/init.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/server_node_efd/server 00:02:22.592 Installing /home/vagrant/spdk_repo/dpdk/examples/server_node_efd/server/init.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/server_node_efd/server 00:02:22.592 Installing /home/vagrant/spdk_repo/dpdk/examples/server_node_efd/server/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/server_node_efd/server 00:02:22.592 Installing /home/vagrant/spdk_repo/dpdk/examples/server_node_efd/shared/common.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/server_node_efd/shared 00:02:22.592 Installing /home/vagrant/spdk_repo/dpdk/examples/service_cores/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/service_cores 00:02:22.592 Installing /home/vagrant/spdk_repo/dpdk/examples/service_cores/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/service_cores 00:02:22.592 Installing /home/vagrant/spdk_repo/dpdk/examples/skeleton/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/skeleton 00:02:22.592 Installing /home/vagrant/spdk_repo/dpdk/examples/skeleton/basicfwd.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/skeleton 00:02:22.592 Installing /home/vagrant/spdk_repo/dpdk/examples/timer/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/timer 00:02:22.592 Installing /home/vagrant/spdk_repo/dpdk/examples/timer/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/timer 00:02:22.592 Installing /home/vagrant/spdk_repo/dpdk/examples/vdpa/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vdpa 00:02:22.592 Installing /home/vagrant/spdk_repo/dpdk/examples/vdpa/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vdpa 00:02:22.592 Installing /home/vagrant/spdk_repo/dpdk/examples/vdpa/vdpa_blk_compact.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vdpa 00:02:22.592 Installing /home/vagrant/spdk_repo/dpdk/examples/vhost/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vhost 00:02:22.592 Installing /home/vagrant/spdk_repo/dpdk/examples/vhost/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vhost 00:02:22.592 Installing /home/vagrant/spdk_repo/dpdk/examples/vhost/main.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vhost 00:02:22.592 Installing /home/vagrant/spdk_repo/dpdk/examples/vhost/virtio_net.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vhost 00:02:22.592 Installing /home/vagrant/spdk_repo/dpdk/examples/vhost_blk/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vhost_blk 00:02:22.592 Installing /home/vagrant/spdk_repo/dpdk/examples/vhost_blk/blk.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vhost_blk 00:02:22.592 Installing /home/vagrant/spdk_repo/dpdk/examples/vhost_blk/blk_spec.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vhost_blk 00:02:22.592 Installing /home/vagrant/spdk_repo/dpdk/examples/vhost_blk/vhost_blk.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vhost_blk 00:02:22.592 Installing /home/vagrant/spdk_repo/dpdk/examples/vhost_blk/vhost_blk.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vhost_blk 00:02:22.592 Installing /home/vagrant/spdk_repo/dpdk/examples/vhost_blk/vhost_blk_compat.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vhost_blk 00:02:22.592 Installing /home/vagrant/spdk_repo/dpdk/examples/vhost_crypto/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vhost_crypto 00:02:22.592 Installing /home/vagrant/spdk_repo/dpdk/examples/vhost_crypto/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vhost_crypto 00:02:22.592 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:02:22.592 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/channel_manager.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:02:22.592 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/channel_manager.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:02:22.592 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/channel_monitor.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:02:22.592 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/channel_monitor.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:02:22.592 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:02:22.592 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/oob_monitor.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:02:22.592 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/oob_monitor_nop.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:02:22.592 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/oob_monitor_x86.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:02:22.593 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/parse.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:02:22.593 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/parse.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:02:22.593 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/power_manager.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:02:22.593 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/power_manager.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:02:22.593 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/vm_power_cli.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:02:22.593 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/vm_power_cli.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:02:22.593 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/guest_cli/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager/guest_cli 00:02:22.593 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/guest_cli/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager/guest_cli 00:02:22.593 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/guest_cli/parse.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager/guest_cli 00:02:22.593 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/guest_cli/parse.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager/guest_cli 00:02:22.593 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/guest_cli/vm_power_cli_guest.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager/guest_cli 00:02:22.593 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/guest_cli/vm_power_cli_guest.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager/guest_cli 00:02:22.593 Installing /home/vagrant/spdk_repo/dpdk/examples/vmdq/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vmdq 00:02:22.593 Installing /home/vagrant/spdk_repo/dpdk/examples/vmdq/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vmdq 00:02:22.593 Installing /home/vagrant/spdk_repo/dpdk/examples/vmdq_dcb/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vmdq_dcb 00:02:22.593 Installing /home/vagrant/spdk_repo/dpdk/examples/vmdq_dcb/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vmdq_dcb 00:02:22.593 Installing lib/librte_kvargs.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:22.593 Installing lib/librte_kvargs.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:22.593 Installing lib/librte_telemetry.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:22.593 Installing lib/librte_telemetry.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:22.593 Installing lib/librte_eal.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:22.593 Installing lib/librte_eal.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:22.593 Installing lib/librte_ring.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:22.593 Installing lib/librte_ring.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:22.593 Installing lib/librte_rcu.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:22.593 Installing lib/librte_rcu.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:22.593 Installing lib/librte_mempool.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:22.593 Installing lib/librte_mempool.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:22.593 Installing lib/librte_mbuf.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:22.593 Installing lib/librte_mbuf.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:22.593 Installing lib/librte_net.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:22.854 Installing lib/librte_net.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:22.854 Installing lib/librte_meter.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:22.854 Installing lib/librte_meter.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:22.854 Installing lib/librte_ethdev.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:22.854 Installing lib/librte_ethdev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:22.854 Installing lib/librte_pci.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:22.854 Installing lib/librte_pci.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:22.854 Installing lib/librte_cmdline.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:22.854 Installing lib/librte_cmdline.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:22.854 Installing lib/librte_metrics.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:22.854 Installing lib/librte_metrics.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:22.854 Installing lib/librte_hash.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:22.854 Installing lib/librte_hash.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:22.854 Installing lib/librte_timer.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:22.854 Installing lib/librte_timer.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:22.854 Installing lib/librte_acl.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:22.854 Installing lib/librte_acl.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:22.854 Installing lib/librte_bbdev.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:22.854 Installing lib/librte_bbdev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:22.854 Installing lib/librte_bitratestats.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:22.854 Installing lib/librte_bitratestats.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:22.854 Installing lib/librte_bpf.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:22.854 Installing lib/librte_bpf.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:22.854 Installing lib/librte_cfgfile.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:22.854 Installing lib/librte_cfgfile.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:22.854 Installing lib/librte_compressdev.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:22.854 Installing lib/librte_compressdev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:22.854 Installing lib/librte_cryptodev.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:22.854 Installing lib/librte_cryptodev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:22.854 Installing lib/librte_distributor.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:22.854 Installing lib/librte_distributor.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:22.854 Installing lib/librte_efd.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:22.854 Installing lib/librte_efd.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:22.854 Installing lib/librte_eventdev.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:22.854 Installing lib/librte_eventdev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:22.854 Installing lib/librte_gpudev.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:22.854 Installing lib/librte_gpudev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:22.854 Installing lib/librte_gro.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:22.854 Installing lib/librte_gro.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:22.854 Installing lib/librte_gso.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:22.854 Installing lib/librte_gso.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:22.854 Installing lib/librte_ip_frag.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:22.854 Installing lib/librte_ip_frag.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:22.854 Installing lib/librte_jobstats.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:22.854 Installing lib/librte_jobstats.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:22.854 Installing lib/librte_latencystats.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:22.854 Installing lib/librte_latencystats.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:22.854 Installing lib/librte_lpm.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:22.854 Installing lib/librte_lpm.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:22.854 Installing lib/librte_member.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:22.854 Installing lib/librte_member.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:22.854 Installing lib/librte_pcapng.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:22.854 Installing lib/librte_pcapng.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:22.854 Installing lib/librte_power.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:22.854 Installing lib/librte_power.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:22.854 Installing lib/librte_rawdev.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:22.854 Installing lib/librte_rawdev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:22.854 Installing lib/librte_regexdev.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:22.854 Installing lib/librte_regexdev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:22.854 Installing lib/librte_dmadev.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:22.854 Installing lib/librte_dmadev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:22.854 Installing lib/librte_rib.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:22.854 Installing lib/librte_rib.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:22.854 Installing lib/librte_reorder.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:22.854 Installing lib/librte_reorder.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:22.854 Installing lib/librte_sched.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:22.854 Installing lib/librte_sched.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:22.855 Installing lib/librte_security.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:22.855 Installing lib/librte_security.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:22.855 Installing lib/librte_stack.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:22.855 Installing lib/librte_stack.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:22.855 Installing lib/librte_vhost.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:22.855 Installing lib/librte_vhost.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:22.855 Installing lib/librte_ipsec.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:22.855 Installing lib/librte_ipsec.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:22.855 Installing lib/librte_fib.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:22.855 Installing lib/librte_fib.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:22.855 Installing lib/librte_port.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:22.855 Installing lib/librte_port.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:22.855 Installing lib/librte_pdump.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:22.855 Installing lib/librte_pdump.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:22.855 Installing lib/librte_table.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:22.855 Installing lib/librte_table.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:22.855 Installing lib/librte_pipeline.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:22.855 Installing lib/librte_pipeline.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:22.855 Installing lib/librte_graph.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:22.855 Installing lib/librte_graph.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:22.855 Installing lib/librte_node.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:22.855 Installing lib/librte_node.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:22.855 Installing drivers/librte_bus_pci.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:22.855 Installing drivers/librte_bus_pci.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0 00:02:22.855 Installing drivers/librte_bus_vdev.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:22.855 Installing drivers/librte_bus_vdev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0 00:02:22.855 Installing drivers/librte_mempool_ring.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:22.855 Installing drivers/librte_mempool_ring.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0 00:02:22.855 Installing drivers/librte_net_i40e.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:22.855 Installing drivers/librte_net_i40e.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0 00:02:22.855 Installing app/dpdk-dumpcap to /home/vagrant/spdk_repo/dpdk/build/bin 00:02:22.855 Installing app/dpdk-pdump to /home/vagrant/spdk_repo/dpdk/build/bin 00:02:22.855 Installing app/dpdk-proc-info to /home/vagrant/spdk_repo/dpdk/build/bin 00:02:22.855 Installing app/dpdk-test-acl to /home/vagrant/spdk_repo/dpdk/build/bin 00:02:22.855 Installing app/dpdk-test-bbdev to /home/vagrant/spdk_repo/dpdk/build/bin 00:02:22.855 Installing app/dpdk-test-cmdline to /home/vagrant/spdk_repo/dpdk/build/bin 00:02:22.855 Installing app/dpdk-test-compress-perf to /home/vagrant/spdk_repo/dpdk/build/bin 00:02:22.855 Installing app/dpdk-test-crypto-perf to /home/vagrant/spdk_repo/dpdk/build/bin 00:02:22.855 Installing app/dpdk-test-eventdev to /home/vagrant/spdk_repo/dpdk/build/bin 00:02:22.855 Installing app/dpdk-test-fib to /home/vagrant/spdk_repo/dpdk/build/bin 00:02:22.855 Installing app/dpdk-test-flow-perf to /home/vagrant/spdk_repo/dpdk/build/bin 00:02:22.855 Installing app/dpdk-test-gpudev to /home/vagrant/spdk_repo/dpdk/build/bin 00:02:22.855 Installing app/dpdk-test-pipeline to /home/vagrant/spdk_repo/dpdk/build/bin 00:02:22.855 Installing app/dpdk-testpmd to /home/vagrant/spdk_repo/dpdk/build/bin 00:02:22.855 Installing app/dpdk-test-regex to /home/vagrant/spdk_repo/dpdk/build/bin 00:02:22.855 Installing app/dpdk-test-sad to /home/vagrant/spdk_repo/dpdk/build/bin 00:02:22.855 Installing app/dpdk-test-security-perf to /home/vagrant/spdk_repo/dpdk/build/bin 00:02:22.855 Installing /home/vagrant/spdk_repo/dpdk/config/rte_config.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:22.855 Installing /home/vagrant/spdk_repo/dpdk/lib/kvargs/rte_kvargs.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:22.855 Installing /home/vagrant/spdk_repo/dpdk/lib/telemetry/rte_telemetry.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:22.855 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/generic/rte_atomic.h to /home/vagrant/spdk_repo/dpdk/build/include/generic 00:02:22.855 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/generic/rte_byteorder.h to /home/vagrant/spdk_repo/dpdk/build/include/generic 00:02:22.855 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/generic/rte_cpuflags.h to /home/vagrant/spdk_repo/dpdk/build/include/generic 00:02:22.855 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/generic/rte_cycles.h to /home/vagrant/spdk_repo/dpdk/build/include/generic 00:02:22.855 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/generic/rte_io.h to /home/vagrant/spdk_repo/dpdk/build/include/generic 00:02:22.855 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/generic/rte_memcpy.h to /home/vagrant/spdk_repo/dpdk/build/include/generic 00:02:22.855 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/generic/rte_pause.h to /home/vagrant/spdk_repo/dpdk/build/include/generic 00:02:22.855 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/generic/rte_power_intrinsics.h to /home/vagrant/spdk_repo/dpdk/build/include/generic 00:02:22.855 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/generic/rte_prefetch.h to /home/vagrant/spdk_repo/dpdk/build/include/generic 00:02:22.855 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/generic/rte_rwlock.h to /home/vagrant/spdk_repo/dpdk/build/include/generic 00:02:22.855 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/generic/rte_spinlock.h to /home/vagrant/spdk_repo/dpdk/build/include/generic 00:02:22.855 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/generic/rte_vect.h to /home/vagrant/spdk_repo/dpdk/build/include/generic 00:02:22.855 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_atomic.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:22.855 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_byteorder.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:22.855 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_cpuflags.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:22.855 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_cycles.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:22.855 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_io.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:22.855 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_memcpy.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:22.855 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_pause.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:22.855 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_power_intrinsics.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:22.855 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_prefetch.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:22.855 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_rtm.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:22.855 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_rwlock.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:22.855 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_spinlock.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:22.855 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_vect.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:22.855 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_atomic_32.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:22.855 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_atomic_64.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:22.855 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_byteorder_32.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:22.855 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_byteorder_64.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:22.855 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_alarm.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:22.855 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_bitmap.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:22.855 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_bitops.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:22.855 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_branch_prediction.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:22.855 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_bus.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:22.855 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_class.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:22.855 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_common.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:22.855 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_compat.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:22.855 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_debug.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:22.855 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_dev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:22.855 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_devargs.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:22.855 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_eal.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:22.855 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_eal_memconfig.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:22.855 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_eal_trace.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:22.855 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_errno.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:22.855 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_epoll.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:22.855 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_fbarray.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:22.855 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_hexdump.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:22.855 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_hypervisor.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:22.855 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_interrupts.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:22.855 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_keepalive.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:22.855 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_launch.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:22.855 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_lcore.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:22.855 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_log.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:22.855 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_malloc.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:22.855 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_mcslock.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:22.855 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_memory.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:22.855 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_memzone.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:22.855 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_pci_dev_feature_defs.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:22.855 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_pci_dev_features.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:22.855 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_per_lcore.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.117 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_pflock.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.117 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_random.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.117 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_reciprocal.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.117 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_seqcount.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.117 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_seqlock.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.117 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_service.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.117 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_service_component.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.117 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_string_fns.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.117 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_tailq.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.117 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_thread.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.117 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_ticketlock.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.117 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_time.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.117 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_trace.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.117 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_trace_point.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.117 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_trace_point_register.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.117 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_uuid.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.117 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_version.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.117 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_vfio.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.117 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/linux/include/rte_os.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.117 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.117 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring_core.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.117 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring_elem.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.117 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring_elem_pvt.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.117 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring_c11_pvt.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.117 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring_generic_pvt.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.117 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring_hts.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.118 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring_hts_elem_pvt.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.118 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring_peek.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.118 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring_peek_elem_pvt.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.118 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring_peek_zc.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.118 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring_rts.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.118 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring_rts_elem_pvt.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.118 Installing /home/vagrant/spdk_repo/dpdk/lib/rcu/rte_rcu_qsbr.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.118 Installing /home/vagrant/spdk_repo/dpdk/lib/mempool/rte_mempool.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.118 Installing /home/vagrant/spdk_repo/dpdk/lib/mempool/rte_mempool_trace.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.118 Installing /home/vagrant/spdk_repo/dpdk/lib/mempool/rte_mempool_trace_fp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.118 Installing /home/vagrant/spdk_repo/dpdk/lib/mbuf/rte_mbuf.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.118 Installing /home/vagrant/spdk_repo/dpdk/lib/mbuf/rte_mbuf_core.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.118 Installing /home/vagrant/spdk_repo/dpdk/lib/mbuf/rte_mbuf_ptype.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.118 Installing /home/vagrant/spdk_repo/dpdk/lib/mbuf/rte_mbuf_pool_ops.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.118 Installing /home/vagrant/spdk_repo/dpdk/lib/mbuf/rte_mbuf_dyn.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.118 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_ip.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.118 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_tcp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.118 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_udp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.118 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_esp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.118 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_sctp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.118 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_icmp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.118 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_arp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.118 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_ether.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.118 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_macsec.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.118 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_vxlan.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.118 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_gre.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.118 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_gtp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.118 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_net.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.118 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_net_crc.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.118 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_mpls.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.118 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_higig.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.118 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_ecpri.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.118 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_geneve.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.118 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_l2tpv2.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.118 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_ppp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.118 Installing /home/vagrant/spdk_repo/dpdk/lib/meter/rte_meter.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.118 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_cman.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.118 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_ethdev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.118 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_ethdev_trace.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.118 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_ethdev_trace_fp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.118 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_dev_info.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.118 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_flow.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.118 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_flow_driver.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.118 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_mtr.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.118 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_mtr_driver.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.118 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_tm.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.118 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_tm_driver.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.118 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_ethdev_core.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.118 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_eth_ctrl.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.118 Installing /home/vagrant/spdk_repo/dpdk/lib/pci/rte_pci.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.118 Installing /home/vagrant/spdk_repo/dpdk/lib/cmdline/cmdline.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.118 Installing /home/vagrant/spdk_repo/dpdk/lib/cmdline/cmdline_parse.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.118 Installing /home/vagrant/spdk_repo/dpdk/lib/cmdline/cmdline_parse_num.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.118 Installing /home/vagrant/spdk_repo/dpdk/lib/cmdline/cmdline_parse_ipaddr.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.118 Installing /home/vagrant/spdk_repo/dpdk/lib/cmdline/cmdline_parse_etheraddr.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.118 Installing /home/vagrant/spdk_repo/dpdk/lib/cmdline/cmdline_parse_string.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.118 Installing /home/vagrant/spdk_repo/dpdk/lib/cmdline/cmdline_rdline.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.118 Installing /home/vagrant/spdk_repo/dpdk/lib/cmdline/cmdline_vt100.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.118 Installing /home/vagrant/spdk_repo/dpdk/lib/cmdline/cmdline_socket.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.118 Installing /home/vagrant/spdk_repo/dpdk/lib/cmdline/cmdline_cirbuf.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.118 Installing /home/vagrant/spdk_repo/dpdk/lib/cmdline/cmdline_parse_portlist.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.118 Installing /home/vagrant/spdk_repo/dpdk/lib/metrics/rte_metrics.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.118 Installing /home/vagrant/spdk_repo/dpdk/lib/metrics/rte_metrics_telemetry.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.118 Installing /home/vagrant/spdk_repo/dpdk/lib/hash/rte_fbk_hash.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.118 Installing /home/vagrant/spdk_repo/dpdk/lib/hash/rte_hash_crc.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.118 Installing /home/vagrant/spdk_repo/dpdk/lib/hash/rte_hash.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.118 Installing /home/vagrant/spdk_repo/dpdk/lib/hash/rte_jhash.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.118 Installing /home/vagrant/spdk_repo/dpdk/lib/hash/rte_thash.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.118 Installing /home/vagrant/spdk_repo/dpdk/lib/hash/rte_thash_gfni.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.118 Installing /home/vagrant/spdk_repo/dpdk/lib/hash/rte_crc_arm64.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.118 Installing /home/vagrant/spdk_repo/dpdk/lib/hash/rte_crc_generic.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.118 Installing /home/vagrant/spdk_repo/dpdk/lib/hash/rte_crc_sw.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.118 Installing /home/vagrant/spdk_repo/dpdk/lib/hash/rte_crc_x86.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.118 Installing /home/vagrant/spdk_repo/dpdk/lib/hash/rte_thash_x86_gfni.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.118 Installing /home/vagrant/spdk_repo/dpdk/lib/timer/rte_timer.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.118 Installing /home/vagrant/spdk_repo/dpdk/lib/acl/rte_acl.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.118 Installing /home/vagrant/spdk_repo/dpdk/lib/acl/rte_acl_osdep.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.118 Installing /home/vagrant/spdk_repo/dpdk/lib/bbdev/rte_bbdev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.118 Installing /home/vagrant/spdk_repo/dpdk/lib/bbdev/rte_bbdev_pmd.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.118 Installing /home/vagrant/spdk_repo/dpdk/lib/bbdev/rte_bbdev_op.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.118 Installing /home/vagrant/spdk_repo/dpdk/lib/bitratestats/rte_bitrate.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.118 Installing /home/vagrant/spdk_repo/dpdk/lib/bpf/bpf_def.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.118 Installing /home/vagrant/spdk_repo/dpdk/lib/bpf/rte_bpf.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.118 Installing /home/vagrant/spdk_repo/dpdk/lib/bpf/rte_bpf_ethdev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.118 Installing /home/vagrant/spdk_repo/dpdk/lib/cfgfile/rte_cfgfile.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.118 Installing /home/vagrant/spdk_repo/dpdk/lib/compressdev/rte_compressdev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.118 Installing /home/vagrant/spdk_repo/dpdk/lib/compressdev/rte_comp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.118 Installing /home/vagrant/spdk_repo/dpdk/lib/cryptodev/rte_cryptodev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.118 Installing /home/vagrant/spdk_repo/dpdk/lib/cryptodev/rte_cryptodev_trace.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.118 Installing /home/vagrant/spdk_repo/dpdk/lib/cryptodev/rte_cryptodev_trace_fp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.118 Installing /home/vagrant/spdk_repo/dpdk/lib/cryptodev/rte_crypto.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.118 Installing /home/vagrant/spdk_repo/dpdk/lib/cryptodev/rte_crypto_sym.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.118 Installing /home/vagrant/spdk_repo/dpdk/lib/cryptodev/rte_crypto_asym.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.118 Installing /home/vagrant/spdk_repo/dpdk/lib/cryptodev/rte_cryptodev_core.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.118 Installing /home/vagrant/spdk_repo/dpdk/lib/distributor/rte_distributor.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.118 Installing /home/vagrant/spdk_repo/dpdk/lib/efd/rte_efd.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.118 Installing /home/vagrant/spdk_repo/dpdk/lib/eventdev/rte_event_crypto_adapter.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.118 Installing /home/vagrant/spdk_repo/dpdk/lib/eventdev/rte_event_eth_rx_adapter.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.118 Installing /home/vagrant/spdk_repo/dpdk/lib/eventdev/rte_event_eth_tx_adapter.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.118 Installing /home/vagrant/spdk_repo/dpdk/lib/eventdev/rte_event_ring.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.118 Installing /home/vagrant/spdk_repo/dpdk/lib/eventdev/rte_event_timer_adapter.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.118 Installing /home/vagrant/spdk_repo/dpdk/lib/eventdev/rte_eventdev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.118 Installing /home/vagrant/spdk_repo/dpdk/lib/eventdev/rte_eventdev_trace_fp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.118 Installing /home/vagrant/spdk_repo/dpdk/lib/eventdev/rte_eventdev_core.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.119 Installing /home/vagrant/spdk_repo/dpdk/lib/gpudev/rte_gpudev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.119 Installing /home/vagrant/spdk_repo/dpdk/lib/gro/rte_gro.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.119 Installing /home/vagrant/spdk_repo/dpdk/lib/gso/rte_gso.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.119 Installing /home/vagrant/spdk_repo/dpdk/lib/ip_frag/rte_ip_frag.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.119 Installing /home/vagrant/spdk_repo/dpdk/lib/jobstats/rte_jobstats.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.119 Installing /home/vagrant/spdk_repo/dpdk/lib/latencystats/rte_latencystats.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.119 Installing /home/vagrant/spdk_repo/dpdk/lib/lpm/rte_lpm.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.119 Installing /home/vagrant/spdk_repo/dpdk/lib/lpm/rte_lpm6.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.119 Installing /home/vagrant/spdk_repo/dpdk/lib/lpm/rte_lpm_altivec.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.119 Installing /home/vagrant/spdk_repo/dpdk/lib/lpm/rte_lpm_neon.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.119 Installing /home/vagrant/spdk_repo/dpdk/lib/lpm/rte_lpm_scalar.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.119 Installing /home/vagrant/spdk_repo/dpdk/lib/lpm/rte_lpm_sse.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.119 Installing /home/vagrant/spdk_repo/dpdk/lib/lpm/rte_lpm_sve.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.119 Installing /home/vagrant/spdk_repo/dpdk/lib/member/rte_member.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.119 Installing /home/vagrant/spdk_repo/dpdk/lib/pcapng/rte_pcapng.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.119 Installing /home/vagrant/spdk_repo/dpdk/lib/power/rte_power.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.119 Installing /home/vagrant/spdk_repo/dpdk/lib/power/rte_power_empty_poll.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.119 Installing /home/vagrant/spdk_repo/dpdk/lib/power/rte_power_intel_uncore.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.119 Installing /home/vagrant/spdk_repo/dpdk/lib/power/rte_power_pmd_mgmt.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.119 Installing /home/vagrant/spdk_repo/dpdk/lib/power/rte_power_guest_channel.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.119 Installing /home/vagrant/spdk_repo/dpdk/lib/rawdev/rte_rawdev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.119 Installing /home/vagrant/spdk_repo/dpdk/lib/rawdev/rte_rawdev_pmd.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.119 Installing /home/vagrant/spdk_repo/dpdk/lib/regexdev/rte_regexdev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.119 Installing /home/vagrant/spdk_repo/dpdk/lib/regexdev/rte_regexdev_driver.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.119 Installing /home/vagrant/spdk_repo/dpdk/lib/regexdev/rte_regexdev_core.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.119 Installing /home/vagrant/spdk_repo/dpdk/lib/dmadev/rte_dmadev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.119 Installing /home/vagrant/spdk_repo/dpdk/lib/dmadev/rte_dmadev_core.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.119 Installing /home/vagrant/spdk_repo/dpdk/lib/rib/rte_rib.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.119 Installing /home/vagrant/spdk_repo/dpdk/lib/rib/rte_rib6.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.119 Installing /home/vagrant/spdk_repo/dpdk/lib/reorder/rte_reorder.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.119 Installing /home/vagrant/spdk_repo/dpdk/lib/sched/rte_approx.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.119 Installing /home/vagrant/spdk_repo/dpdk/lib/sched/rte_red.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.119 Installing /home/vagrant/spdk_repo/dpdk/lib/sched/rte_sched.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.119 Installing /home/vagrant/spdk_repo/dpdk/lib/sched/rte_sched_common.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.119 Installing /home/vagrant/spdk_repo/dpdk/lib/sched/rte_pie.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.119 Installing /home/vagrant/spdk_repo/dpdk/lib/security/rte_security.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.119 Installing /home/vagrant/spdk_repo/dpdk/lib/security/rte_security_driver.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.119 Installing /home/vagrant/spdk_repo/dpdk/lib/stack/rte_stack.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.119 Installing /home/vagrant/spdk_repo/dpdk/lib/stack/rte_stack_std.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.119 Installing /home/vagrant/spdk_repo/dpdk/lib/stack/rte_stack_lf.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.119 Installing /home/vagrant/spdk_repo/dpdk/lib/stack/rte_stack_lf_generic.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.119 Installing /home/vagrant/spdk_repo/dpdk/lib/stack/rte_stack_lf_c11.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.119 Installing /home/vagrant/spdk_repo/dpdk/lib/stack/rte_stack_lf_stubs.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.119 Installing /home/vagrant/spdk_repo/dpdk/lib/vhost/rte_vdpa.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.119 Installing /home/vagrant/spdk_repo/dpdk/lib/vhost/rte_vhost.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.119 Installing /home/vagrant/spdk_repo/dpdk/lib/vhost/rte_vhost_async.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.119 Installing /home/vagrant/spdk_repo/dpdk/lib/vhost/rte_vhost_crypto.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.119 Installing /home/vagrant/spdk_repo/dpdk/lib/ipsec/rte_ipsec.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.119 Installing /home/vagrant/spdk_repo/dpdk/lib/ipsec/rte_ipsec_sa.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.119 Installing /home/vagrant/spdk_repo/dpdk/lib/ipsec/rte_ipsec_sad.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.119 Installing /home/vagrant/spdk_repo/dpdk/lib/ipsec/rte_ipsec_group.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.119 Installing /home/vagrant/spdk_repo/dpdk/lib/fib/rte_fib.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.119 Installing /home/vagrant/spdk_repo/dpdk/lib/fib/rte_fib6.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.119 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_port_ethdev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.119 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_port_fd.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.119 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_port_frag.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.119 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_port_ras.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.119 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_port.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.119 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_port_ring.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.119 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_port_sched.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.119 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_port_source_sink.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.119 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_port_sym_crypto.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.119 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_port_eventdev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.119 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_swx_port.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.119 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_swx_port_ethdev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.119 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_swx_port_fd.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.119 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_swx_port_ring.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.119 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_swx_port_source_sink.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.119 Installing /home/vagrant/spdk_repo/dpdk/lib/pdump/rte_pdump.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.119 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_lru.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.119 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_swx_hash_func.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.119 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_swx_table.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.119 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_swx_table_em.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.119 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_swx_table_learner.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.119 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_swx_table_selector.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.119 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_swx_table_wm.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.119 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_table.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.119 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_table_acl.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.119 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_table_array.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.119 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_table_hash.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.119 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_table_hash_cuckoo.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.119 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_table_hash_func.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.119 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_table_lpm.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.119 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_table_lpm_ipv6.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.119 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_table_stub.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.119 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_lru_arm64.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.119 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_lru_x86.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.119 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_table_hash_func_arm64.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.119 Installing /home/vagrant/spdk_repo/dpdk/lib/pipeline/rte_pipeline.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.119 Installing /home/vagrant/spdk_repo/dpdk/lib/pipeline/rte_port_in_action.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.119 Installing /home/vagrant/spdk_repo/dpdk/lib/pipeline/rte_table_action.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.119 Installing /home/vagrant/spdk_repo/dpdk/lib/pipeline/rte_swx_pipeline.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.119 Installing /home/vagrant/spdk_repo/dpdk/lib/pipeline/rte_swx_extern.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.119 Installing /home/vagrant/spdk_repo/dpdk/lib/pipeline/rte_swx_ctl.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.119 Installing /home/vagrant/spdk_repo/dpdk/lib/graph/rte_graph.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.119 Installing /home/vagrant/spdk_repo/dpdk/lib/graph/rte_graph_worker.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.119 Installing /home/vagrant/spdk_repo/dpdk/lib/node/rte_node_ip4_api.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.119 Installing /home/vagrant/spdk_repo/dpdk/lib/node/rte_node_eth_api.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.119 Installing /home/vagrant/spdk_repo/dpdk/drivers/bus/pci/rte_bus_pci.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.120 Installing /home/vagrant/spdk_repo/dpdk/drivers/bus/vdev/rte_bus_vdev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.120 Installing /home/vagrant/spdk_repo/dpdk/drivers/net/i40e/rte_pmd_i40e.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.120 Installing /home/vagrant/spdk_repo/dpdk/usertools/dpdk-devbind.py to /home/vagrant/spdk_repo/dpdk/build/bin 00:02:23.120 Installing /home/vagrant/spdk_repo/dpdk/usertools/dpdk-pmdinfo.py to /home/vagrant/spdk_repo/dpdk/build/bin 00:02:23.120 Installing /home/vagrant/spdk_repo/dpdk/usertools/dpdk-telemetry.py to /home/vagrant/spdk_repo/dpdk/build/bin 00:02:23.120 Installing /home/vagrant/spdk_repo/dpdk/usertools/dpdk-hugepages.py to /home/vagrant/spdk_repo/dpdk/build/bin 00:02:23.120 Installing /home/vagrant/spdk_repo/dpdk/build-tmp/rte_build_config.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:23.120 Installing /home/vagrant/spdk_repo/dpdk/build-tmp/meson-private/libdpdk-libs.pc to /home/vagrant/spdk_repo/dpdk/build/lib/pkgconfig 00:02:23.120 Installing /home/vagrant/spdk_repo/dpdk/build-tmp/meson-private/libdpdk.pc to /home/vagrant/spdk_repo/dpdk/build/lib/pkgconfig 00:02:23.120 Installing symlink pointing to librte_kvargs.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_kvargs.so.23 00:02:23.120 Installing symlink pointing to librte_kvargs.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_kvargs.so 00:02:23.120 Installing symlink pointing to librte_telemetry.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_telemetry.so.23 00:02:23.120 Installing symlink pointing to librte_telemetry.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_telemetry.so 00:02:23.120 Installing symlink pointing to librte_eal.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_eal.so.23 00:02:23.120 Installing symlink pointing to librte_eal.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_eal.so 00:02:23.120 Installing symlink pointing to librte_ring.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_ring.so.23 00:02:23.120 Installing symlink pointing to librte_ring.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_ring.so 00:02:23.120 Installing symlink pointing to librte_rcu.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_rcu.so.23 00:02:23.120 Installing symlink pointing to librte_rcu.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_rcu.so 00:02:23.120 Installing symlink pointing to librte_mempool.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_mempool.so.23 00:02:23.120 Installing symlink pointing to librte_mempool.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_mempool.so 00:02:23.120 Installing symlink pointing to librte_mbuf.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_mbuf.so.23 00:02:23.120 Installing symlink pointing to librte_mbuf.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_mbuf.so 00:02:23.120 Installing symlink pointing to librte_net.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_net.so.23 00:02:23.120 Installing symlink pointing to librte_net.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_net.so 00:02:23.120 Installing symlink pointing to librte_meter.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_meter.so.23 00:02:23.120 Installing symlink pointing to librte_meter.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_meter.so 00:02:23.120 Installing symlink pointing to librte_ethdev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_ethdev.so.23 00:02:23.120 Installing symlink pointing to librte_ethdev.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_ethdev.so 00:02:23.120 Installing symlink pointing to librte_pci.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_pci.so.23 00:02:23.120 Installing symlink pointing to librte_pci.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_pci.so 00:02:23.120 Installing symlink pointing to librte_cmdline.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_cmdline.so.23 00:02:23.120 Installing symlink pointing to librte_cmdline.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_cmdline.so 00:02:23.120 Installing symlink pointing to librte_metrics.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_metrics.so.23 00:02:23.120 Installing symlink pointing to librte_metrics.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_metrics.so 00:02:23.120 Installing symlink pointing to librte_hash.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_hash.so.23 00:02:23.120 Installing symlink pointing to librte_hash.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_hash.so 00:02:23.120 Installing symlink pointing to librte_timer.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_timer.so.23 00:02:23.120 Installing symlink pointing to librte_timer.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_timer.so 00:02:23.120 Installing symlink pointing to librte_acl.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_acl.so.23 00:02:23.120 Installing symlink pointing to librte_acl.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_acl.so 00:02:23.120 Installing symlink pointing to librte_bbdev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_bbdev.so.23 00:02:23.120 Installing symlink pointing to librte_bbdev.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_bbdev.so 00:02:23.120 Installing symlink pointing to librte_bitratestats.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_bitratestats.so.23 00:02:23.120 Installing symlink pointing to librte_bitratestats.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_bitratestats.so 00:02:23.120 Installing symlink pointing to librte_bpf.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_bpf.so.23 00:02:23.120 Installing symlink pointing to librte_bpf.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_bpf.so 00:02:23.120 Installing symlink pointing to librte_cfgfile.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_cfgfile.so.23 00:02:23.120 Installing symlink pointing to librte_cfgfile.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_cfgfile.so 00:02:23.120 Installing symlink pointing to librte_compressdev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_compressdev.so.23 00:02:23.120 Installing symlink pointing to librte_compressdev.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_compressdev.so 00:02:23.120 Installing symlink pointing to librte_cryptodev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_cryptodev.so.23 00:02:23.120 Installing symlink pointing to librte_cryptodev.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_cryptodev.so 00:02:23.120 Installing symlink pointing to librte_distributor.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_distributor.so.23 00:02:23.120 Installing symlink pointing to librte_distributor.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_distributor.so 00:02:23.120 Installing symlink pointing to librte_efd.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_efd.so.23 00:02:23.120 Installing symlink pointing to librte_efd.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_efd.so 00:02:23.120 Installing symlink pointing to librte_eventdev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_eventdev.so.23 00:02:23.120 Installing symlink pointing to librte_eventdev.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_eventdev.so 00:02:23.120 Installing symlink pointing to librte_gpudev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_gpudev.so.23 00:02:23.120 './librte_bus_pci.so' -> 'dpdk/pmds-23.0/librte_bus_pci.so' 00:02:23.120 './librte_bus_pci.so.23' -> 'dpdk/pmds-23.0/librte_bus_pci.so.23' 00:02:23.120 './librte_bus_pci.so.23.0' -> 'dpdk/pmds-23.0/librte_bus_pci.so.23.0' 00:02:23.120 './librte_bus_vdev.so' -> 'dpdk/pmds-23.0/librte_bus_vdev.so' 00:02:23.120 './librte_bus_vdev.so.23' -> 'dpdk/pmds-23.0/librte_bus_vdev.so.23' 00:02:23.120 './librte_bus_vdev.so.23.0' -> 'dpdk/pmds-23.0/librte_bus_vdev.so.23.0' 00:02:23.120 './librte_mempool_ring.so' -> 'dpdk/pmds-23.0/librte_mempool_ring.so' 00:02:23.120 './librte_mempool_ring.so.23' -> 'dpdk/pmds-23.0/librte_mempool_ring.so.23' 00:02:23.120 './librte_mempool_ring.so.23.0' -> 'dpdk/pmds-23.0/librte_mempool_ring.so.23.0' 00:02:23.120 './librte_net_i40e.so' -> 'dpdk/pmds-23.0/librte_net_i40e.so' 00:02:23.120 './librte_net_i40e.so.23' -> 'dpdk/pmds-23.0/librte_net_i40e.so.23' 00:02:23.120 './librte_net_i40e.so.23.0' -> 'dpdk/pmds-23.0/librte_net_i40e.so.23.0' 00:02:23.120 Installing symlink pointing to librte_gpudev.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_gpudev.so 00:02:23.120 Installing symlink pointing to librte_gro.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_gro.so.23 00:02:23.120 Installing symlink pointing to librte_gro.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_gro.so 00:02:23.120 Installing symlink pointing to librte_gso.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_gso.so.23 00:02:23.120 Installing symlink pointing to librte_gso.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_gso.so 00:02:23.120 Installing symlink pointing to librte_ip_frag.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_ip_frag.so.23 00:02:23.120 Installing symlink pointing to librte_ip_frag.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_ip_frag.so 00:02:23.120 Installing symlink pointing to librte_jobstats.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_jobstats.so.23 00:02:23.120 Installing symlink pointing to librte_jobstats.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_jobstats.so 00:02:23.120 Installing symlink pointing to librte_latencystats.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_latencystats.so.23 00:02:23.120 Installing symlink pointing to librte_latencystats.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_latencystats.so 00:02:23.120 Installing symlink pointing to librte_lpm.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_lpm.so.23 00:02:23.120 Installing symlink pointing to librte_lpm.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_lpm.so 00:02:23.120 Installing symlink pointing to librte_member.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_member.so.23 00:02:23.120 Installing symlink pointing to librte_member.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_member.so 00:02:23.120 Installing symlink pointing to librte_pcapng.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_pcapng.so.23 00:02:23.120 Installing symlink pointing to librte_pcapng.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_pcapng.so 00:02:23.120 Installing symlink pointing to librte_power.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_power.so.23 00:02:23.120 Installing symlink pointing to librte_power.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_power.so 00:02:23.120 Installing symlink pointing to librte_rawdev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_rawdev.so.23 00:02:23.120 Installing symlink pointing to librte_rawdev.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_rawdev.so 00:02:23.120 Installing symlink pointing to librte_regexdev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_regexdev.so.23 00:02:23.120 Installing symlink pointing to librte_regexdev.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_regexdev.so 00:02:23.120 Installing symlink pointing to librte_dmadev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_dmadev.so.23 00:02:23.120 Installing symlink pointing to librte_dmadev.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_dmadev.so 00:02:23.120 Installing symlink pointing to librte_rib.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_rib.so.23 00:02:23.120 Installing symlink pointing to librte_rib.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_rib.so 00:02:23.120 Installing symlink pointing to librte_reorder.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_reorder.so.23 00:02:23.120 Installing symlink pointing to librte_reorder.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_reorder.so 00:02:23.120 Installing symlink pointing to librte_sched.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_sched.so.23 00:02:23.120 Installing symlink pointing to librte_sched.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_sched.so 00:02:23.120 Installing symlink pointing to librte_security.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_security.so.23 00:02:23.120 Installing symlink pointing to librte_security.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_security.so 00:02:23.120 Installing symlink pointing to librte_stack.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_stack.so.23 00:02:23.121 Installing symlink pointing to librte_stack.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_stack.so 00:02:23.121 Installing symlink pointing to librte_vhost.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_vhost.so.23 00:02:23.121 Installing symlink pointing to librte_vhost.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_vhost.so 00:02:23.121 Installing symlink pointing to librte_ipsec.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_ipsec.so.23 00:02:23.121 Installing symlink pointing to librte_ipsec.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_ipsec.so 00:02:23.121 Installing symlink pointing to librte_fib.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_fib.so.23 00:02:23.121 Installing symlink pointing to librte_fib.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_fib.so 00:02:23.121 Installing symlink pointing to librte_port.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_port.so.23 00:02:23.121 Installing symlink pointing to librte_port.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_port.so 00:02:23.121 Installing symlink pointing to librte_pdump.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_pdump.so.23 00:02:23.121 Installing symlink pointing to librte_pdump.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_pdump.so 00:02:23.121 Installing symlink pointing to librte_table.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_table.so.23 00:02:23.121 Installing symlink pointing to librte_table.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_table.so 00:02:23.121 Installing symlink pointing to librte_pipeline.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_pipeline.so.23 00:02:23.121 Installing symlink pointing to librte_pipeline.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_pipeline.so 00:02:23.121 Installing symlink pointing to librte_graph.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_graph.so.23 00:02:23.121 Installing symlink pointing to librte_graph.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_graph.so 00:02:23.121 Installing symlink pointing to librte_node.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_node.so.23 00:02:23.121 Installing symlink pointing to librte_node.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_node.so 00:02:23.121 Installing symlink pointing to librte_bus_pci.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0/librte_bus_pci.so.23 00:02:23.121 Installing symlink pointing to librte_bus_pci.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0/librte_bus_pci.so 00:02:23.121 Installing symlink pointing to librte_bus_vdev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0/librte_bus_vdev.so.23 00:02:23.121 Installing symlink pointing to librte_bus_vdev.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0/librte_bus_vdev.so 00:02:23.121 Installing symlink pointing to librte_mempool_ring.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0/librte_mempool_ring.so.23 00:02:23.121 Installing symlink pointing to librte_mempool_ring.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0/librte_mempool_ring.so 00:02:23.121 Installing symlink pointing to librte_net_i40e.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0/librte_net_i40e.so.23 00:02:23.121 Installing symlink pointing to librte_net_i40e.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0/librte_net_i40e.so 00:02:23.121 Running custom install script '/bin/sh /home/vagrant/spdk_repo/dpdk/config/../buildtools/symlink-drivers-solibs.sh lib dpdk/pmds-23.0' 00:02:23.121 23:36:11 build_native_dpdk -- common/autobuild_common.sh@220 -- $ cat 00:02:23.121 23:36:11 build_native_dpdk -- common/autobuild_common.sh@225 -- $ cd /home/vagrant/spdk_repo/spdk 00:02:23.121 00:02:23.121 real 0m42.465s 00:02:23.121 user 4m13.172s 00:02:23.121 sys 0m49.630s 00:02:23.121 23:36:11 build_native_dpdk -- common/autotest_common.sh@1130 -- $ xtrace_disable 00:02:23.121 23:36:11 build_native_dpdk -- common/autotest_common.sh@10 -- $ set +x 00:02:23.121 ************************************ 00:02:23.121 END TEST build_native_dpdk 00:02:23.121 ************************************ 00:02:23.121 23:36:11 -- spdk/autobuild.sh@31 -- $ case "$SPDK_TEST_AUTOBUILD" in 00:02:23.121 23:36:11 -- spdk/autobuild.sh@47 -- $ [[ 0 -eq 1 ]] 00:02:23.121 23:36:11 -- spdk/autobuild.sh@51 -- $ [[ 0 -eq 1 ]] 00:02:23.121 23:36:11 -- spdk/autobuild.sh@55 -- $ [[ -n '' ]] 00:02:23.121 23:36:11 -- spdk/autobuild.sh@57 -- $ [[ 0 -eq 1 ]] 00:02:23.121 23:36:11 -- spdk/autobuild.sh@59 -- $ [[ 0 -eq 1 ]] 00:02:23.121 23:36:11 -- spdk/autobuild.sh@62 -- $ [[ 0 -eq 1 ]] 00:02:23.121 23:36:11 -- spdk/autobuild.sh@67 -- $ /home/vagrant/spdk_repo/spdk/configure --enable-debug --enable-werror --with-rdma --with-idxd --with-fio=/usr/src/fio --with-iscsi-initiator --disable-unit-tests --enable-ubsan --enable-asan --enable-coverage --with-ublk --with-raid5f --with-dpdk=/home/vagrant/spdk_repo/dpdk/build --with-shared 00:02:23.380 Using /home/vagrant/spdk_repo/dpdk/build/lib/pkgconfig for additional libs... 00:02:23.380 DPDK libraries: /home/vagrant/spdk_repo/dpdk/build/lib 00:02:23.380 DPDK includes: //home/vagrant/spdk_repo/dpdk/build/include 00:02:23.380 Using default SPDK env in /home/vagrant/spdk_repo/spdk/lib/env_dpdk 00:02:23.957 Using 'verbs' RDMA provider 00:02:39.779 Configuring ISA-L (logfile: /home/vagrant/spdk_repo/spdk/.spdk-isal.log)...done. 00:02:54.666 Configuring ISA-L-crypto (logfile: /home/vagrant/spdk_repo/spdk/.spdk-isal-crypto.log)...done. 00:02:54.666 Creating mk/config.mk...done. 00:02:54.666 Creating mk/cc.flags.mk...done. 00:02:54.666 Type 'make' to build. 00:02:54.666 23:36:42 -- spdk/autobuild.sh@70 -- $ run_test make make -j10 00:02:54.666 23:36:42 -- common/autotest_common.sh@1105 -- $ '[' 3 -le 1 ']' 00:02:54.666 23:36:42 -- common/autotest_common.sh@1111 -- $ xtrace_disable 00:02:54.666 23:36:42 -- common/autotest_common.sh@10 -- $ set +x 00:02:54.666 ************************************ 00:02:54.666 START TEST make 00:02:54.666 ************************************ 00:02:54.666 23:36:42 make -- common/autotest_common.sh@1129 -- $ make -j10 00:02:54.666 make[1]: Nothing to be done for 'all'. 00:03:41.381 CC lib/ut/ut.o 00:03:41.381 CC lib/ut_mock/mock.o 00:03:41.381 CC lib/log/log.o 00:03:41.381 CC lib/log/log_flags.o 00:03:41.381 CC lib/log/log_deprecated.o 00:03:41.381 LIB libspdk_ut.a 00:03:41.381 SO libspdk_ut.so.2.0 00:03:41.381 LIB libspdk_log.a 00:03:41.381 LIB libspdk_ut_mock.a 00:03:41.381 SO libspdk_ut_mock.so.6.0 00:03:41.381 SYMLINK libspdk_ut.so 00:03:41.381 SO libspdk_log.so.7.1 00:03:41.381 SYMLINK libspdk_log.so 00:03:41.381 SYMLINK libspdk_ut_mock.so 00:03:41.381 CC lib/ioat/ioat.o 00:03:41.381 CC lib/dma/dma.o 00:03:41.381 CXX lib/trace_parser/trace.o 00:03:41.381 CC lib/util/base64.o 00:03:41.381 CC lib/util/bit_array.o 00:03:41.381 CC lib/util/crc32c.o 00:03:41.381 CC lib/util/cpuset.o 00:03:41.381 CC lib/util/crc16.o 00:03:41.381 CC lib/util/crc32.o 00:03:41.381 CC lib/vfio_user/host/vfio_user_pci.o 00:03:41.381 CC lib/util/crc32_ieee.o 00:03:41.381 CC lib/util/crc64.o 00:03:41.381 CC lib/vfio_user/host/vfio_user.o 00:03:41.381 LIB libspdk_dma.a 00:03:41.381 CC lib/util/dif.o 00:03:41.381 SO libspdk_dma.so.5.0 00:03:41.381 CC lib/util/fd.o 00:03:41.381 CC lib/util/fd_group.o 00:03:41.381 SYMLINK libspdk_dma.so 00:03:41.381 CC lib/util/file.o 00:03:41.381 LIB libspdk_ioat.a 00:03:41.381 CC lib/util/hexlify.o 00:03:41.381 CC lib/util/iov.o 00:03:41.381 SO libspdk_ioat.so.7.0 00:03:41.381 SYMLINK libspdk_ioat.so 00:03:41.381 CC lib/util/math.o 00:03:41.381 CC lib/util/net.o 00:03:41.381 CC lib/util/pipe.o 00:03:41.381 LIB libspdk_vfio_user.a 00:03:41.381 CC lib/util/strerror_tls.o 00:03:41.381 SO libspdk_vfio_user.so.5.0 00:03:41.381 CC lib/util/string.o 00:03:41.381 CC lib/util/uuid.o 00:03:41.381 SYMLINK libspdk_vfio_user.so 00:03:41.381 CC lib/util/xor.o 00:03:41.381 CC lib/util/zipf.o 00:03:41.381 CC lib/util/md5.o 00:03:41.381 LIB libspdk_util.a 00:03:41.381 LIB libspdk_trace_parser.a 00:03:41.381 SO libspdk_trace_parser.so.6.0 00:03:41.381 SO libspdk_util.so.10.1 00:03:41.381 SYMLINK libspdk_trace_parser.so 00:03:41.381 SYMLINK libspdk_util.so 00:03:41.381 CC lib/conf/conf.o 00:03:41.381 CC lib/vmd/vmd.o 00:03:41.381 CC lib/vmd/led.o 00:03:41.381 CC lib/env_dpdk/env.o 00:03:41.381 CC lib/json/json_parse.o 00:03:41.381 CC lib/env_dpdk/pci.o 00:03:41.381 CC lib/env_dpdk/memory.o 00:03:41.381 CC lib/idxd/idxd.o 00:03:41.381 CC lib/idxd/idxd_user.o 00:03:41.381 CC lib/rdma_utils/rdma_utils.o 00:03:41.381 CC lib/idxd/idxd_kernel.o 00:03:41.381 LIB libspdk_conf.a 00:03:41.381 SO libspdk_conf.so.6.0 00:03:41.381 CC lib/json/json_util.o 00:03:41.381 CC lib/json/json_write.o 00:03:41.381 LIB libspdk_rdma_utils.a 00:03:41.381 SYMLINK libspdk_conf.so 00:03:41.381 CC lib/env_dpdk/init.o 00:03:41.381 CC lib/env_dpdk/threads.o 00:03:41.381 SO libspdk_rdma_utils.so.1.0 00:03:41.381 SYMLINK libspdk_rdma_utils.so 00:03:41.381 CC lib/env_dpdk/pci_ioat.o 00:03:41.381 CC lib/env_dpdk/pci_virtio.o 00:03:41.381 CC lib/env_dpdk/pci_vmd.o 00:03:41.381 CC lib/env_dpdk/pci_idxd.o 00:03:41.381 CC lib/env_dpdk/pci_event.o 00:03:41.381 CC lib/env_dpdk/sigbus_handler.o 00:03:41.381 CC lib/env_dpdk/pci_dpdk.o 00:03:41.381 LIB libspdk_json.a 00:03:41.381 CC lib/env_dpdk/pci_dpdk_2207.o 00:03:41.381 SO libspdk_json.so.6.0 00:03:41.381 CC lib/rdma_provider/common.o 00:03:41.381 CC lib/rdma_provider/rdma_provider_verbs.o 00:03:41.381 CC lib/env_dpdk/pci_dpdk_2211.o 00:03:41.381 SYMLINK libspdk_json.so 00:03:41.381 LIB libspdk_idxd.a 00:03:41.381 SO libspdk_idxd.so.12.1 00:03:41.381 LIB libspdk_vmd.a 00:03:41.381 SO libspdk_vmd.so.6.0 00:03:41.381 SYMLINK libspdk_idxd.so 00:03:41.381 SYMLINK libspdk_vmd.so 00:03:41.381 LIB libspdk_rdma_provider.a 00:03:41.381 CC lib/jsonrpc/jsonrpc_server.o 00:03:41.381 CC lib/jsonrpc/jsonrpc_server_tcp.o 00:03:41.381 CC lib/jsonrpc/jsonrpc_client.o 00:03:41.381 CC lib/jsonrpc/jsonrpc_client_tcp.o 00:03:41.381 SO libspdk_rdma_provider.so.7.0 00:03:41.381 SYMLINK libspdk_rdma_provider.so 00:03:41.381 LIB libspdk_jsonrpc.a 00:03:41.381 SO libspdk_jsonrpc.so.6.0 00:03:41.381 SYMLINK libspdk_jsonrpc.so 00:03:41.381 LIB libspdk_env_dpdk.a 00:03:41.381 CC lib/rpc/rpc.o 00:03:41.381 SO libspdk_env_dpdk.so.15.1 00:03:41.381 SYMLINK libspdk_env_dpdk.so 00:03:41.381 LIB libspdk_rpc.a 00:03:41.381 SO libspdk_rpc.so.6.0 00:03:41.381 SYMLINK libspdk_rpc.so 00:03:41.381 CC lib/notify/notify.o 00:03:41.381 CC lib/notify/notify_rpc.o 00:03:41.381 CC lib/trace/trace.o 00:03:41.381 CC lib/trace/trace_flags.o 00:03:41.381 CC lib/trace/trace_rpc.o 00:03:41.381 CC lib/keyring/keyring.o 00:03:41.381 CC lib/keyring/keyring_rpc.o 00:03:41.381 LIB libspdk_notify.a 00:03:41.381 SO libspdk_notify.so.6.0 00:03:41.381 SYMLINK libspdk_notify.so 00:03:41.381 LIB libspdk_keyring.a 00:03:41.381 LIB libspdk_trace.a 00:03:41.381 SO libspdk_keyring.so.2.0 00:03:41.381 SO libspdk_trace.so.11.0 00:03:41.381 SYMLINK libspdk_keyring.so 00:03:41.381 SYMLINK libspdk_trace.so 00:03:41.381 CC lib/thread/thread.o 00:03:41.381 CC lib/thread/iobuf.o 00:03:41.381 CC lib/sock/sock.o 00:03:41.381 CC lib/sock/sock_rpc.o 00:03:41.949 LIB libspdk_sock.a 00:03:41.949 SO libspdk_sock.so.10.0 00:03:41.949 SYMLINK libspdk_sock.so 00:03:42.209 CC lib/nvme/nvme_ctrlr_cmd.o 00:03:42.469 CC lib/nvme/nvme_ctrlr.o 00:03:42.469 CC lib/nvme/nvme_fabric.o 00:03:42.469 CC lib/nvme/nvme_ns_cmd.o 00:03:42.469 CC lib/nvme/nvme_pcie_common.o 00:03:42.469 CC lib/nvme/nvme_ns.o 00:03:42.469 CC lib/nvme/nvme_pcie.o 00:03:42.469 CC lib/nvme/nvme_qpair.o 00:03:42.469 CC lib/nvme/nvme.o 00:03:43.036 CC lib/nvme/nvme_quirks.o 00:03:43.036 CC lib/nvme/nvme_transport.o 00:03:43.037 LIB libspdk_thread.a 00:03:43.037 CC lib/nvme/nvme_discovery.o 00:03:43.037 CC lib/nvme/nvme_ctrlr_ocssd_cmd.o 00:03:43.037 SO libspdk_thread.so.11.0 00:03:43.037 CC lib/nvme/nvme_ns_ocssd_cmd.o 00:03:43.037 SYMLINK libspdk_thread.so 00:03:43.295 CC lib/nvme/nvme_tcp.o 00:03:43.295 CC lib/nvme/nvme_opal.o 00:03:43.295 CC lib/accel/accel.o 00:03:43.295 CC lib/accel/accel_rpc.o 00:03:43.554 CC lib/accel/accel_sw.o 00:03:43.554 CC lib/nvme/nvme_io_msg.o 00:03:43.554 CC lib/nvme/nvme_poll_group.o 00:03:43.554 CC lib/nvme/nvme_zns.o 00:03:43.813 CC lib/blob/blobstore.o 00:03:43.813 CC lib/init/json_config.o 00:03:43.813 CC lib/blob/request.o 00:03:43.813 CC lib/virtio/virtio.o 00:03:44.071 CC lib/blob/zeroes.o 00:03:44.071 CC lib/init/subsystem.o 00:03:44.071 CC lib/nvme/nvme_stubs.o 00:03:44.071 CC lib/blob/blob_bs_dev.o 00:03:44.071 CC lib/virtio/virtio_vhost_user.o 00:03:44.329 CC lib/virtio/virtio_vfio_user.o 00:03:44.329 CC lib/virtio/virtio_pci.o 00:03:44.329 CC lib/init/subsystem_rpc.o 00:03:44.329 CC lib/init/rpc.o 00:03:44.329 CC lib/nvme/nvme_cuse.o 00:03:44.329 CC lib/nvme/nvme_auth.o 00:03:44.589 LIB libspdk_accel.a 00:03:44.589 CC lib/nvme/nvme_rdma.o 00:03:44.589 LIB libspdk_init.a 00:03:44.589 SO libspdk_init.so.6.0 00:03:44.589 SO libspdk_accel.so.16.0 00:03:44.589 LIB libspdk_virtio.a 00:03:44.589 SO libspdk_virtio.so.7.0 00:03:44.589 SYMLINK libspdk_init.so 00:03:44.589 SYMLINK libspdk_accel.so 00:03:44.589 SYMLINK libspdk_virtio.so 00:03:44.849 CC lib/fsdev/fsdev.o 00:03:44.849 CC lib/fsdev/fsdev_io.o 00:03:44.849 CC lib/fsdev/fsdev_rpc.o 00:03:44.849 CC lib/event/reactor.o 00:03:44.849 CC lib/event/app.o 00:03:44.849 CC lib/bdev/bdev.o 00:03:44.849 CC lib/bdev/bdev_rpc.o 00:03:45.111 CC lib/bdev/bdev_zone.o 00:03:45.111 CC lib/bdev/part.o 00:03:45.111 CC lib/event/log_rpc.o 00:03:45.111 CC lib/event/app_rpc.o 00:03:45.111 CC lib/event/scheduler_static.o 00:03:45.387 CC lib/bdev/scsi_nvme.o 00:03:45.387 LIB libspdk_fsdev.a 00:03:45.387 SO libspdk_fsdev.so.2.0 00:03:45.387 LIB libspdk_event.a 00:03:45.387 SYMLINK libspdk_fsdev.so 00:03:45.387 SO libspdk_event.so.14.0 00:03:45.662 SYMLINK libspdk_event.so 00:03:45.921 CC lib/fuse_dispatcher/fuse_dispatcher.o 00:03:45.921 LIB libspdk_nvme.a 00:03:46.181 SO libspdk_nvme.so.15.0 00:03:46.441 SYMLINK libspdk_nvme.so 00:03:46.441 LIB libspdk_fuse_dispatcher.a 00:03:46.441 SO libspdk_fuse_dispatcher.so.1.0 00:03:46.702 SYMLINK libspdk_fuse_dispatcher.so 00:03:47.273 LIB libspdk_blob.a 00:03:47.273 SO libspdk_blob.so.12.0 00:03:47.533 SYMLINK libspdk_blob.so 00:03:47.533 LIB libspdk_bdev.a 00:03:47.533 SO libspdk_bdev.so.17.0 00:03:47.793 SYMLINK libspdk_bdev.so 00:03:47.793 CC lib/lvol/lvol.o 00:03:47.793 CC lib/blobfs/tree.o 00:03:47.793 CC lib/blobfs/blobfs.o 00:03:47.793 CC lib/ftl/ftl_core.o 00:03:47.793 CC lib/ftl/ftl_init.o 00:03:47.793 CC lib/ftl/ftl_layout.o 00:03:47.793 CC lib/nbd/nbd.o 00:03:47.793 CC lib/scsi/dev.o 00:03:47.793 CC lib/ublk/ublk.o 00:03:47.793 CC lib/nvmf/ctrlr.o 00:03:48.053 CC lib/scsi/lun.o 00:03:48.053 CC lib/ftl/ftl_debug.o 00:03:48.053 CC lib/ftl/ftl_io.o 00:03:48.313 CC lib/ftl/ftl_sb.o 00:03:48.313 CC lib/scsi/port.o 00:03:48.313 CC lib/ftl/ftl_l2p.o 00:03:48.313 CC lib/scsi/scsi.o 00:03:48.313 CC lib/nbd/nbd_rpc.o 00:03:48.313 CC lib/scsi/scsi_bdev.o 00:03:48.313 CC lib/scsi/scsi_pr.o 00:03:48.313 CC lib/ftl/ftl_l2p_flat.o 00:03:48.313 CC lib/scsi/scsi_rpc.o 00:03:48.313 CC lib/scsi/task.o 00:03:48.573 LIB libspdk_nbd.a 00:03:48.573 SO libspdk_nbd.so.7.0 00:03:48.573 SYMLINK libspdk_nbd.so 00:03:48.573 CC lib/ublk/ublk_rpc.o 00:03:48.573 CC lib/nvmf/ctrlr_discovery.o 00:03:48.573 CC lib/ftl/ftl_nv_cache.o 00:03:48.573 CC lib/nvmf/ctrlr_bdev.o 00:03:48.573 LIB libspdk_blobfs.a 00:03:48.573 CC lib/nvmf/subsystem.o 00:03:48.573 SO libspdk_blobfs.so.11.0 00:03:48.573 CC lib/nvmf/nvmf.o 00:03:48.573 LIB libspdk_ublk.a 00:03:48.833 SYMLINK libspdk_blobfs.so 00:03:48.833 CC lib/ftl/ftl_band.o 00:03:48.833 SO libspdk_ublk.so.3.0 00:03:48.833 SYMLINK libspdk_ublk.so 00:03:48.833 CC lib/nvmf/nvmf_rpc.o 00:03:48.833 LIB libspdk_lvol.a 00:03:48.833 LIB libspdk_scsi.a 00:03:48.833 SO libspdk_lvol.so.11.0 00:03:48.833 SO libspdk_scsi.so.9.0 00:03:48.833 SYMLINK libspdk_lvol.so 00:03:48.833 CC lib/ftl/ftl_band_ops.o 00:03:49.093 SYMLINK libspdk_scsi.so 00:03:49.093 CC lib/ftl/ftl_writer.o 00:03:49.093 CC lib/ftl/ftl_rq.o 00:03:49.093 CC lib/ftl/ftl_reloc.o 00:03:49.093 CC lib/ftl/ftl_l2p_cache.o 00:03:49.093 CC lib/ftl/ftl_p2l.o 00:03:49.352 CC lib/nvmf/transport.o 00:03:49.352 CC lib/iscsi/conn.o 00:03:49.352 CC lib/iscsi/init_grp.o 00:03:49.611 CC lib/iscsi/iscsi.o 00:03:49.611 CC lib/iscsi/param.o 00:03:49.611 CC lib/nvmf/tcp.o 00:03:49.611 CC lib/nvmf/stubs.o 00:03:49.611 CC lib/ftl/ftl_p2l_log.o 00:03:49.871 CC lib/ftl/mngt/ftl_mngt.o 00:03:49.871 CC lib/nvmf/mdns_server.o 00:03:49.871 CC lib/iscsi/portal_grp.o 00:03:49.871 CC lib/nvmf/rdma.o 00:03:50.129 CC lib/iscsi/tgt_node.o 00:03:50.129 CC lib/iscsi/iscsi_subsystem.o 00:03:50.129 CC lib/ftl/mngt/ftl_mngt_bdev.o 00:03:50.129 CC lib/ftl/mngt/ftl_mngt_shutdown.o 00:03:50.129 CC lib/ftl/mngt/ftl_mngt_startup.o 00:03:50.129 CC lib/ftl/mngt/ftl_mngt_md.o 00:03:50.129 CC lib/ftl/mngt/ftl_mngt_misc.o 00:03:50.388 CC lib/ftl/mngt/ftl_mngt_ioch.o 00:03:50.388 CC lib/iscsi/iscsi_rpc.o 00:03:50.388 CC lib/vhost/vhost.o 00:03:50.388 CC lib/vhost/vhost_rpc.o 00:03:50.388 CC lib/ftl/mngt/ftl_mngt_l2p.o 00:03:50.388 CC lib/vhost/vhost_scsi.o 00:03:50.388 CC lib/vhost/vhost_blk.o 00:03:50.388 CC lib/vhost/rte_vhost_user.o 00:03:50.646 CC lib/ftl/mngt/ftl_mngt_band.o 00:03:50.646 CC lib/nvmf/auth.o 00:03:50.904 CC lib/ftl/mngt/ftl_mngt_self_test.o 00:03:50.904 CC lib/ftl/mngt/ftl_mngt_p2l.o 00:03:51.163 CC lib/ftl/mngt/ftl_mngt_recovery.o 00:03:51.164 CC lib/iscsi/task.o 00:03:51.164 CC lib/ftl/mngt/ftl_mngt_upgrade.o 00:03:51.164 CC lib/ftl/utils/ftl_conf.o 00:03:51.424 LIB libspdk_iscsi.a 00:03:51.424 CC lib/ftl/utils/ftl_md.o 00:03:51.424 CC lib/ftl/utils/ftl_mempool.o 00:03:51.424 CC lib/ftl/utils/ftl_bitmap.o 00:03:51.424 CC lib/ftl/utils/ftl_property.o 00:03:51.424 CC lib/ftl/utils/ftl_layout_tracker_bdev.o 00:03:51.424 SO libspdk_iscsi.so.8.0 00:03:51.424 LIB libspdk_vhost.a 00:03:51.424 CC lib/ftl/upgrade/ftl_layout_upgrade.o 00:03:51.424 CC lib/ftl/upgrade/ftl_sb_upgrade.o 00:03:51.424 SO libspdk_vhost.so.8.0 00:03:51.424 CC lib/ftl/upgrade/ftl_p2l_upgrade.o 00:03:51.424 CC lib/ftl/upgrade/ftl_band_upgrade.o 00:03:51.424 SYMLINK libspdk_iscsi.so 00:03:51.424 CC lib/ftl/upgrade/ftl_chunk_upgrade.o 00:03:51.684 CC lib/ftl/upgrade/ftl_trim_upgrade.o 00:03:51.684 CC lib/ftl/upgrade/ftl_sb_v3.o 00:03:51.684 SYMLINK libspdk_vhost.so 00:03:51.684 CC lib/ftl/upgrade/ftl_sb_v5.o 00:03:51.684 CC lib/ftl/nvc/ftl_nvc_dev.o 00:03:51.684 CC lib/ftl/nvc/ftl_nvc_bdev_vss.o 00:03:51.684 CC lib/ftl/nvc/ftl_nvc_bdev_non_vss.o 00:03:51.684 CC lib/ftl/nvc/ftl_nvc_bdev_common.o 00:03:51.684 CC lib/ftl/base/ftl_base_dev.o 00:03:51.684 CC lib/ftl/base/ftl_base_bdev.o 00:03:51.684 CC lib/ftl/ftl_trace.o 00:03:51.944 LIB libspdk_ftl.a 00:03:52.204 SO libspdk_ftl.so.9.0 00:03:52.204 LIB libspdk_nvmf.a 00:03:52.462 SO libspdk_nvmf.so.20.0 00:03:52.462 SYMLINK libspdk_ftl.so 00:03:52.721 SYMLINK libspdk_nvmf.so 00:03:52.980 CC module/env_dpdk/env_dpdk_rpc.o 00:03:53.240 CC module/scheduler/dynamic/scheduler_dynamic.o 00:03:53.240 CC module/fsdev/aio/fsdev_aio.o 00:03:53.240 CC module/keyring/file/keyring.o 00:03:53.240 CC module/keyring/linux/keyring.o 00:03:53.240 CC module/accel/error/accel_error.o 00:03:53.240 CC module/scheduler/dpdk_governor/dpdk_governor.o 00:03:53.240 CC module/blob/bdev/blob_bdev.o 00:03:53.240 CC module/sock/posix/posix.o 00:03:53.240 CC module/scheduler/gscheduler/gscheduler.o 00:03:53.240 LIB libspdk_env_dpdk_rpc.a 00:03:53.240 SO libspdk_env_dpdk_rpc.so.6.0 00:03:53.240 SYMLINK libspdk_env_dpdk_rpc.so 00:03:53.240 CC module/fsdev/aio/fsdev_aio_rpc.o 00:03:53.240 CC module/keyring/file/keyring_rpc.o 00:03:53.240 CC module/keyring/linux/keyring_rpc.o 00:03:53.240 LIB libspdk_scheduler_dpdk_governor.a 00:03:53.240 LIB libspdk_scheduler_gscheduler.a 00:03:53.240 SO libspdk_scheduler_dpdk_governor.so.4.0 00:03:53.240 SO libspdk_scheduler_gscheduler.so.4.0 00:03:53.240 LIB libspdk_scheduler_dynamic.a 00:03:53.500 CC module/accel/error/accel_error_rpc.o 00:03:53.500 SO libspdk_scheduler_dynamic.so.4.0 00:03:53.500 SYMLINK libspdk_scheduler_gscheduler.so 00:03:53.500 SYMLINK libspdk_scheduler_dpdk_governor.so 00:03:53.500 LIB libspdk_keyring_file.a 00:03:53.500 SYMLINK libspdk_scheduler_dynamic.so 00:03:53.500 CC module/fsdev/aio/linux_aio_mgr.o 00:03:53.500 LIB libspdk_blob_bdev.a 00:03:53.500 LIB libspdk_keyring_linux.a 00:03:53.500 SO libspdk_keyring_file.so.2.0 00:03:53.500 SO libspdk_blob_bdev.so.12.0 00:03:53.500 SO libspdk_keyring_linux.so.1.0 00:03:53.500 SYMLINK libspdk_keyring_file.so 00:03:53.500 LIB libspdk_accel_error.a 00:03:53.500 SYMLINK libspdk_blob_bdev.so 00:03:53.500 SYMLINK libspdk_keyring_linux.so 00:03:53.500 SO libspdk_accel_error.so.2.0 00:03:53.500 CC module/accel/ioat/accel_ioat.o 00:03:53.500 CC module/accel/ioat/accel_ioat_rpc.o 00:03:53.500 CC module/accel/iaa/accel_iaa.o 00:03:53.500 CC module/accel/dsa/accel_dsa.o 00:03:53.500 SYMLINK libspdk_accel_error.so 00:03:53.500 CC module/accel/iaa/accel_iaa_rpc.o 00:03:53.760 LIB libspdk_accel_ioat.a 00:03:53.760 CC module/bdev/delay/vbdev_delay.o 00:03:53.760 CC module/blobfs/bdev/blobfs_bdev.o 00:03:53.760 LIB libspdk_accel_iaa.a 00:03:53.760 SO libspdk_accel_ioat.so.6.0 00:03:53.760 CC module/bdev/error/vbdev_error.o 00:03:53.760 SO libspdk_accel_iaa.so.3.0 00:03:53.760 SYMLINK libspdk_accel_ioat.so 00:03:53.760 CC module/bdev/error/vbdev_error_rpc.o 00:03:53.760 CC module/bdev/gpt/gpt.o 00:03:53.760 CC module/accel/dsa/accel_dsa_rpc.o 00:03:53.760 LIB libspdk_fsdev_aio.a 00:03:53.760 CC module/bdev/lvol/vbdev_lvol.o 00:03:53.760 SYMLINK libspdk_accel_iaa.so 00:03:53.760 CC module/blobfs/bdev/blobfs_bdev_rpc.o 00:03:54.021 SO libspdk_fsdev_aio.so.1.0 00:03:54.021 CC module/bdev/gpt/vbdev_gpt.o 00:03:54.021 LIB libspdk_sock_posix.a 00:03:54.021 LIB libspdk_accel_dsa.a 00:03:54.021 SYMLINK libspdk_fsdev_aio.so 00:03:54.021 CC module/bdev/lvol/vbdev_lvol_rpc.o 00:03:54.021 SO libspdk_accel_dsa.so.5.0 00:03:54.021 SO libspdk_sock_posix.so.6.0 00:03:54.021 LIB libspdk_blobfs_bdev.a 00:03:54.021 LIB libspdk_bdev_error.a 00:03:54.021 SO libspdk_blobfs_bdev.so.6.0 00:03:54.021 SYMLINK libspdk_accel_dsa.so 00:03:54.021 SYMLINK libspdk_sock_posix.so 00:03:54.021 CC module/bdev/delay/vbdev_delay_rpc.o 00:03:54.021 SO libspdk_bdev_error.so.6.0 00:03:54.021 SYMLINK libspdk_blobfs_bdev.so 00:03:54.021 CC module/bdev/malloc/bdev_malloc.o 00:03:54.021 CC module/bdev/malloc/bdev_malloc_rpc.o 00:03:54.282 SYMLINK libspdk_bdev_error.so 00:03:54.282 CC module/bdev/null/bdev_null.o 00:03:54.282 LIB libspdk_bdev_gpt.a 00:03:54.282 CC module/bdev/nvme/bdev_nvme.o 00:03:54.282 SO libspdk_bdev_gpt.so.6.0 00:03:54.283 LIB libspdk_bdev_delay.a 00:03:54.283 CC module/bdev/passthru/vbdev_passthru.o 00:03:54.283 SO libspdk_bdev_delay.so.6.0 00:03:54.283 CC module/bdev/null/bdev_null_rpc.o 00:03:54.283 CC module/bdev/raid/bdev_raid.o 00:03:54.283 SYMLINK libspdk_bdev_gpt.so 00:03:54.283 CC module/bdev/raid/bdev_raid_rpc.o 00:03:54.283 SYMLINK libspdk_bdev_delay.so 00:03:54.283 CC module/bdev/raid/bdev_raid_sb.o 00:03:54.283 CC module/bdev/raid/raid0.o 00:03:54.542 LIB libspdk_bdev_lvol.a 00:03:54.542 CC module/bdev/nvme/bdev_nvme_rpc.o 00:03:54.542 SO libspdk_bdev_lvol.so.6.0 00:03:54.542 LIB libspdk_bdev_null.a 00:03:54.542 SO libspdk_bdev_null.so.6.0 00:03:54.542 LIB libspdk_bdev_malloc.a 00:03:54.542 CC module/bdev/raid/raid1.o 00:03:54.542 SYMLINK libspdk_bdev_lvol.so 00:03:54.542 CC module/bdev/raid/concat.o 00:03:54.542 SO libspdk_bdev_malloc.so.6.0 00:03:54.542 SYMLINK libspdk_bdev_null.so 00:03:54.542 CC module/bdev/passthru/vbdev_passthru_rpc.o 00:03:54.542 CC module/bdev/raid/raid5f.o 00:03:54.542 SYMLINK libspdk_bdev_malloc.so 00:03:54.542 CC module/bdev/nvme/nvme_rpc.o 00:03:54.801 LIB libspdk_bdev_passthru.a 00:03:54.801 CC module/bdev/split/vbdev_split.o 00:03:54.801 SO libspdk_bdev_passthru.so.6.0 00:03:54.802 CC module/bdev/zone_block/vbdev_zone_block.o 00:03:54.802 CC module/bdev/zone_block/vbdev_zone_block_rpc.o 00:03:54.802 CC module/bdev/nvme/bdev_mdns_client.o 00:03:54.802 SYMLINK libspdk_bdev_passthru.so 00:03:54.802 CC module/bdev/nvme/vbdev_opal.o 00:03:54.802 CC module/bdev/nvme/vbdev_opal_rpc.o 00:03:54.802 CC module/bdev/aio/bdev_aio.o 00:03:54.802 CC module/bdev/aio/bdev_aio_rpc.o 00:03:55.076 CC module/bdev/split/vbdev_split_rpc.o 00:03:55.076 LIB libspdk_bdev_zone_block.a 00:03:55.076 CC module/bdev/nvme/bdev_nvme_cuse_rpc.o 00:03:55.076 SO libspdk_bdev_zone_block.so.6.0 00:03:55.076 LIB libspdk_bdev_split.a 00:03:55.076 SYMLINK libspdk_bdev_zone_block.so 00:03:55.076 SO libspdk_bdev_split.so.6.0 00:03:55.076 CC module/bdev/ftl/bdev_ftl.o 00:03:55.076 CC module/bdev/ftl/bdev_ftl_rpc.o 00:03:55.339 SYMLINK libspdk_bdev_split.so 00:03:55.339 CC module/bdev/iscsi/bdev_iscsi.o 00:03:55.339 CC module/bdev/iscsi/bdev_iscsi_rpc.o 00:03:55.339 LIB libspdk_bdev_aio.a 00:03:55.339 CC module/bdev/virtio/bdev_virtio_scsi.o 00:03:55.339 CC module/bdev/virtio/bdev_virtio_blk.o 00:03:55.339 CC module/bdev/virtio/bdev_virtio_rpc.o 00:03:55.339 SO libspdk_bdev_aio.so.6.0 00:03:55.339 LIB libspdk_bdev_raid.a 00:03:55.339 SYMLINK libspdk_bdev_aio.so 00:03:55.339 SO libspdk_bdev_raid.so.6.0 00:03:55.339 LIB libspdk_bdev_ftl.a 00:03:55.599 SO libspdk_bdev_ftl.so.6.0 00:03:55.599 SYMLINK libspdk_bdev_raid.so 00:03:55.599 SYMLINK libspdk_bdev_ftl.so 00:03:55.599 LIB libspdk_bdev_iscsi.a 00:03:55.599 SO libspdk_bdev_iscsi.so.6.0 00:03:55.599 SYMLINK libspdk_bdev_iscsi.so 00:03:55.859 LIB libspdk_bdev_virtio.a 00:03:55.859 SO libspdk_bdev_virtio.so.6.0 00:03:55.859 SYMLINK libspdk_bdev_virtio.so 00:03:56.800 LIB libspdk_bdev_nvme.a 00:03:57.060 SO libspdk_bdev_nvme.so.7.1 00:03:57.060 SYMLINK libspdk_bdev_nvme.so 00:03:57.631 CC module/event/subsystems/iobuf/iobuf.o 00:03:57.631 CC module/event/subsystems/iobuf/iobuf_rpc.o 00:03:57.631 CC module/event/subsystems/vmd/vmd.o 00:03:57.631 CC module/event/subsystems/vmd/vmd_rpc.o 00:03:57.631 CC module/event/subsystems/vhost_blk/vhost_blk.o 00:03:57.631 CC module/event/subsystems/sock/sock.o 00:03:57.631 CC module/event/subsystems/fsdev/fsdev.o 00:03:57.631 CC module/event/subsystems/scheduler/scheduler.o 00:03:57.631 CC module/event/subsystems/keyring/keyring.o 00:03:57.891 LIB libspdk_event_keyring.a 00:03:57.891 LIB libspdk_event_vhost_blk.a 00:03:57.891 LIB libspdk_event_sock.a 00:03:57.891 LIB libspdk_event_scheduler.a 00:03:57.891 LIB libspdk_event_fsdev.a 00:03:57.891 LIB libspdk_event_vmd.a 00:03:57.891 SO libspdk_event_keyring.so.1.0 00:03:57.891 LIB libspdk_event_iobuf.a 00:03:57.891 SO libspdk_event_vhost_blk.so.3.0 00:03:57.891 SO libspdk_event_scheduler.so.4.0 00:03:57.891 SO libspdk_event_sock.so.5.0 00:03:57.891 SO libspdk_event_fsdev.so.1.0 00:03:57.891 SO libspdk_event_vmd.so.6.0 00:03:57.891 SO libspdk_event_iobuf.so.3.0 00:03:57.891 SYMLINK libspdk_event_keyring.so 00:03:57.891 SYMLINK libspdk_event_vhost_blk.so 00:03:57.891 SYMLINK libspdk_event_scheduler.so 00:03:57.891 SYMLINK libspdk_event_fsdev.so 00:03:57.891 SYMLINK libspdk_event_sock.so 00:03:57.891 SYMLINK libspdk_event_vmd.so 00:03:57.891 SYMLINK libspdk_event_iobuf.so 00:03:58.152 CC module/event/subsystems/accel/accel.o 00:03:58.412 LIB libspdk_event_accel.a 00:03:58.412 SO libspdk_event_accel.so.6.0 00:03:58.673 SYMLINK libspdk_event_accel.so 00:03:58.940 CC module/event/subsystems/bdev/bdev.o 00:03:59.207 LIB libspdk_event_bdev.a 00:03:59.207 SO libspdk_event_bdev.so.6.0 00:03:59.207 SYMLINK libspdk_event_bdev.so 00:03:59.467 CC module/event/subsystems/nvmf/nvmf_rpc.o 00:03:59.467 CC module/event/subsystems/nvmf/nvmf_tgt.o 00:03:59.467 CC module/event/subsystems/ublk/ublk.o 00:03:59.467 CC module/event/subsystems/nbd/nbd.o 00:03:59.467 CC module/event/subsystems/scsi/scsi.o 00:03:59.726 LIB libspdk_event_nvmf.a 00:03:59.726 LIB libspdk_event_nbd.a 00:03:59.726 LIB libspdk_event_scsi.a 00:03:59.726 LIB libspdk_event_ublk.a 00:03:59.726 SO libspdk_event_nbd.so.6.0 00:03:59.726 SO libspdk_event_nvmf.so.6.0 00:03:59.726 SO libspdk_event_scsi.so.6.0 00:03:59.726 SO libspdk_event_ublk.so.3.0 00:03:59.726 SYMLINK libspdk_event_nbd.so 00:03:59.986 SYMLINK libspdk_event_nvmf.so 00:03:59.986 SYMLINK libspdk_event_scsi.so 00:03:59.986 SYMLINK libspdk_event_ublk.so 00:04:00.245 CC module/event/subsystems/iscsi/iscsi.o 00:04:00.245 CC module/event/subsystems/vhost_scsi/vhost_scsi.o 00:04:00.518 LIB libspdk_event_iscsi.a 00:04:00.518 LIB libspdk_event_vhost_scsi.a 00:04:00.518 SO libspdk_event_iscsi.so.6.0 00:04:00.518 SO libspdk_event_vhost_scsi.so.3.0 00:04:00.518 SYMLINK libspdk_event_iscsi.so 00:04:00.518 SYMLINK libspdk_event_vhost_scsi.so 00:04:00.784 SO libspdk.so.6.0 00:04:00.784 SYMLINK libspdk.so 00:04:01.044 TEST_HEADER include/spdk/accel.h 00:04:01.044 TEST_HEADER include/spdk/accel_module.h 00:04:01.044 CXX app/trace/trace.o 00:04:01.044 TEST_HEADER include/spdk/assert.h 00:04:01.044 TEST_HEADER include/spdk/barrier.h 00:04:01.044 CC test/rpc_client/rpc_client_test.o 00:04:01.044 TEST_HEADER include/spdk/base64.h 00:04:01.044 TEST_HEADER include/spdk/bdev.h 00:04:01.044 TEST_HEADER include/spdk/bdev_module.h 00:04:01.044 CC app/trace_record/trace_record.o 00:04:01.044 TEST_HEADER include/spdk/bdev_zone.h 00:04:01.044 TEST_HEADER include/spdk/bit_array.h 00:04:01.044 TEST_HEADER include/spdk/bit_pool.h 00:04:01.044 TEST_HEADER include/spdk/blob_bdev.h 00:04:01.044 TEST_HEADER include/spdk/blobfs_bdev.h 00:04:01.044 TEST_HEADER include/spdk/blobfs.h 00:04:01.044 TEST_HEADER include/spdk/blob.h 00:04:01.044 TEST_HEADER include/spdk/conf.h 00:04:01.044 TEST_HEADER include/spdk/config.h 00:04:01.044 TEST_HEADER include/spdk/cpuset.h 00:04:01.044 TEST_HEADER include/spdk/crc16.h 00:04:01.044 TEST_HEADER include/spdk/crc32.h 00:04:01.044 TEST_HEADER include/spdk/crc64.h 00:04:01.044 TEST_HEADER include/spdk/dif.h 00:04:01.044 TEST_HEADER include/spdk/dma.h 00:04:01.044 TEST_HEADER include/spdk/endian.h 00:04:01.044 TEST_HEADER include/spdk/env_dpdk.h 00:04:01.044 TEST_HEADER include/spdk/env.h 00:04:01.044 TEST_HEADER include/spdk/event.h 00:04:01.044 TEST_HEADER include/spdk/fd_group.h 00:04:01.044 TEST_HEADER include/spdk/fd.h 00:04:01.044 CC app/nvmf_tgt/nvmf_main.o 00:04:01.044 TEST_HEADER include/spdk/file.h 00:04:01.044 TEST_HEADER include/spdk/fsdev.h 00:04:01.044 TEST_HEADER include/spdk/fsdev_module.h 00:04:01.044 TEST_HEADER include/spdk/ftl.h 00:04:01.044 TEST_HEADER include/spdk/fuse_dispatcher.h 00:04:01.044 TEST_HEADER include/spdk/gpt_spec.h 00:04:01.044 TEST_HEADER include/spdk/hexlify.h 00:04:01.044 TEST_HEADER include/spdk/histogram_data.h 00:04:01.044 TEST_HEADER include/spdk/idxd.h 00:04:01.044 TEST_HEADER include/spdk/idxd_spec.h 00:04:01.044 TEST_HEADER include/spdk/init.h 00:04:01.044 TEST_HEADER include/spdk/ioat.h 00:04:01.044 TEST_HEADER include/spdk/ioat_spec.h 00:04:01.044 TEST_HEADER include/spdk/iscsi_spec.h 00:04:01.044 TEST_HEADER include/spdk/json.h 00:04:01.044 TEST_HEADER include/spdk/jsonrpc.h 00:04:01.044 CC test/thread/poller_perf/poller_perf.o 00:04:01.044 TEST_HEADER include/spdk/keyring.h 00:04:01.044 TEST_HEADER include/spdk/keyring_module.h 00:04:01.044 TEST_HEADER include/spdk/likely.h 00:04:01.044 TEST_HEADER include/spdk/log.h 00:04:01.044 TEST_HEADER include/spdk/lvol.h 00:04:01.044 CC examples/util/zipf/zipf.o 00:04:01.044 TEST_HEADER include/spdk/md5.h 00:04:01.044 TEST_HEADER include/spdk/memory.h 00:04:01.044 TEST_HEADER include/spdk/mmio.h 00:04:01.044 TEST_HEADER include/spdk/nbd.h 00:04:01.044 TEST_HEADER include/spdk/net.h 00:04:01.044 TEST_HEADER include/spdk/notify.h 00:04:01.044 TEST_HEADER include/spdk/nvme.h 00:04:01.044 TEST_HEADER include/spdk/nvme_intel.h 00:04:01.044 TEST_HEADER include/spdk/nvme_ocssd.h 00:04:01.044 CC test/app/bdev_svc/bdev_svc.o 00:04:01.044 TEST_HEADER include/spdk/nvme_ocssd_spec.h 00:04:01.044 TEST_HEADER include/spdk/nvme_spec.h 00:04:01.044 CC test/dma/test_dma/test_dma.o 00:04:01.044 TEST_HEADER include/spdk/nvme_zns.h 00:04:01.044 TEST_HEADER include/spdk/nvmf_cmd.h 00:04:01.044 TEST_HEADER include/spdk/nvmf_fc_spec.h 00:04:01.044 TEST_HEADER include/spdk/nvmf.h 00:04:01.044 TEST_HEADER include/spdk/nvmf_spec.h 00:04:01.044 TEST_HEADER include/spdk/nvmf_transport.h 00:04:01.044 TEST_HEADER include/spdk/opal.h 00:04:01.044 TEST_HEADER include/spdk/opal_spec.h 00:04:01.044 TEST_HEADER include/spdk/pci_ids.h 00:04:01.044 TEST_HEADER include/spdk/pipe.h 00:04:01.044 TEST_HEADER include/spdk/queue.h 00:04:01.044 TEST_HEADER include/spdk/reduce.h 00:04:01.044 TEST_HEADER include/spdk/rpc.h 00:04:01.044 TEST_HEADER include/spdk/scheduler.h 00:04:01.044 TEST_HEADER include/spdk/scsi.h 00:04:01.044 CC test/env/mem_callbacks/mem_callbacks.o 00:04:01.044 TEST_HEADER include/spdk/scsi_spec.h 00:04:01.044 TEST_HEADER include/spdk/sock.h 00:04:01.044 TEST_HEADER include/spdk/stdinc.h 00:04:01.044 TEST_HEADER include/spdk/string.h 00:04:01.044 TEST_HEADER include/spdk/thread.h 00:04:01.044 TEST_HEADER include/spdk/trace.h 00:04:01.044 TEST_HEADER include/spdk/trace_parser.h 00:04:01.044 TEST_HEADER include/spdk/tree.h 00:04:01.044 TEST_HEADER include/spdk/ublk.h 00:04:01.044 TEST_HEADER include/spdk/util.h 00:04:01.044 TEST_HEADER include/spdk/uuid.h 00:04:01.044 TEST_HEADER include/spdk/version.h 00:04:01.044 TEST_HEADER include/spdk/vfio_user_pci.h 00:04:01.044 TEST_HEADER include/spdk/vfio_user_spec.h 00:04:01.044 TEST_HEADER include/spdk/vhost.h 00:04:01.044 TEST_HEADER include/spdk/vmd.h 00:04:01.044 TEST_HEADER include/spdk/xor.h 00:04:01.044 TEST_HEADER include/spdk/zipf.h 00:04:01.044 CXX test/cpp_headers/accel.o 00:04:01.302 LINK rpc_client_test 00:04:01.302 LINK nvmf_tgt 00:04:01.302 LINK poller_perf 00:04:01.302 LINK zipf 00:04:01.302 LINK spdk_trace_record 00:04:01.302 LINK bdev_svc 00:04:01.302 CXX test/cpp_headers/accel_module.o 00:04:01.302 LINK mem_callbacks 00:04:01.302 LINK spdk_trace 00:04:01.302 CXX test/cpp_headers/assert.o 00:04:01.302 CC test/env/vtophys/vtophys.o 00:04:01.561 CC test/app/histogram_perf/histogram_perf.o 00:04:01.561 CC app/iscsi_tgt/iscsi_tgt.o 00:04:01.561 CC examples/ioat/perf/perf.o 00:04:01.561 CXX test/cpp_headers/barrier.o 00:04:01.561 LINK vtophys 00:04:01.561 CC test/app/jsoncat/jsoncat.o 00:04:01.561 CC test/app/fuzz/nvme_fuzz/nvme_fuzz.o 00:04:01.561 CC test/app/fuzz/iscsi_fuzz/iscsi_fuzz.o 00:04:01.561 LINK test_dma 00:04:01.561 CC test/event/event_perf/event_perf.o 00:04:01.561 LINK jsoncat 00:04:01.561 CXX test/cpp_headers/base64.o 00:04:01.561 LINK histogram_perf 00:04:01.820 LINK iscsi_tgt 00:04:01.820 LINK ioat_perf 00:04:01.820 LINK event_perf 00:04:01.820 CC test/env/env_dpdk_post_init/env_dpdk_post_init.o 00:04:01.820 CXX test/cpp_headers/bdev.o 00:04:01.820 CC examples/ioat/verify/verify.o 00:04:01.820 CC test/env/memory/memory_ut.o 00:04:01.821 CC test/app/fuzz/vhost_fuzz/vhost_fuzz_rpc.o 00:04:01.821 LINK env_dpdk_post_init 00:04:02.079 LINK nvme_fuzz 00:04:02.079 CC test/event/reactor/reactor.o 00:04:02.079 CXX test/cpp_headers/bdev_module.o 00:04:02.079 CC test/app/fuzz/vhost_fuzz/vhost_fuzz.o 00:04:02.079 CC examples/vmd/lsvmd/lsvmd.o 00:04:02.079 CC app/spdk_tgt/spdk_tgt.o 00:04:02.079 LINK reactor 00:04:02.079 LINK verify 00:04:02.079 LINK lsvmd 00:04:02.337 CC test/event/reactor_perf/reactor_perf.o 00:04:02.337 CXX test/cpp_headers/bdev_zone.o 00:04:02.337 CC test/app/stub/stub.o 00:04:02.337 LINK spdk_tgt 00:04:02.337 LINK reactor_perf 00:04:02.337 CC examples/vmd/led/led.o 00:04:02.337 CXX test/cpp_headers/bit_array.o 00:04:02.337 CC test/event/app_repeat/app_repeat.o 00:04:02.337 LINK stub 00:04:02.337 LINK vhost_fuzz 00:04:02.596 CC test/event/scheduler/scheduler.o 00:04:02.596 LINK led 00:04:02.596 CXX test/cpp_headers/bit_pool.o 00:04:02.596 LINK app_repeat 00:04:02.596 CC app/spdk_lspci/spdk_lspci.o 00:04:02.596 CXX test/cpp_headers/blob_bdev.o 00:04:02.596 LINK scheduler 00:04:02.596 CC test/accel/dif/dif.o 00:04:02.596 LINK memory_ut 00:04:02.855 LINK spdk_lspci 00:04:02.855 CC test/blobfs/mkfs/mkfs.o 00:04:02.855 CC examples/idxd/perf/perf.o 00:04:02.855 CC test/lvol/esnap/esnap.o 00:04:02.855 CXX test/cpp_headers/blobfs_bdev.o 00:04:02.855 CC test/nvme/aer/aer.o 00:04:02.855 LINK mkfs 00:04:02.855 CC test/nvme/reset/reset.o 00:04:03.123 CC app/spdk_nvme_perf/perf.o 00:04:03.123 CC test/env/pci/pci_ut.o 00:04:03.123 CXX test/cpp_headers/blobfs.o 00:04:03.123 LINK idxd_perf 00:04:03.123 CXX test/cpp_headers/blob.o 00:04:03.123 LINK aer 00:04:03.387 LINK reset 00:04:03.387 CC test/nvme/sgl/sgl.o 00:04:03.387 CXX test/cpp_headers/conf.o 00:04:03.387 CXX test/cpp_headers/config.o 00:04:03.387 CC examples/interrupt_tgt/interrupt_tgt.o 00:04:03.387 CXX test/cpp_headers/cpuset.o 00:04:03.387 LINK iscsi_fuzz 00:04:03.387 LINK pci_ut 00:04:03.387 CC test/nvme/e2edp/nvme_dp.o 00:04:03.646 LINK dif 00:04:03.646 LINK sgl 00:04:03.646 CC examples/thread/thread/thread_ex.o 00:04:03.646 CXX test/cpp_headers/crc16.o 00:04:03.646 LINK interrupt_tgt 00:04:03.646 CXX test/cpp_headers/crc32.o 00:04:03.646 CXX test/cpp_headers/crc64.o 00:04:03.646 CXX test/cpp_headers/dif.o 00:04:03.905 LINK nvme_dp 00:04:03.905 CXX test/cpp_headers/dma.o 00:04:03.905 LINK thread 00:04:03.905 CXX test/cpp_headers/endian.o 00:04:03.905 CXX test/cpp_headers/env_dpdk.o 00:04:03.905 LINK spdk_nvme_perf 00:04:03.905 CC examples/sock/hello_world/hello_sock.o 00:04:03.905 CC test/nvme/overhead/overhead.o 00:04:04.164 CC test/bdev/bdevio/bdevio.o 00:04:04.164 CXX test/cpp_headers/env.o 00:04:04.164 CC test/nvme/err_injection/err_injection.o 00:04:04.164 CC test/nvme/startup/startup.o 00:04:04.164 CC test/nvme/reserve/reserve.o 00:04:04.164 CC test/nvme/simple_copy/simple_copy.o 00:04:04.164 CC app/spdk_nvme_identify/identify.o 00:04:04.164 CXX test/cpp_headers/event.o 00:04:04.164 LINK startup 00:04:04.164 LINK err_injection 00:04:04.164 LINK hello_sock 00:04:04.423 LINK overhead 00:04:04.423 LINK reserve 00:04:04.423 CXX test/cpp_headers/fd_group.o 00:04:04.423 LINK simple_copy 00:04:04.423 LINK bdevio 00:04:04.423 CC test/nvme/connect_stress/connect_stress.o 00:04:04.682 CC examples/accel/perf/accel_perf.o 00:04:04.682 CXX test/cpp_headers/fd.o 00:04:04.682 CC test/nvme/boot_partition/boot_partition.o 00:04:04.682 CC examples/nvme/hello_world/hello_world.o 00:04:04.682 CC examples/blob/hello_world/hello_blob.o 00:04:04.682 CC examples/nvme/reconnect/reconnect.o 00:04:04.682 LINK connect_stress 00:04:04.682 CXX test/cpp_headers/file.o 00:04:04.682 CC examples/nvme/nvme_manage/nvme_manage.o 00:04:04.682 LINK boot_partition 00:04:04.941 LINK hello_blob 00:04:04.941 LINK hello_world 00:04:04.941 CXX test/cpp_headers/fsdev.o 00:04:04.941 CC examples/nvme/arbitration/arbitration.o 00:04:04.941 CC test/nvme/compliance/nvme_compliance.o 00:04:05.200 CXX test/cpp_headers/fsdev_module.o 00:04:05.200 LINK reconnect 00:04:05.200 CC test/nvme/fused_ordering/fused_ordering.o 00:04:05.200 LINK spdk_nvme_identify 00:04:05.200 LINK accel_perf 00:04:05.200 CC examples/blob/cli/blobcli.o 00:04:05.200 CXX test/cpp_headers/ftl.o 00:04:05.460 CC examples/nvme/hotplug/hotplug.o 00:04:05.460 LINK nvme_manage 00:04:05.460 LINK fused_ordering 00:04:05.460 LINK arbitration 00:04:05.460 LINK nvme_compliance 00:04:05.460 CC app/spdk_nvme_discover/discovery_aer.o 00:04:05.460 CXX test/cpp_headers/fuse_dispatcher.o 00:04:05.460 CC examples/fsdev/hello_world/hello_fsdev.o 00:04:05.460 CC examples/nvme/cmb_copy/cmb_copy.o 00:04:05.718 LINK hotplug 00:04:05.719 CXX test/cpp_headers/gpt_spec.o 00:04:05.719 LINK spdk_nvme_discover 00:04:05.719 CC examples/nvme/abort/abort.o 00:04:05.719 CC test/nvme/doorbell_aers/doorbell_aers.o 00:04:05.719 LINK blobcli 00:04:05.719 LINK cmb_copy 00:04:05.719 CXX test/cpp_headers/hexlify.o 00:04:05.719 CC examples/bdev/hello_world/hello_bdev.o 00:04:05.719 CC examples/nvme/pmr_persistence/pmr_persistence.o 00:04:05.719 LINK hello_fsdev 00:04:05.977 LINK doorbell_aers 00:04:05.977 CXX test/cpp_headers/histogram_data.o 00:04:05.977 CC app/spdk_top/spdk_top.o 00:04:05.977 LINK pmr_persistence 00:04:05.977 CC test/nvme/fdp/fdp.o 00:04:05.977 LINK hello_bdev 00:04:05.977 CC test/nvme/cuse/cuse.o 00:04:06.236 LINK abort 00:04:06.236 CXX test/cpp_headers/idxd.o 00:04:06.236 CXX test/cpp_headers/idxd_spec.o 00:04:06.236 CC app/vhost/vhost.o 00:04:06.236 CC examples/bdev/bdevperf/bdevperf.o 00:04:06.236 CXX test/cpp_headers/init.o 00:04:06.236 CXX test/cpp_headers/ioat.o 00:04:06.236 CXX test/cpp_headers/ioat_spec.o 00:04:06.495 CXX test/cpp_headers/iscsi_spec.o 00:04:06.495 LINK fdp 00:04:06.495 CC app/spdk_dd/spdk_dd.o 00:04:06.495 LINK vhost 00:04:06.495 CXX test/cpp_headers/json.o 00:04:06.495 CXX test/cpp_headers/jsonrpc.o 00:04:06.495 CXX test/cpp_headers/keyring.o 00:04:06.753 CXX test/cpp_headers/keyring_module.o 00:04:06.753 CXX test/cpp_headers/likely.o 00:04:06.753 CXX test/cpp_headers/log.o 00:04:06.753 CC app/fio/nvme/fio_plugin.o 00:04:06.753 CXX test/cpp_headers/lvol.o 00:04:06.753 CXX test/cpp_headers/md5.o 00:04:06.753 CXX test/cpp_headers/memory.o 00:04:06.753 LINK spdk_dd 00:04:07.017 CC app/fio/bdev/fio_plugin.o 00:04:07.017 CXX test/cpp_headers/mmio.o 00:04:07.017 CXX test/cpp_headers/nbd.o 00:04:07.017 CXX test/cpp_headers/net.o 00:04:07.018 LINK spdk_top 00:04:07.018 CXX test/cpp_headers/notify.o 00:04:07.018 CXX test/cpp_headers/nvme.o 00:04:07.018 CXX test/cpp_headers/nvme_intel.o 00:04:07.281 CXX test/cpp_headers/nvme_ocssd.o 00:04:07.281 CXX test/cpp_headers/nvme_ocssd_spec.o 00:04:07.281 CXX test/cpp_headers/nvme_spec.o 00:04:07.281 CXX test/cpp_headers/nvme_zns.o 00:04:07.281 CXX test/cpp_headers/nvmf_cmd.o 00:04:07.281 LINK bdevperf 00:04:07.281 CXX test/cpp_headers/nvmf_fc_spec.o 00:04:07.281 LINK spdk_nvme 00:04:07.281 CXX test/cpp_headers/nvmf.o 00:04:07.281 CXX test/cpp_headers/nvmf_spec.o 00:04:07.539 LINK spdk_bdev 00:04:07.539 CXX test/cpp_headers/nvmf_transport.o 00:04:07.539 CXX test/cpp_headers/opal.o 00:04:07.539 CXX test/cpp_headers/opal_spec.o 00:04:07.539 LINK cuse 00:04:07.539 CXX test/cpp_headers/pci_ids.o 00:04:07.539 CXX test/cpp_headers/pipe.o 00:04:07.539 CXX test/cpp_headers/queue.o 00:04:07.539 CXX test/cpp_headers/reduce.o 00:04:07.539 CXX test/cpp_headers/rpc.o 00:04:07.539 CXX test/cpp_headers/scheduler.o 00:04:07.539 CXX test/cpp_headers/scsi.o 00:04:07.540 CXX test/cpp_headers/scsi_spec.o 00:04:07.540 CC examples/nvmf/nvmf/nvmf.o 00:04:07.799 CXX test/cpp_headers/sock.o 00:04:07.799 CXX test/cpp_headers/stdinc.o 00:04:07.799 CXX test/cpp_headers/string.o 00:04:07.799 CXX test/cpp_headers/thread.o 00:04:07.799 CXX test/cpp_headers/trace.o 00:04:07.799 CXX test/cpp_headers/trace_parser.o 00:04:07.799 CXX test/cpp_headers/tree.o 00:04:07.799 CXX test/cpp_headers/ublk.o 00:04:07.799 CXX test/cpp_headers/util.o 00:04:07.799 CXX test/cpp_headers/uuid.o 00:04:07.799 CXX test/cpp_headers/version.o 00:04:07.799 CXX test/cpp_headers/vfio_user_pci.o 00:04:07.799 CXX test/cpp_headers/vfio_user_spec.o 00:04:07.799 CXX test/cpp_headers/vhost.o 00:04:07.799 CXX test/cpp_headers/vmd.o 00:04:08.058 CXX test/cpp_headers/xor.o 00:04:08.058 CXX test/cpp_headers/zipf.o 00:04:08.058 LINK nvmf 00:04:08.997 LINK esnap 00:04:09.564 00:04:09.564 real 1m15.256s 00:04:09.564 user 5m59.406s 00:04:09.564 sys 1m7.637s 00:04:09.564 23:37:57 make -- common/autotest_common.sh@1130 -- $ xtrace_disable 00:04:09.564 23:37:57 make -- common/autotest_common.sh@10 -- $ set +x 00:04:09.564 ************************************ 00:04:09.564 END TEST make 00:04:09.564 ************************************ 00:04:09.564 23:37:57 -- spdk/autobuild.sh@1 -- $ stop_monitor_resources 00:04:09.564 23:37:57 -- pm/common@29 -- $ signal_monitor_resources TERM 00:04:09.564 23:37:57 -- pm/common@40 -- $ local monitor pid pids signal=TERM 00:04:09.564 23:37:57 -- pm/common@42 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:04:09.564 23:37:57 -- pm/common@43 -- $ [[ -e /home/vagrant/spdk_repo/spdk/../output/power/collect-cpu-load.pid ]] 00:04:09.564 23:37:57 -- pm/common@44 -- $ pid=6201 00:04:09.564 23:37:57 -- pm/common@50 -- $ kill -TERM 6201 00:04:09.564 23:37:57 -- pm/common@42 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:04:09.564 23:37:57 -- pm/common@43 -- $ [[ -e /home/vagrant/spdk_repo/spdk/../output/power/collect-vmstat.pid ]] 00:04:09.564 23:37:57 -- pm/common@44 -- $ pid=6203 00:04:09.564 23:37:57 -- pm/common@50 -- $ kill -TERM 6203 00:04:09.564 23:37:57 -- spdk/autorun.sh@26 -- $ (( SPDK_TEST_UNITTEST == 1 || SPDK_RUN_FUNCTIONAL_TEST == 1 )) 00:04:09.564 23:37:57 -- spdk/autorun.sh@27 -- $ sudo -E /home/vagrant/spdk_repo/spdk/autotest.sh /home/vagrant/spdk_repo/autorun-spdk.conf 00:04:09.564 23:37:57 -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:04:09.564 23:37:57 -- common/autotest_common.sh@1693 -- # lcov --version 00:04:09.564 23:37:57 -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:04:09.564 23:37:57 -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:04:09.564 23:37:57 -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:04:09.564 23:37:57 -- scripts/common.sh@333 -- # local ver1 ver1_l 00:04:09.564 23:37:57 -- scripts/common.sh@334 -- # local ver2 ver2_l 00:04:09.564 23:37:57 -- scripts/common.sh@336 -- # IFS=.-: 00:04:09.564 23:37:57 -- scripts/common.sh@336 -- # read -ra ver1 00:04:09.564 23:37:57 -- scripts/common.sh@337 -- # IFS=.-: 00:04:09.564 23:37:57 -- scripts/common.sh@337 -- # read -ra ver2 00:04:09.564 23:37:57 -- scripts/common.sh@338 -- # local 'op=<' 00:04:09.564 23:37:57 -- scripts/common.sh@340 -- # ver1_l=2 00:04:09.564 23:37:57 -- scripts/common.sh@341 -- # ver2_l=1 00:04:09.564 23:37:57 -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:04:09.564 23:37:57 -- scripts/common.sh@344 -- # case "$op" in 00:04:09.564 23:37:57 -- scripts/common.sh@345 -- # : 1 00:04:09.564 23:37:57 -- scripts/common.sh@364 -- # (( v = 0 )) 00:04:09.564 23:37:57 -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:04:09.824 23:37:57 -- scripts/common.sh@365 -- # decimal 1 00:04:09.824 23:37:57 -- scripts/common.sh@353 -- # local d=1 00:04:09.824 23:37:57 -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:04:09.824 23:37:57 -- scripts/common.sh@355 -- # echo 1 00:04:09.824 23:37:57 -- scripts/common.sh@365 -- # ver1[v]=1 00:04:09.824 23:37:57 -- scripts/common.sh@366 -- # decimal 2 00:04:09.824 23:37:57 -- scripts/common.sh@353 -- # local d=2 00:04:09.824 23:37:57 -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:04:09.824 23:37:57 -- scripts/common.sh@355 -- # echo 2 00:04:09.824 23:37:57 -- scripts/common.sh@366 -- # ver2[v]=2 00:04:09.824 23:37:57 -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:04:09.824 23:37:57 -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:04:09.824 23:37:57 -- scripts/common.sh@368 -- # return 0 00:04:09.824 23:37:57 -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:04:09.824 23:37:57 -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:04:09.824 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:04:09.824 --rc genhtml_branch_coverage=1 00:04:09.824 --rc genhtml_function_coverage=1 00:04:09.824 --rc genhtml_legend=1 00:04:09.824 --rc geninfo_all_blocks=1 00:04:09.824 --rc geninfo_unexecuted_blocks=1 00:04:09.824 00:04:09.824 ' 00:04:09.824 23:37:57 -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:04:09.824 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:04:09.824 --rc genhtml_branch_coverage=1 00:04:09.824 --rc genhtml_function_coverage=1 00:04:09.824 --rc genhtml_legend=1 00:04:09.824 --rc geninfo_all_blocks=1 00:04:09.824 --rc geninfo_unexecuted_blocks=1 00:04:09.824 00:04:09.824 ' 00:04:09.824 23:37:57 -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:04:09.824 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:04:09.824 --rc genhtml_branch_coverage=1 00:04:09.824 --rc genhtml_function_coverage=1 00:04:09.824 --rc genhtml_legend=1 00:04:09.824 --rc geninfo_all_blocks=1 00:04:09.824 --rc geninfo_unexecuted_blocks=1 00:04:09.824 00:04:09.824 ' 00:04:09.824 23:37:57 -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:04:09.824 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:04:09.824 --rc genhtml_branch_coverage=1 00:04:09.824 --rc genhtml_function_coverage=1 00:04:09.824 --rc genhtml_legend=1 00:04:09.824 --rc geninfo_all_blocks=1 00:04:09.824 --rc geninfo_unexecuted_blocks=1 00:04:09.824 00:04:09.824 ' 00:04:09.824 23:37:57 -- spdk/autotest.sh@25 -- # source /home/vagrant/spdk_repo/spdk/test/nvmf/common.sh 00:04:09.824 23:37:57 -- nvmf/common.sh@7 -- # uname -s 00:04:09.824 23:37:57 -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:04:09.824 23:37:57 -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:04:09.824 23:37:57 -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:04:09.824 23:37:57 -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:04:09.824 23:37:57 -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:04:09.824 23:37:57 -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:04:09.824 23:37:57 -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:04:09.824 23:37:57 -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:04:09.824 23:37:57 -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:04:09.824 23:37:57 -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:04:09.824 23:37:57 -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:0d30ab93-e078-437c-8ce5-2573a58e7d32 00:04:09.824 23:37:57 -- nvmf/common.sh@18 -- # NVME_HOSTID=0d30ab93-e078-437c-8ce5-2573a58e7d32 00:04:09.824 23:37:57 -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:04:09.824 23:37:57 -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:04:09.824 23:37:57 -- nvmf/common.sh@21 -- # NET_TYPE=phy-fallback 00:04:09.824 23:37:57 -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:04:09.824 23:37:57 -- nvmf/common.sh@49 -- # source /home/vagrant/spdk_repo/spdk/scripts/common.sh 00:04:09.824 23:37:57 -- scripts/common.sh@15 -- # shopt -s extglob 00:04:09.824 23:37:57 -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:04:09.824 23:37:57 -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:04:09.824 23:37:57 -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:04:09.824 23:37:57 -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:04:09.824 23:37:57 -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:04:09.824 23:37:57 -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:04:09.824 23:37:57 -- paths/export.sh@5 -- # export PATH 00:04:09.824 23:37:57 -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:04:09.824 23:37:57 -- nvmf/common.sh@51 -- # : 0 00:04:09.824 23:37:57 -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:04:09.824 23:37:57 -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:04:09.824 23:37:57 -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:04:09.824 23:37:57 -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:04:09.824 23:37:57 -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:04:09.824 23:37:57 -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:04:09.824 /home/vagrant/spdk_repo/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:04:09.824 23:37:57 -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:04:09.824 23:37:57 -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:04:09.824 23:37:57 -- nvmf/common.sh@55 -- # have_pci_nics=0 00:04:09.824 23:37:57 -- spdk/autotest.sh@27 -- # '[' 0 -ne 0 ']' 00:04:09.824 23:37:57 -- spdk/autotest.sh@32 -- # uname -s 00:04:09.824 23:37:57 -- spdk/autotest.sh@32 -- # '[' Linux = Linux ']' 00:04:09.824 23:37:57 -- spdk/autotest.sh@33 -- # old_core_pattern='|/usr/lib/systemd/systemd-coredump %P %u %g %s %t %c %h' 00:04:09.824 23:37:57 -- spdk/autotest.sh@34 -- # mkdir -p /home/vagrant/spdk_repo/spdk/../output/coredumps 00:04:09.824 23:37:57 -- spdk/autotest.sh@39 -- # echo '|/home/vagrant/spdk_repo/spdk/scripts/core-collector.sh %P %s %t' 00:04:09.824 23:37:57 -- spdk/autotest.sh@40 -- # echo /home/vagrant/spdk_repo/spdk/../output/coredumps 00:04:09.824 23:37:57 -- spdk/autotest.sh@44 -- # modprobe nbd 00:04:09.824 23:37:57 -- spdk/autotest.sh@46 -- # type -P udevadm 00:04:09.824 23:37:57 -- spdk/autotest.sh@46 -- # udevadm=/usr/sbin/udevadm 00:04:09.824 23:37:57 -- spdk/autotest.sh@48 -- # udevadm_pid=66462 00:04:09.824 23:37:57 -- spdk/autotest.sh@47 -- # /usr/sbin/udevadm monitor --property 00:04:09.824 23:37:57 -- spdk/autotest.sh@53 -- # start_monitor_resources 00:04:09.824 23:37:57 -- pm/common@17 -- # local monitor 00:04:09.824 23:37:57 -- pm/common@19 -- # for monitor in "${MONITOR_RESOURCES[@]}" 00:04:09.824 23:37:57 -- pm/common@19 -- # for monitor in "${MONITOR_RESOURCES[@]}" 00:04:09.824 23:37:57 -- pm/common@25 -- # sleep 1 00:04:09.824 23:37:57 -- pm/common@21 -- # date +%s 00:04:09.824 23:37:57 -- pm/common@21 -- # date +%s 00:04:09.824 23:37:57 -- pm/common@21 -- # /home/vagrant/spdk_repo/spdk/scripts/perf/pm/collect-cpu-load -d /home/vagrant/spdk_repo/spdk/../output/power -l -p monitor.autotest.sh.1732664277 00:04:09.824 23:37:57 -- pm/common@21 -- # /home/vagrant/spdk_repo/spdk/scripts/perf/pm/collect-vmstat -d /home/vagrant/spdk_repo/spdk/../output/power -l -p monitor.autotest.sh.1732664277 00:04:09.824 Redirecting to /home/vagrant/spdk_repo/spdk/../output/power/monitor.autotest.sh.1732664277_collect-vmstat.pm.log 00:04:09.824 Redirecting to /home/vagrant/spdk_repo/spdk/../output/power/monitor.autotest.sh.1732664277_collect-cpu-load.pm.log 00:04:10.764 23:37:58 -- spdk/autotest.sh@55 -- # trap 'autotest_cleanup || :; exit 1' SIGINT SIGTERM EXIT 00:04:10.764 23:37:58 -- spdk/autotest.sh@57 -- # timing_enter autotest 00:04:10.764 23:37:58 -- common/autotest_common.sh@726 -- # xtrace_disable 00:04:10.764 23:37:58 -- common/autotest_common.sh@10 -- # set +x 00:04:10.764 23:37:58 -- spdk/autotest.sh@59 -- # create_test_list 00:04:10.764 23:37:58 -- common/autotest_common.sh@752 -- # xtrace_disable 00:04:10.764 23:37:58 -- common/autotest_common.sh@10 -- # set +x 00:04:11.024 23:37:58 -- spdk/autotest.sh@61 -- # dirname /home/vagrant/spdk_repo/spdk/autotest.sh 00:04:11.024 23:37:58 -- spdk/autotest.sh@61 -- # readlink -f /home/vagrant/spdk_repo/spdk 00:04:11.024 23:37:58 -- spdk/autotest.sh@61 -- # src=/home/vagrant/spdk_repo/spdk 00:04:11.024 23:37:58 -- spdk/autotest.sh@62 -- # out=/home/vagrant/spdk_repo/spdk/../output 00:04:11.024 23:37:58 -- spdk/autotest.sh@63 -- # cd /home/vagrant/spdk_repo/spdk 00:04:11.024 23:37:58 -- spdk/autotest.sh@65 -- # freebsd_update_contigmem_mod 00:04:11.024 23:37:58 -- common/autotest_common.sh@1457 -- # uname 00:04:11.024 23:37:58 -- common/autotest_common.sh@1457 -- # '[' Linux = FreeBSD ']' 00:04:11.024 23:37:58 -- spdk/autotest.sh@66 -- # freebsd_set_maxsock_buf 00:04:11.024 23:37:58 -- common/autotest_common.sh@1477 -- # uname 00:04:11.024 23:37:58 -- common/autotest_common.sh@1477 -- # [[ Linux = FreeBSD ]] 00:04:11.024 23:37:58 -- spdk/autotest.sh@68 -- # [[ y == y ]] 00:04:11.024 23:37:58 -- spdk/autotest.sh@70 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 --version 00:04:11.024 lcov: LCOV version 1.15 00:04:11.024 23:37:59 -- spdk/autotest.sh@72 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -c --no-external -i -t Baseline -d /home/vagrant/spdk_repo/spdk -o /home/vagrant/spdk_repo/spdk/../output/cov_base.info 00:04:25.924 /home/vagrant/spdk_repo/spdk/lib/nvme/nvme_stubs.gcno:no functions found 00:04:25.924 geninfo: WARNING: GCOV did not produce any data for /home/vagrant/spdk_repo/spdk/lib/nvme/nvme_stubs.gcno 00:04:40.825 23:38:28 -- spdk/autotest.sh@76 -- # timing_enter pre_cleanup 00:04:40.825 23:38:28 -- common/autotest_common.sh@726 -- # xtrace_disable 00:04:40.825 23:38:28 -- common/autotest_common.sh@10 -- # set +x 00:04:40.825 23:38:28 -- spdk/autotest.sh@78 -- # rm -f 00:04:40.825 23:38:28 -- spdk/autotest.sh@81 -- # /home/vagrant/spdk_repo/spdk/scripts/setup.sh reset 00:04:41.460 0000:00:03.0 (1af4 1001): Active devices: mount@vda:vda2,mount@vda:vda3,mount@vda:vda5, so not binding PCI dev 00:04:41.460 0000:00:11.0 (1b36 0010): Already using the nvme driver 00:04:41.460 0000:00:10.0 (1b36 0010): Already using the nvme driver 00:04:41.460 23:38:29 -- spdk/autotest.sh@83 -- # get_zoned_devs 00:04:41.460 23:38:29 -- common/autotest_common.sh@1657 -- # zoned_devs=() 00:04:41.460 23:38:29 -- common/autotest_common.sh@1657 -- # local -gA zoned_devs 00:04:41.460 23:38:29 -- common/autotest_common.sh@1658 -- # local nvme bdf 00:04:41.460 23:38:29 -- common/autotest_common.sh@1660 -- # for nvme in /sys/block/nvme* 00:04:41.460 23:38:29 -- common/autotest_common.sh@1661 -- # is_block_zoned nvme0n1 00:04:41.460 23:38:29 -- common/autotest_common.sh@1650 -- # local device=nvme0n1 00:04:41.460 23:38:29 -- common/autotest_common.sh@1652 -- # [[ -e /sys/block/nvme0n1/queue/zoned ]] 00:04:41.460 23:38:29 -- common/autotest_common.sh@1653 -- # [[ none != none ]] 00:04:41.460 23:38:29 -- common/autotest_common.sh@1660 -- # for nvme in /sys/block/nvme* 00:04:41.460 23:38:29 -- common/autotest_common.sh@1661 -- # is_block_zoned nvme1n1 00:04:41.460 23:38:29 -- common/autotest_common.sh@1650 -- # local device=nvme1n1 00:04:41.460 23:38:29 -- common/autotest_common.sh@1652 -- # [[ -e /sys/block/nvme1n1/queue/zoned ]] 00:04:41.460 23:38:29 -- common/autotest_common.sh@1653 -- # [[ none != none ]] 00:04:41.460 23:38:29 -- common/autotest_common.sh@1660 -- # for nvme in /sys/block/nvme* 00:04:41.460 23:38:29 -- common/autotest_common.sh@1661 -- # is_block_zoned nvme1n2 00:04:41.460 23:38:29 -- common/autotest_common.sh@1650 -- # local device=nvme1n2 00:04:41.460 23:38:29 -- common/autotest_common.sh@1652 -- # [[ -e /sys/block/nvme1n2/queue/zoned ]] 00:04:41.460 23:38:29 -- common/autotest_common.sh@1653 -- # [[ none != none ]] 00:04:41.460 23:38:29 -- common/autotest_common.sh@1660 -- # for nvme in /sys/block/nvme* 00:04:41.460 23:38:29 -- common/autotest_common.sh@1661 -- # is_block_zoned nvme1n3 00:04:41.460 23:38:29 -- common/autotest_common.sh@1650 -- # local device=nvme1n3 00:04:41.460 23:38:29 -- common/autotest_common.sh@1652 -- # [[ -e /sys/block/nvme1n3/queue/zoned ]] 00:04:41.460 23:38:29 -- common/autotest_common.sh@1653 -- # [[ none != none ]] 00:04:41.460 23:38:29 -- spdk/autotest.sh@85 -- # (( 0 > 0 )) 00:04:41.460 23:38:29 -- spdk/autotest.sh@97 -- # for dev in /dev/nvme*n!(*p*) 00:04:41.460 23:38:29 -- spdk/autotest.sh@99 -- # [[ -z '' ]] 00:04:41.460 23:38:29 -- spdk/autotest.sh@100 -- # block_in_use /dev/nvme0n1 00:04:41.460 23:38:29 -- scripts/common.sh@381 -- # local block=/dev/nvme0n1 pt 00:04:41.460 23:38:29 -- scripts/common.sh@390 -- # /home/vagrant/spdk_repo/spdk/scripts/spdk-gpt.py /dev/nvme0n1 00:04:41.460 No valid GPT data, bailing 00:04:41.460 23:38:29 -- scripts/common.sh@394 -- # blkid -s PTTYPE -o value /dev/nvme0n1 00:04:41.460 23:38:29 -- scripts/common.sh@394 -- # pt= 00:04:41.460 23:38:29 -- scripts/common.sh@395 -- # return 1 00:04:41.460 23:38:29 -- spdk/autotest.sh@101 -- # dd if=/dev/zero of=/dev/nvme0n1 bs=1M count=1 00:04:41.460 1+0 records in 00:04:41.460 1+0 records out 00:04:41.460 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.00630345 s, 166 MB/s 00:04:41.460 23:38:29 -- spdk/autotest.sh@97 -- # for dev in /dev/nvme*n!(*p*) 00:04:41.460 23:38:29 -- spdk/autotest.sh@99 -- # [[ -z '' ]] 00:04:41.460 23:38:29 -- spdk/autotest.sh@100 -- # block_in_use /dev/nvme1n1 00:04:41.460 23:38:29 -- scripts/common.sh@381 -- # local block=/dev/nvme1n1 pt 00:04:41.460 23:38:29 -- scripts/common.sh@390 -- # /home/vagrant/spdk_repo/spdk/scripts/spdk-gpt.py /dev/nvme1n1 00:04:41.734 No valid GPT data, bailing 00:04:41.734 23:38:29 -- scripts/common.sh@394 -- # blkid -s PTTYPE -o value /dev/nvme1n1 00:04:41.734 23:38:29 -- scripts/common.sh@394 -- # pt= 00:04:41.734 23:38:29 -- scripts/common.sh@395 -- # return 1 00:04:41.734 23:38:29 -- spdk/autotest.sh@101 -- # dd if=/dev/zero of=/dev/nvme1n1 bs=1M count=1 00:04:41.734 1+0 records in 00:04:41.734 1+0 records out 00:04:41.734 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.00685795 s, 153 MB/s 00:04:41.734 23:38:29 -- spdk/autotest.sh@97 -- # for dev in /dev/nvme*n!(*p*) 00:04:41.734 23:38:29 -- spdk/autotest.sh@99 -- # [[ -z '' ]] 00:04:41.734 23:38:29 -- spdk/autotest.sh@100 -- # block_in_use /dev/nvme1n2 00:04:41.734 23:38:29 -- scripts/common.sh@381 -- # local block=/dev/nvme1n2 pt 00:04:41.734 23:38:29 -- scripts/common.sh@390 -- # /home/vagrant/spdk_repo/spdk/scripts/spdk-gpt.py /dev/nvme1n2 00:04:41.734 No valid GPT data, bailing 00:04:41.734 23:38:29 -- scripts/common.sh@394 -- # blkid -s PTTYPE -o value /dev/nvme1n2 00:04:41.734 23:38:29 -- scripts/common.sh@394 -- # pt= 00:04:41.734 23:38:29 -- scripts/common.sh@395 -- # return 1 00:04:41.734 23:38:29 -- spdk/autotest.sh@101 -- # dd if=/dev/zero of=/dev/nvme1n2 bs=1M count=1 00:04:41.734 1+0 records in 00:04:41.734 1+0 records out 00:04:41.734 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.00621783 s, 169 MB/s 00:04:41.734 23:38:29 -- spdk/autotest.sh@97 -- # for dev in /dev/nvme*n!(*p*) 00:04:41.734 23:38:29 -- spdk/autotest.sh@99 -- # [[ -z '' ]] 00:04:41.734 23:38:29 -- spdk/autotest.sh@100 -- # block_in_use /dev/nvme1n3 00:04:41.734 23:38:29 -- scripts/common.sh@381 -- # local block=/dev/nvme1n3 pt 00:04:41.734 23:38:29 -- scripts/common.sh@390 -- # /home/vagrant/spdk_repo/spdk/scripts/spdk-gpt.py /dev/nvme1n3 00:04:41.734 No valid GPT data, bailing 00:04:41.734 23:38:29 -- scripts/common.sh@394 -- # blkid -s PTTYPE -o value /dev/nvme1n3 00:04:41.734 23:38:29 -- scripts/common.sh@394 -- # pt= 00:04:41.734 23:38:29 -- scripts/common.sh@395 -- # return 1 00:04:41.734 23:38:29 -- spdk/autotest.sh@101 -- # dd if=/dev/zero of=/dev/nvme1n3 bs=1M count=1 00:04:41.734 1+0 records in 00:04:41.734 1+0 records out 00:04:41.734 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.00424613 s, 247 MB/s 00:04:41.734 23:38:29 -- spdk/autotest.sh@105 -- # sync 00:04:41.734 23:38:29 -- spdk/autotest.sh@107 -- # xtrace_disable_per_cmd reap_spdk_processes 00:04:41.734 23:38:29 -- common/autotest_common.sh@22 -- # eval 'reap_spdk_processes 12> /dev/null' 00:04:41.734 23:38:29 -- common/autotest_common.sh@22 -- # reap_spdk_processes 00:04:45.026 23:38:32 -- spdk/autotest.sh@111 -- # uname -s 00:04:45.026 23:38:32 -- spdk/autotest.sh@111 -- # [[ Linux == Linux ]] 00:04:45.026 23:38:32 -- spdk/autotest.sh@111 -- # [[ 0 -eq 1 ]] 00:04:45.026 23:38:32 -- spdk/autotest.sh@115 -- # /home/vagrant/spdk_repo/spdk/scripts/setup.sh status 00:04:45.595 0000:00:03.0 (1af4 1001): Active devices: mount@vda:vda2,mount@vda:vda3,mount@vda:vda5, so not binding PCI dev 00:04:45.595 Hugepages 00:04:45.595 node hugesize free / total 00:04:45.595 node0 1048576kB 0 / 0 00:04:45.595 node0 2048kB 0 / 0 00:04:45.595 00:04:45.595 Type BDF Vendor Device NUMA Driver Device Block devices 00:04:45.854 virtio 0000:00:03.0 1af4 1001 unknown virtio-pci - vda 00:04:45.854 NVMe 0000:00:10.0 1b36 0010 unknown nvme nvme0 nvme0n1 00:04:45.854 NVMe 0000:00:11.0 1b36 0010 unknown nvme nvme1 nvme1n1 nvme1n2 nvme1n3 00:04:45.854 23:38:33 -- spdk/autotest.sh@117 -- # uname -s 00:04:45.854 23:38:33 -- spdk/autotest.sh@117 -- # [[ Linux == Linux ]] 00:04:45.854 23:38:33 -- spdk/autotest.sh@119 -- # nvme_namespace_revert 00:04:45.854 23:38:33 -- common/autotest_common.sh@1516 -- # /home/vagrant/spdk_repo/spdk/scripts/setup.sh 00:04:46.794 0000:00:03.0 (1af4 1001): Active devices: mount@vda:vda2,mount@vda:vda3,mount@vda:vda5, so not binding PCI dev 00:04:46.794 0000:00:10.0 (1b36 0010): nvme -> uio_pci_generic 00:04:47.053 0000:00:11.0 (1b36 0010): nvme -> uio_pci_generic 00:04:47.053 23:38:34 -- common/autotest_common.sh@1517 -- # sleep 1 00:04:47.992 23:38:36 -- common/autotest_common.sh@1518 -- # bdfs=() 00:04:47.992 23:38:36 -- common/autotest_common.sh@1518 -- # local bdfs 00:04:47.992 23:38:36 -- common/autotest_common.sh@1520 -- # bdfs=($(get_nvme_bdfs)) 00:04:47.992 23:38:36 -- common/autotest_common.sh@1520 -- # get_nvme_bdfs 00:04:47.992 23:38:36 -- common/autotest_common.sh@1498 -- # bdfs=() 00:04:47.992 23:38:36 -- common/autotest_common.sh@1498 -- # local bdfs 00:04:47.992 23:38:36 -- common/autotest_common.sh@1499 -- # bdfs=($("$rootdir/scripts/gen_nvme.sh" | jq -r '.config[].params.traddr')) 00:04:47.992 23:38:36 -- common/autotest_common.sh@1499 -- # /home/vagrant/spdk_repo/spdk/scripts/gen_nvme.sh 00:04:47.992 23:38:36 -- common/autotest_common.sh@1499 -- # jq -r '.config[].params.traddr' 00:04:47.992 23:38:36 -- common/autotest_common.sh@1500 -- # (( 2 == 0 )) 00:04:47.992 23:38:36 -- common/autotest_common.sh@1504 -- # printf '%s\n' 0000:00:10.0 0000:00:11.0 00:04:47.992 23:38:36 -- common/autotest_common.sh@1522 -- # /home/vagrant/spdk_repo/spdk/scripts/setup.sh reset 00:04:48.581 0000:00:03.0 (1af4 1001): Active devices: mount@vda:vda2,mount@vda:vda3,mount@vda:vda5, so not binding PCI dev 00:04:48.581 Waiting for block devices as requested 00:04:48.581 0000:00:11.0 (1b36 0010): uio_pci_generic -> nvme 00:04:48.841 0000:00:10.0 (1b36 0010): uio_pci_generic -> nvme 00:04:48.841 23:38:36 -- common/autotest_common.sh@1524 -- # for bdf in "${bdfs[@]}" 00:04:48.841 23:38:36 -- common/autotest_common.sh@1525 -- # get_nvme_ctrlr_from_bdf 0000:00:10.0 00:04:48.841 23:38:36 -- common/autotest_common.sh@1487 -- # readlink -f /sys/class/nvme/nvme0 /sys/class/nvme/nvme1 00:04:48.841 23:38:36 -- common/autotest_common.sh@1487 -- # grep 0000:00:10.0/nvme/nvme 00:04:48.841 23:38:36 -- common/autotest_common.sh@1487 -- # bdf_sysfs_path=/sys/devices/pci0000:00/0000:00:10.0/nvme/nvme1 00:04:48.841 23:38:36 -- common/autotest_common.sh@1488 -- # [[ -z /sys/devices/pci0000:00/0000:00:10.0/nvme/nvme1 ]] 00:04:48.841 23:38:36 -- common/autotest_common.sh@1492 -- # basename /sys/devices/pci0000:00/0000:00:10.0/nvme/nvme1 00:04:48.841 23:38:36 -- common/autotest_common.sh@1492 -- # printf '%s\n' nvme1 00:04:48.841 23:38:36 -- common/autotest_common.sh@1525 -- # nvme_ctrlr=/dev/nvme1 00:04:48.841 23:38:36 -- common/autotest_common.sh@1526 -- # [[ -z /dev/nvme1 ]] 00:04:48.841 23:38:36 -- common/autotest_common.sh@1531 -- # nvme id-ctrl /dev/nvme1 00:04:48.841 23:38:36 -- common/autotest_common.sh@1531 -- # grep oacs 00:04:48.841 23:38:36 -- common/autotest_common.sh@1531 -- # cut -d: -f2 00:04:48.841 23:38:36 -- common/autotest_common.sh@1531 -- # oacs=' 0x12a' 00:04:48.841 23:38:36 -- common/autotest_common.sh@1532 -- # oacs_ns_manage=8 00:04:48.841 23:38:36 -- common/autotest_common.sh@1534 -- # [[ 8 -ne 0 ]] 00:04:48.841 23:38:36 -- common/autotest_common.sh@1540 -- # nvme id-ctrl /dev/nvme1 00:04:48.841 23:38:36 -- common/autotest_common.sh@1540 -- # grep unvmcap 00:04:48.841 23:38:36 -- common/autotest_common.sh@1540 -- # cut -d: -f2 00:04:48.841 23:38:36 -- common/autotest_common.sh@1540 -- # unvmcap=' 0' 00:04:48.841 23:38:36 -- common/autotest_common.sh@1541 -- # [[ 0 -eq 0 ]] 00:04:48.841 23:38:36 -- common/autotest_common.sh@1543 -- # continue 00:04:48.841 23:38:36 -- common/autotest_common.sh@1524 -- # for bdf in "${bdfs[@]}" 00:04:48.841 23:38:36 -- common/autotest_common.sh@1525 -- # get_nvme_ctrlr_from_bdf 0000:00:11.0 00:04:48.841 23:38:36 -- common/autotest_common.sh@1487 -- # readlink -f /sys/class/nvme/nvme0 /sys/class/nvme/nvme1 00:04:48.841 23:38:36 -- common/autotest_common.sh@1487 -- # grep 0000:00:11.0/nvme/nvme 00:04:48.841 23:38:36 -- common/autotest_common.sh@1487 -- # bdf_sysfs_path=/sys/devices/pci0000:00/0000:00:11.0/nvme/nvme0 00:04:48.841 23:38:36 -- common/autotest_common.sh@1488 -- # [[ -z /sys/devices/pci0000:00/0000:00:11.0/nvme/nvme0 ]] 00:04:48.841 23:38:36 -- common/autotest_common.sh@1492 -- # basename /sys/devices/pci0000:00/0000:00:11.0/nvme/nvme0 00:04:48.841 23:38:36 -- common/autotest_common.sh@1492 -- # printf '%s\n' nvme0 00:04:48.841 23:38:36 -- common/autotest_common.sh@1525 -- # nvme_ctrlr=/dev/nvme0 00:04:48.841 23:38:36 -- common/autotest_common.sh@1526 -- # [[ -z /dev/nvme0 ]] 00:04:48.841 23:38:36 -- common/autotest_common.sh@1531 -- # nvme id-ctrl /dev/nvme0 00:04:48.841 23:38:36 -- common/autotest_common.sh@1531 -- # grep oacs 00:04:48.841 23:38:36 -- common/autotest_common.sh@1531 -- # cut -d: -f2 00:04:48.841 23:38:36 -- common/autotest_common.sh@1531 -- # oacs=' 0x12a' 00:04:48.841 23:38:36 -- common/autotest_common.sh@1532 -- # oacs_ns_manage=8 00:04:48.841 23:38:36 -- common/autotest_common.sh@1534 -- # [[ 8 -ne 0 ]] 00:04:48.841 23:38:36 -- common/autotest_common.sh@1540 -- # nvme id-ctrl /dev/nvme0 00:04:48.841 23:38:36 -- common/autotest_common.sh@1540 -- # grep unvmcap 00:04:48.841 23:38:36 -- common/autotest_common.sh@1540 -- # cut -d: -f2 00:04:48.841 23:38:36 -- common/autotest_common.sh@1540 -- # unvmcap=' 0' 00:04:48.841 23:38:36 -- common/autotest_common.sh@1541 -- # [[ 0 -eq 0 ]] 00:04:48.841 23:38:36 -- common/autotest_common.sh@1543 -- # continue 00:04:48.842 23:38:36 -- spdk/autotest.sh@122 -- # timing_exit pre_cleanup 00:04:48.842 23:38:36 -- common/autotest_common.sh@732 -- # xtrace_disable 00:04:48.842 23:38:36 -- common/autotest_common.sh@10 -- # set +x 00:04:49.101 23:38:37 -- spdk/autotest.sh@125 -- # timing_enter afterboot 00:04:49.101 23:38:37 -- common/autotest_common.sh@726 -- # xtrace_disable 00:04:49.101 23:38:37 -- common/autotest_common.sh@10 -- # set +x 00:04:49.101 23:38:37 -- spdk/autotest.sh@126 -- # /home/vagrant/spdk_repo/spdk/scripts/setup.sh 00:04:49.669 0000:00:03.0 (1af4 1001): Active devices: mount@vda:vda2,mount@vda:vda3,mount@vda:vda5, so not binding PCI dev 00:04:49.928 0000:00:10.0 (1b36 0010): nvme -> uio_pci_generic 00:04:49.928 0000:00:11.0 (1b36 0010): nvme -> uio_pci_generic 00:04:49.928 23:38:38 -- spdk/autotest.sh@127 -- # timing_exit afterboot 00:04:49.928 23:38:38 -- common/autotest_common.sh@732 -- # xtrace_disable 00:04:49.928 23:38:38 -- common/autotest_common.sh@10 -- # set +x 00:04:50.187 23:38:38 -- spdk/autotest.sh@131 -- # opal_revert_cleanup 00:04:50.187 23:38:38 -- common/autotest_common.sh@1578 -- # mapfile -t bdfs 00:04:50.187 23:38:38 -- common/autotest_common.sh@1578 -- # get_nvme_bdfs_by_id 0x0a54 00:04:50.187 23:38:38 -- common/autotest_common.sh@1563 -- # bdfs=() 00:04:50.187 23:38:38 -- common/autotest_common.sh@1563 -- # _bdfs=() 00:04:50.187 23:38:38 -- common/autotest_common.sh@1563 -- # local bdfs _bdfs 00:04:50.187 23:38:38 -- common/autotest_common.sh@1564 -- # _bdfs=($(get_nvme_bdfs)) 00:04:50.187 23:38:38 -- common/autotest_common.sh@1564 -- # get_nvme_bdfs 00:04:50.187 23:38:38 -- common/autotest_common.sh@1498 -- # bdfs=() 00:04:50.187 23:38:38 -- common/autotest_common.sh@1498 -- # local bdfs 00:04:50.187 23:38:38 -- common/autotest_common.sh@1499 -- # bdfs=($("$rootdir/scripts/gen_nvme.sh" | jq -r '.config[].params.traddr')) 00:04:50.187 23:38:38 -- common/autotest_common.sh@1499 -- # /home/vagrant/spdk_repo/spdk/scripts/gen_nvme.sh 00:04:50.187 23:38:38 -- common/autotest_common.sh@1499 -- # jq -r '.config[].params.traddr' 00:04:50.187 23:38:38 -- common/autotest_common.sh@1500 -- # (( 2 == 0 )) 00:04:50.187 23:38:38 -- common/autotest_common.sh@1504 -- # printf '%s\n' 0000:00:10.0 0000:00:11.0 00:04:50.187 23:38:38 -- common/autotest_common.sh@1565 -- # for bdf in "${_bdfs[@]}" 00:04:50.187 23:38:38 -- common/autotest_common.sh@1566 -- # cat /sys/bus/pci/devices/0000:00:10.0/device 00:04:50.187 23:38:38 -- common/autotest_common.sh@1566 -- # device=0x0010 00:04:50.187 23:38:38 -- common/autotest_common.sh@1567 -- # [[ 0x0010 == \0\x\0\a\5\4 ]] 00:04:50.187 23:38:38 -- common/autotest_common.sh@1565 -- # for bdf in "${_bdfs[@]}" 00:04:50.187 23:38:38 -- common/autotest_common.sh@1566 -- # cat /sys/bus/pci/devices/0000:00:11.0/device 00:04:50.187 23:38:38 -- common/autotest_common.sh@1566 -- # device=0x0010 00:04:50.187 23:38:38 -- common/autotest_common.sh@1567 -- # [[ 0x0010 == \0\x\0\a\5\4 ]] 00:04:50.187 23:38:38 -- common/autotest_common.sh@1572 -- # (( 0 > 0 )) 00:04:50.187 23:38:38 -- common/autotest_common.sh@1572 -- # return 0 00:04:50.187 23:38:38 -- common/autotest_common.sh@1579 -- # [[ -z '' ]] 00:04:50.187 23:38:38 -- common/autotest_common.sh@1580 -- # return 0 00:04:50.187 23:38:38 -- spdk/autotest.sh@137 -- # '[' 0 -eq 1 ']' 00:04:50.187 23:38:38 -- spdk/autotest.sh@141 -- # '[' 1 -eq 1 ']' 00:04:50.187 23:38:38 -- spdk/autotest.sh@142 -- # [[ 0 -eq 1 ]] 00:04:50.187 23:38:38 -- spdk/autotest.sh@142 -- # [[ 0 -eq 1 ]] 00:04:50.187 23:38:38 -- spdk/autotest.sh@149 -- # timing_enter lib 00:04:50.187 23:38:38 -- common/autotest_common.sh@726 -- # xtrace_disable 00:04:50.187 23:38:38 -- common/autotest_common.sh@10 -- # set +x 00:04:50.187 23:38:38 -- spdk/autotest.sh@151 -- # [[ 0 -eq 1 ]] 00:04:50.187 23:38:38 -- spdk/autotest.sh@155 -- # run_test env /home/vagrant/spdk_repo/spdk/test/env/env.sh 00:04:50.187 23:38:38 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:04:50.187 23:38:38 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:04:50.187 23:38:38 -- common/autotest_common.sh@10 -- # set +x 00:04:50.187 ************************************ 00:04:50.187 START TEST env 00:04:50.187 ************************************ 00:04:50.187 23:38:38 env -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/env/env.sh 00:04:50.187 * Looking for test storage... 00:04:50.446 * Found test storage at /home/vagrant/spdk_repo/spdk/test/env 00:04:50.446 23:38:38 env -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:04:50.446 23:38:38 env -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:04:50.446 23:38:38 env -- common/autotest_common.sh@1693 -- # lcov --version 00:04:50.446 23:38:38 env -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:04:50.446 23:38:38 env -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:04:50.446 23:38:38 env -- scripts/common.sh@333 -- # local ver1 ver1_l 00:04:50.446 23:38:38 env -- scripts/common.sh@334 -- # local ver2 ver2_l 00:04:50.446 23:38:38 env -- scripts/common.sh@336 -- # IFS=.-: 00:04:50.446 23:38:38 env -- scripts/common.sh@336 -- # read -ra ver1 00:04:50.446 23:38:38 env -- scripts/common.sh@337 -- # IFS=.-: 00:04:50.446 23:38:38 env -- scripts/common.sh@337 -- # read -ra ver2 00:04:50.446 23:38:38 env -- scripts/common.sh@338 -- # local 'op=<' 00:04:50.446 23:38:38 env -- scripts/common.sh@340 -- # ver1_l=2 00:04:50.446 23:38:38 env -- scripts/common.sh@341 -- # ver2_l=1 00:04:50.446 23:38:38 env -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:04:50.446 23:38:38 env -- scripts/common.sh@344 -- # case "$op" in 00:04:50.446 23:38:38 env -- scripts/common.sh@345 -- # : 1 00:04:50.447 23:38:38 env -- scripts/common.sh@364 -- # (( v = 0 )) 00:04:50.447 23:38:38 env -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:04:50.447 23:38:38 env -- scripts/common.sh@365 -- # decimal 1 00:04:50.447 23:38:38 env -- scripts/common.sh@353 -- # local d=1 00:04:50.447 23:38:38 env -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:04:50.447 23:38:38 env -- scripts/common.sh@355 -- # echo 1 00:04:50.447 23:38:38 env -- scripts/common.sh@365 -- # ver1[v]=1 00:04:50.447 23:38:38 env -- scripts/common.sh@366 -- # decimal 2 00:04:50.447 23:38:38 env -- scripts/common.sh@353 -- # local d=2 00:04:50.447 23:38:38 env -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:04:50.447 23:38:38 env -- scripts/common.sh@355 -- # echo 2 00:04:50.447 23:38:38 env -- scripts/common.sh@366 -- # ver2[v]=2 00:04:50.447 23:38:38 env -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:04:50.447 23:38:38 env -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:04:50.447 23:38:38 env -- scripts/common.sh@368 -- # return 0 00:04:50.447 23:38:38 env -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:04:50.447 23:38:38 env -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:04:50.447 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:04:50.447 --rc genhtml_branch_coverage=1 00:04:50.447 --rc genhtml_function_coverage=1 00:04:50.447 --rc genhtml_legend=1 00:04:50.447 --rc geninfo_all_blocks=1 00:04:50.447 --rc geninfo_unexecuted_blocks=1 00:04:50.447 00:04:50.447 ' 00:04:50.447 23:38:38 env -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:04:50.447 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:04:50.447 --rc genhtml_branch_coverage=1 00:04:50.447 --rc genhtml_function_coverage=1 00:04:50.447 --rc genhtml_legend=1 00:04:50.447 --rc geninfo_all_blocks=1 00:04:50.447 --rc geninfo_unexecuted_blocks=1 00:04:50.447 00:04:50.447 ' 00:04:50.447 23:38:38 env -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:04:50.447 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:04:50.447 --rc genhtml_branch_coverage=1 00:04:50.447 --rc genhtml_function_coverage=1 00:04:50.447 --rc genhtml_legend=1 00:04:50.447 --rc geninfo_all_blocks=1 00:04:50.447 --rc geninfo_unexecuted_blocks=1 00:04:50.447 00:04:50.447 ' 00:04:50.447 23:38:38 env -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:04:50.447 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:04:50.447 --rc genhtml_branch_coverage=1 00:04:50.447 --rc genhtml_function_coverage=1 00:04:50.447 --rc genhtml_legend=1 00:04:50.447 --rc geninfo_all_blocks=1 00:04:50.447 --rc geninfo_unexecuted_blocks=1 00:04:50.447 00:04:50.447 ' 00:04:50.447 23:38:38 env -- env/env.sh@10 -- # run_test env_memory /home/vagrant/spdk_repo/spdk/test/env/memory/memory_ut 00:04:50.447 23:38:38 env -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:04:50.447 23:38:38 env -- common/autotest_common.sh@1111 -- # xtrace_disable 00:04:50.447 23:38:38 env -- common/autotest_common.sh@10 -- # set +x 00:04:50.447 ************************************ 00:04:50.447 START TEST env_memory 00:04:50.447 ************************************ 00:04:50.447 23:38:38 env.env_memory -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/env/memory/memory_ut 00:04:50.447 00:04:50.447 00:04:50.447 CUnit - A unit testing framework for C - Version 2.1-3 00:04:50.447 http://cunit.sourceforge.net/ 00:04:50.447 00:04:50.447 00:04:50.447 Suite: memory 00:04:50.447 Test: alloc and free memory map ...[2024-11-26 23:38:38.492802] /home/vagrant/spdk_repo/spdk/lib/env_dpdk/memory.c: 283:spdk_mem_map_alloc: *ERROR*: Initial mem_map notify failed 00:04:50.447 passed 00:04:50.447 Test: mem map translation ...[2024-11-26 23:38:38.555748] /home/vagrant/spdk_repo/spdk/lib/env_dpdk/memory.c: 595:spdk_mem_map_set_translation: *ERROR*: invalid spdk_mem_map_set_translation parameters, vaddr=2097152 len=1234 00:04:50.447 [2024-11-26 23:38:38.555856] /home/vagrant/spdk_repo/spdk/lib/env_dpdk/memory.c: 595:spdk_mem_map_set_translation: *ERROR*: invalid spdk_mem_map_set_translation parameters, vaddr=1234 len=2097152 00:04:50.447 [2024-11-26 23:38:38.555947] /home/vagrant/spdk_repo/spdk/lib/env_dpdk/memory.c: 589:spdk_mem_map_set_translation: *ERROR*: invalid usermode virtual address 281474976710656 00:04:50.447 [2024-11-26 23:38:38.555989] /home/vagrant/spdk_repo/spdk/lib/env_dpdk/memory.c: 605:spdk_mem_map_set_translation: *ERROR*: could not get 0xffffffe00000 map 00:04:50.707 passed 00:04:50.707 Test: mem map registration ...[2024-11-26 23:38:38.623288] /home/vagrant/spdk_repo/spdk/lib/env_dpdk/memory.c: 347:spdk_mem_register: *ERROR*: invalid spdk_mem_register parameters, vaddr=200000 len=1234 00:04:50.707 [2024-11-26 23:38:38.623425] /home/vagrant/spdk_repo/spdk/lib/env_dpdk/memory.c: 347:spdk_mem_register: *ERROR*: invalid spdk_mem_register parameters, vaddr=4d2 len=2097152 00:04:50.707 passed 00:04:50.707 Test: mem map adjacent registrations ...passed 00:04:50.707 00:04:50.707 Run Summary: Type Total Ran Passed Failed Inactive 00:04:50.707 suites 1 1 n/a 0 0 00:04:50.707 tests 4 4 4 0 0 00:04:50.707 asserts 152 152 152 0 n/a 00:04:50.707 00:04:50.707 Elapsed time = 0.272 seconds 00:04:50.707 00:04:50.707 real 0m0.312s 00:04:50.707 user 0m0.281s 00:04:50.707 sys 0m0.024s 00:04:50.707 23:38:38 env.env_memory -- common/autotest_common.sh@1130 -- # xtrace_disable 00:04:50.707 23:38:38 env.env_memory -- common/autotest_common.sh@10 -- # set +x 00:04:50.707 ************************************ 00:04:50.707 END TEST env_memory 00:04:50.707 ************************************ 00:04:50.707 23:38:38 env -- env/env.sh@11 -- # run_test env_vtophys /home/vagrant/spdk_repo/spdk/test/env/vtophys/vtophys 00:04:50.707 23:38:38 env -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:04:50.707 23:38:38 env -- common/autotest_common.sh@1111 -- # xtrace_disable 00:04:50.707 23:38:38 env -- common/autotest_common.sh@10 -- # set +x 00:04:50.707 ************************************ 00:04:50.707 START TEST env_vtophys 00:04:50.707 ************************************ 00:04:50.707 23:38:38 env.env_vtophys -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/env/vtophys/vtophys 00:04:50.966 EAL: lib.eal log level changed from notice to debug 00:04:50.966 EAL: Detected lcore 0 as core 0 on socket 0 00:04:50.966 EAL: Detected lcore 1 as core 0 on socket 0 00:04:50.966 EAL: Detected lcore 2 as core 0 on socket 0 00:04:50.966 EAL: Detected lcore 3 as core 0 on socket 0 00:04:50.966 EAL: Detected lcore 4 as core 0 on socket 0 00:04:50.966 EAL: Detected lcore 5 as core 0 on socket 0 00:04:50.966 EAL: Detected lcore 6 as core 0 on socket 0 00:04:50.966 EAL: Detected lcore 7 as core 0 on socket 0 00:04:50.966 EAL: Detected lcore 8 as core 0 on socket 0 00:04:50.966 EAL: Detected lcore 9 as core 0 on socket 0 00:04:50.966 EAL: Maximum logical cores by configuration: 128 00:04:50.967 EAL: Detected CPU lcores: 10 00:04:50.967 EAL: Detected NUMA nodes: 1 00:04:50.967 EAL: Checking presence of .so 'librte_eal.so.23.0' 00:04:50.967 EAL: Detected shared linkage of DPDK 00:04:50.967 EAL: open shared lib /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0/librte_bus_pci.so.23.0 00:04:50.967 EAL: open shared lib /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0/librte_bus_vdev.so.23.0 00:04:50.967 EAL: Registered [vdev] bus. 00:04:50.967 EAL: bus.vdev log level changed from disabled to notice 00:04:50.967 EAL: open shared lib /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0/librte_mempool_ring.so.23.0 00:04:50.967 EAL: open shared lib /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0/librte_net_i40e.so.23.0 00:04:50.967 EAL: pmd.net.i40e.init log level changed from disabled to notice 00:04:50.967 EAL: pmd.net.i40e.driver log level changed from disabled to notice 00:04:50.967 EAL: open shared lib /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0/librte_bus_pci.so 00:04:50.967 EAL: open shared lib /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0/librte_bus_vdev.so 00:04:50.967 EAL: open shared lib /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0/librte_mempool_ring.so 00:04:50.967 EAL: open shared lib /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0/librte_net_i40e.so 00:04:50.967 EAL: No shared files mode enabled, IPC will be disabled 00:04:50.967 EAL: No shared files mode enabled, IPC is disabled 00:04:50.967 EAL: Selected IOVA mode 'PA' 00:04:50.967 EAL: Probing VFIO support... 00:04:50.967 EAL: Module /sys/module/vfio not found! error 2 (No such file or directory) 00:04:50.967 EAL: VFIO modules not loaded, skipping VFIO support... 00:04:50.967 EAL: Ask a virtual area of 0x2e000 bytes 00:04:50.967 EAL: Virtual area found at 0x200000000000 (size = 0x2e000) 00:04:50.967 EAL: Setting up physically contiguous memory... 00:04:50.967 EAL: Setting maximum number of open files to 524288 00:04:50.967 EAL: Detected memory type: socket_id:0 hugepage_sz:2097152 00:04:50.967 EAL: Creating 4 segment lists: n_segs:8192 socket_id:0 hugepage_sz:2097152 00:04:50.967 EAL: Ask a virtual area of 0x61000 bytes 00:04:50.967 EAL: Virtual area found at 0x20000002e000 (size = 0x61000) 00:04:50.967 EAL: Memseg list allocated at socket 0, page size 0x800kB 00:04:50.967 EAL: Ask a virtual area of 0x400000000 bytes 00:04:50.967 EAL: Virtual area found at 0x200000200000 (size = 0x400000000) 00:04:50.967 EAL: VA reserved for memseg list at 0x200000200000, size 400000000 00:04:50.967 EAL: Ask a virtual area of 0x61000 bytes 00:04:50.967 EAL: Virtual area found at 0x200400200000 (size = 0x61000) 00:04:50.967 EAL: Memseg list allocated at socket 0, page size 0x800kB 00:04:50.967 EAL: Ask a virtual area of 0x400000000 bytes 00:04:50.967 EAL: Virtual area found at 0x200400400000 (size = 0x400000000) 00:04:50.967 EAL: VA reserved for memseg list at 0x200400400000, size 400000000 00:04:50.967 EAL: Ask a virtual area of 0x61000 bytes 00:04:50.967 EAL: Virtual area found at 0x200800400000 (size = 0x61000) 00:04:50.967 EAL: Memseg list allocated at socket 0, page size 0x800kB 00:04:50.967 EAL: Ask a virtual area of 0x400000000 bytes 00:04:50.967 EAL: Virtual area found at 0x200800600000 (size = 0x400000000) 00:04:50.967 EAL: VA reserved for memseg list at 0x200800600000, size 400000000 00:04:50.967 EAL: Ask a virtual area of 0x61000 bytes 00:04:50.967 EAL: Virtual area found at 0x200c00600000 (size = 0x61000) 00:04:50.967 EAL: Memseg list allocated at socket 0, page size 0x800kB 00:04:50.967 EAL: Ask a virtual area of 0x400000000 bytes 00:04:50.967 EAL: Virtual area found at 0x200c00800000 (size = 0x400000000) 00:04:50.967 EAL: VA reserved for memseg list at 0x200c00800000, size 400000000 00:04:50.967 EAL: Hugepages will be freed exactly as allocated. 00:04:50.967 EAL: No shared files mode enabled, IPC is disabled 00:04:50.967 EAL: No shared files mode enabled, IPC is disabled 00:04:50.967 EAL: TSC frequency is ~2290000 KHz 00:04:50.967 EAL: Main lcore 0 is ready (tid=7f6971c9fa40;cpuset=[0]) 00:04:50.967 EAL: Trying to obtain current memory policy. 00:04:50.967 EAL: Setting policy MPOL_PREFERRED for socket 0 00:04:50.967 EAL: Restoring previous memory policy: 0 00:04:50.967 EAL: request: mp_malloc_sync 00:04:50.967 EAL: No shared files mode enabled, IPC is disabled 00:04:50.967 EAL: Heap on socket 0 was expanded by 2MB 00:04:50.967 EAL: Module /sys/module/vfio not found! error 2 (No such file or directory) 00:04:50.967 EAL: No shared files mode enabled, IPC is disabled 00:04:50.967 EAL: No PCI address specified using 'addr=' in: bus=pci 00:04:50.967 EAL: Mem event callback 'spdk:(nil)' registered 00:04:50.967 EAL: Module /sys/module/vfio_pci not found! error 2 (No such file or directory) 00:04:50.967 00:04:50.967 00:04:50.967 CUnit - A unit testing framework for C - Version 2.1-3 00:04:50.967 http://cunit.sourceforge.net/ 00:04:50.967 00:04:50.967 00:04:50.967 Suite: components_suite 00:04:51.225 Test: vtophys_malloc_test ...passed 00:04:51.225 Test: vtophys_spdk_malloc_test ...EAL: Trying to obtain current memory policy. 00:04:51.225 EAL: Setting policy MPOL_PREFERRED for socket 0 00:04:51.225 EAL: Restoring previous memory policy: 4 00:04:51.225 EAL: Calling mem event callback 'spdk:(nil)' 00:04:51.225 EAL: request: mp_malloc_sync 00:04:51.225 EAL: No shared files mode enabled, IPC is disabled 00:04:51.225 EAL: Heap on socket 0 was expanded by 4MB 00:04:51.225 EAL: Calling mem event callback 'spdk:(nil)' 00:04:51.225 EAL: request: mp_malloc_sync 00:04:51.225 EAL: No shared files mode enabled, IPC is disabled 00:04:51.225 EAL: Heap on socket 0 was shrunk by 4MB 00:04:51.225 EAL: Trying to obtain current memory policy. 00:04:51.225 EAL: Setting policy MPOL_PREFERRED for socket 0 00:04:51.225 EAL: Restoring previous memory policy: 4 00:04:51.225 EAL: Calling mem event callback 'spdk:(nil)' 00:04:51.225 EAL: request: mp_malloc_sync 00:04:51.225 EAL: No shared files mode enabled, IPC is disabled 00:04:51.225 EAL: Heap on socket 0 was expanded by 6MB 00:04:51.225 EAL: Calling mem event callback 'spdk:(nil)' 00:04:51.225 EAL: request: mp_malloc_sync 00:04:51.225 EAL: No shared files mode enabled, IPC is disabled 00:04:51.225 EAL: Heap on socket 0 was shrunk by 6MB 00:04:51.225 EAL: Trying to obtain current memory policy. 00:04:51.225 EAL: Setting policy MPOL_PREFERRED for socket 0 00:04:51.225 EAL: Restoring previous memory policy: 4 00:04:51.225 EAL: Calling mem event callback 'spdk:(nil)' 00:04:51.225 EAL: request: mp_malloc_sync 00:04:51.225 EAL: No shared files mode enabled, IPC is disabled 00:04:51.225 EAL: Heap on socket 0 was expanded by 10MB 00:04:51.225 EAL: Calling mem event callback 'spdk:(nil)' 00:04:51.225 EAL: request: mp_malloc_sync 00:04:51.225 EAL: No shared files mode enabled, IPC is disabled 00:04:51.225 EAL: Heap on socket 0 was shrunk by 10MB 00:04:51.225 EAL: Trying to obtain current memory policy. 00:04:51.225 EAL: Setting policy MPOL_PREFERRED for socket 0 00:04:51.225 EAL: Restoring previous memory policy: 4 00:04:51.225 EAL: Calling mem event callback 'spdk:(nil)' 00:04:51.225 EAL: request: mp_malloc_sync 00:04:51.225 EAL: No shared files mode enabled, IPC is disabled 00:04:51.225 EAL: Heap on socket 0 was expanded by 18MB 00:04:51.225 EAL: Calling mem event callback 'spdk:(nil)' 00:04:51.225 EAL: request: mp_malloc_sync 00:04:51.225 EAL: No shared files mode enabled, IPC is disabled 00:04:51.225 EAL: Heap on socket 0 was shrunk by 18MB 00:04:51.225 EAL: Trying to obtain current memory policy. 00:04:51.225 EAL: Setting policy MPOL_PREFERRED for socket 0 00:04:51.225 EAL: Restoring previous memory policy: 4 00:04:51.225 EAL: Calling mem event callback 'spdk:(nil)' 00:04:51.225 EAL: request: mp_malloc_sync 00:04:51.225 EAL: No shared files mode enabled, IPC is disabled 00:04:51.225 EAL: Heap on socket 0 was expanded by 34MB 00:04:51.484 EAL: Calling mem event callback 'spdk:(nil)' 00:04:51.484 EAL: request: mp_malloc_sync 00:04:51.484 EAL: No shared files mode enabled, IPC is disabled 00:04:51.484 EAL: Heap on socket 0 was shrunk by 34MB 00:04:51.484 EAL: Trying to obtain current memory policy. 00:04:51.484 EAL: Setting policy MPOL_PREFERRED for socket 0 00:04:51.484 EAL: Restoring previous memory policy: 4 00:04:51.484 EAL: Calling mem event callback 'spdk:(nil)' 00:04:51.484 EAL: request: mp_malloc_sync 00:04:51.484 EAL: No shared files mode enabled, IPC is disabled 00:04:51.484 EAL: Heap on socket 0 was expanded by 66MB 00:04:51.484 EAL: Calling mem event callback 'spdk:(nil)' 00:04:51.484 EAL: request: mp_malloc_sync 00:04:51.484 EAL: No shared files mode enabled, IPC is disabled 00:04:51.484 EAL: Heap on socket 0 was shrunk by 66MB 00:04:51.484 EAL: Trying to obtain current memory policy. 00:04:51.484 EAL: Setting policy MPOL_PREFERRED for socket 0 00:04:51.484 EAL: Restoring previous memory policy: 4 00:04:51.484 EAL: Calling mem event callback 'spdk:(nil)' 00:04:51.484 EAL: request: mp_malloc_sync 00:04:51.484 EAL: No shared files mode enabled, IPC is disabled 00:04:51.484 EAL: Heap on socket 0 was expanded by 130MB 00:04:51.484 EAL: Calling mem event callback 'spdk:(nil)' 00:04:51.484 EAL: request: mp_malloc_sync 00:04:51.484 EAL: No shared files mode enabled, IPC is disabled 00:04:51.484 EAL: Heap on socket 0 was shrunk by 130MB 00:04:51.484 EAL: Trying to obtain current memory policy. 00:04:51.484 EAL: Setting policy MPOL_PREFERRED for socket 0 00:04:51.484 EAL: Restoring previous memory policy: 4 00:04:51.484 EAL: Calling mem event callback 'spdk:(nil)' 00:04:51.484 EAL: request: mp_malloc_sync 00:04:51.484 EAL: No shared files mode enabled, IPC is disabled 00:04:51.484 EAL: Heap on socket 0 was expanded by 258MB 00:04:51.484 EAL: Calling mem event callback 'spdk:(nil)' 00:04:51.484 EAL: request: mp_malloc_sync 00:04:51.484 EAL: No shared files mode enabled, IPC is disabled 00:04:51.484 EAL: Heap on socket 0 was shrunk by 258MB 00:04:51.484 EAL: Trying to obtain current memory policy. 00:04:51.484 EAL: Setting policy MPOL_PREFERRED for socket 0 00:04:51.743 EAL: Restoring previous memory policy: 4 00:04:51.743 EAL: Calling mem event callback 'spdk:(nil)' 00:04:51.743 EAL: request: mp_malloc_sync 00:04:51.743 EAL: No shared files mode enabled, IPC is disabled 00:04:51.743 EAL: Heap on socket 0 was expanded by 514MB 00:04:51.743 EAL: Calling mem event callback 'spdk:(nil)' 00:04:51.743 EAL: request: mp_malloc_sync 00:04:51.743 EAL: No shared files mode enabled, IPC is disabled 00:04:51.743 EAL: Heap on socket 0 was shrunk by 514MB 00:04:51.743 EAL: Trying to obtain current memory policy. 00:04:51.743 EAL: Setting policy MPOL_PREFERRED for socket 0 00:04:52.003 EAL: Restoring previous memory policy: 4 00:04:52.003 EAL: Calling mem event callback 'spdk:(nil)' 00:04:52.003 EAL: request: mp_malloc_sync 00:04:52.003 EAL: No shared files mode enabled, IPC is disabled 00:04:52.003 EAL: Heap on socket 0 was expanded by 1026MB 00:04:52.262 EAL: Calling mem event callback 'spdk:(nil)' 00:04:52.262 passed 00:04:52.262 00:04:52.262 Run Summary: Type Total Ran Passed Failed Inactive 00:04:52.262 suites 1 1 n/a 0 0 00:04:52.262 tests 2 2 2 0 0 00:04:52.262 asserts 5218 5218 5218 0 n/a 00:04:52.262 00:04:52.262 Elapsed time = 1.364 seconds 00:04:52.262 EAL: request: mp_malloc_sync 00:04:52.262 EAL: No shared files mode enabled, IPC is disabled 00:04:52.262 EAL: Heap on socket 0 was shrunk by 1026MB 00:04:52.262 EAL: Calling mem event callback 'spdk:(nil)' 00:04:52.262 EAL: request: mp_malloc_sync 00:04:52.262 EAL: No shared files mode enabled, IPC is disabled 00:04:52.262 EAL: Heap on socket 0 was shrunk by 2MB 00:04:52.262 EAL: No shared files mode enabled, IPC is disabled 00:04:52.262 EAL: No shared files mode enabled, IPC is disabled 00:04:52.262 EAL: No shared files mode enabled, IPC is disabled 00:04:52.521 00:04:52.521 real 0m1.607s 00:04:52.521 user 0m0.756s 00:04:52.521 sys 0m0.715s 00:04:52.521 23:38:40 env.env_vtophys -- common/autotest_common.sh@1130 -- # xtrace_disable 00:04:52.521 23:38:40 env.env_vtophys -- common/autotest_common.sh@10 -- # set +x 00:04:52.521 ************************************ 00:04:52.521 END TEST env_vtophys 00:04:52.521 ************************************ 00:04:52.521 23:38:40 env -- env/env.sh@12 -- # run_test env_pci /home/vagrant/spdk_repo/spdk/test/env/pci/pci_ut 00:04:52.521 23:38:40 env -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:04:52.521 23:38:40 env -- common/autotest_common.sh@1111 -- # xtrace_disable 00:04:52.521 23:38:40 env -- common/autotest_common.sh@10 -- # set +x 00:04:52.521 ************************************ 00:04:52.521 START TEST env_pci 00:04:52.521 ************************************ 00:04:52.521 23:38:40 env.env_pci -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/env/pci/pci_ut 00:04:52.521 00:04:52.521 00:04:52.521 CUnit - A unit testing framework for C - Version 2.1-3 00:04:52.521 http://cunit.sourceforge.net/ 00:04:52.521 00:04:52.521 00:04:52.521 Suite: pci 00:04:52.521 Test: pci_hook ...[2024-11-26 23:38:40.512013] /home/vagrant/spdk_repo/spdk/lib/env_dpdk/pci.c:1117:spdk_pci_device_claim: *ERROR*: Cannot create lock on device /var/tmp/spdk_pci_lock_10000:00:01.0, probably process 68727 has claimed it 00:04:52.521 passed 00:04:52.522 00:04:52.522 Run Summary: Type Total Ran Passed Failed Inactive 00:04:52.522 suites 1 1 n/a 0 0 00:04:52.522 tests 1 1 1 0 0 00:04:52.522 asserts 25 25 25 0 n/a 00:04:52.522 00:04:52.522 Elapsed time = 0.006 seconds 00:04:52.522 EAL: Cannot find device (10000:00:01.0) 00:04:52.522 EAL: Failed to attach device on primary process 00:04:52.522 00:04:52.522 real 0m0.094s 00:04:52.522 user 0m0.040s 00:04:52.522 sys 0m0.053s 00:04:52.522 23:38:40 env.env_pci -- common/autotest_common.sh@1130 -- # xtrace_disable 00:04:52.522 23:38:40 env.env_pci -- common/autotest_common.sh@10 -- # set +x 00:04:52.522 ************************************ 00:04:52.522 END TEST env_pci 00:04:52.522 ************************************ 00:04:52.522 23:38:40 env -- env/env.sh@14 -- # argv='-c 0x1 ' 00:04:52.522 23:38:40 env -- env/env.sh@15 -- # uname 00:04:52.522 23:38:40 env -- env/env.sh@15 -- # '[' Linux = Linux ']' 00:04:52.522 23:38:40 env -- env/env.sh@22 -- # argv+=--base-virtaddr=0x200000000000 00:04:52.522 23:38:40 env -- env/env.sh@24 -- # run_test env_dpdk_post_init /home/vagrant/spdk_repo/spdk/test/env/env_dpdk_post_init/env_dpdk_post_init -c 0x1 --base-virtaddr=0x200000000000 00:04:52.522 23:38:40 env -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:04:52.522 23:38:40 env -- common/autotest_common.sh@1111 -- # xtrace_disable 00:04:52.522 23:38:40 env -- common/autotest_common.sh@10 -- # set +x 00:04:52.780 ************************************ 00:04:52.780 START TEST env_dpdk_post_init 00:04:52.780 ************************************ 00:04:52.780 23:38:40 env.env_dpdk_post_init -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/env/env_dpdk_post_init/env_dpdk_post_init -c 0x1 --base-virtaddr=0x200000000000 00:04:52.780 EAL: Detected CPU lcores: 10 00:04:52.780 EAL: Detected NUMA nodes: 1 00:04:52.780 EAL: Detected shared linkage of DPDK 00:04:52.780 EAL: Multi-process socket /var/run/dpdk/rte/mp_socket 00:04:52.780 EAL: Selected IOVA mode 'PA' 00:04:52.780 TELEMETRY: No legacy callbacks, legacy socket not created 00:04:52.780 EAL: Probe PCI driver: spdk_nvme (1b36:0010) device: 0000:00:10.0 (socket -1) 00:04:52.780 EAL: Probe PCI driver: spdk_nvme (1b36:0010) device: 0000:00:11.0 (socket -1) 00:04:52.780 Starting DPDK initialization... 00:04:52.780 Starting SPDK post initialization... 00:04:52.780 SPDK NVMe probe 00:04:52.780 Attaching to 0000:00:10.0 00:04:52.780 Attaching to 0000:00:11.0 00:04:52.780 Attached to 0000:00:10.0 00:04:52.780 Attached to 0000:00:11.0 00:04:52.780 Cleaning up... 00:04:52.780 00:04:52.780 real 0m0.242s 00:04:52.780 user 0m0.074s 00:04:52.780 sys 0m0.070s 00:04:52.780 ************************************ 00:04:52.780 END TEST env_dpdk_post_init 00:04:52.780 ************************************ 00:04:52.780 23:38:40 env.env_dpdk_post_init -- common/autotest_common.sh@1130 -- # xtrace_disable 00:04:52.780 23:38:40 env.env_dpdk_post_init -- common/autotest_common.sh@10 -- # set +x 00:04:53.106 23:38:40 env -- env/env.sh@26 -- # uname 00:04:53.107 23:38:40 env -- env/env.sh@26 -- # '[' Linux = Linux ']' 00:04:53.107 23:38:40 env -- env/env.sh@29 -- # run_test env_mem_callbacks /home/vagrant/spdk_repo/spdk/test/env/mem_callbacks/mem_callbacks 00:04:53.107 23:38:40 env -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:04:53.107 23:38:40 env -- common/autotest_common.sh@1111 -- # xtrace_disable 00:04:53.107 23:38:40 env -- common/autotest_common.sh@10 -- # set +x 00:04:53.107 ************************************ 00:04:53.107 START TEST env_mem_callbacks 00:04:53.107 ************************************ 00:04:53.107 23:38:40 env.env_mem_callbacks -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/env/mem_callbacks/mem_callbacks 00:04:53.107 EAL: Detected CPU lcores: 10 00:04:53.107 EAL: Detected NUMA nodes: 1 00:04:53.107 EAL: Detected shared linkage of DPDK 00:04:53.107 EAL: Multi-process socket /var/run/dpdk/rte/mp_socket 00:04:53.107 EAL: Selected IOVA mode 'PA' 00:04:53.107 TELEMETRY: No legacy callbacks, legacy socket not created 00:04:53.107 00:04:53.107 00:04:53.107 CUnit - A unit testing framework for C - Version 2.1-3 00:04:53.107 http://cunit.sourceforge.net/ 00:04:53.107 00:04:53.107 00:04:53.107 Suite: memory 00:04:53.107 Test: test ... 00:04:53.107 register 0x200000200000 2097152 00:04:53.107 malloc 3145728 00:04:53.107 register 0x200000400000 4194304 00:04:53.107 buf 0x200000500000 len 3145728 PASSED 00:04:53.107 malloc 64 00:04:53.107 buf 0x2000004fff40 len 64 PASSED 00:04:53.107 malloc 4194304 00:04:53.107 register 0x200000800000 6291456 00:04:53.107 buf 0x200000a00000 len 4194304 PASSED 00:04:53.107 free 0x200000500000 3145728 00:04:53.107 free 0x2000004fff40 64 00:04:53.107 unregister 0x200000400000 4194304 PASSED 00:04:53.107 free 0x200000a00000 4194304 00:04:53.107 unregister 0x200000800000 6291456 PASSED 00:04:53.107 malloc 8388608 00:04:53.107 register 0x200000400000 10485760 00:04:53.107 buf 0x200000600000 len 8388608 PASSED 00:04:53.107 free 0x200000600000 8388608 00:04:53.107 unregister 0x200000400000 10485760 PASSED 00:04:53.107 passed 00:04:53.107 00:04:53.107 Run Summary: Type Total Ran Passed Failed Inactive 00:04:53.107 suites 1 1 n/a 0 0 00:04:53.107 tests 1 1 1 0 0 00:04:53.107 asserts 15 15 15 0 n/a 00:04:53.107 00:04:53.107 Elapsed time = 0.012 seconds 00:04:53.107 00:04:53.107 real 0m0.183s 00:04:53.107 user 0m0.036s 00:04:53.107 sys 0m0.044s 00:04:53.107 23:38:41 env.env_mem_callbacks -- common/autotest_common.sh@1130 -- # xtrace_disable 00:04:53.107 23:38:41 env.env_mem_callbacks -- common/autotest_common.sh@10 -- # set +x 00:04:53.107 ************************************ 00:04:53.107 END TEST env_mem_callbacks 00:04:53.107 ************************************ 00:04:53.385 00:04:53.385 real 0m3.021s 00:04:53.385 user 0m1.405s 00:04:53.385 sys 0m1.276s 00:04:53.385 ************************************ 00:04:53.385 END TEST env 00:04:53.385 ************************************ 00:04:53.385 23:38:41 env -- common/autotest_common.sh@1130 -- # xtrace_disable 00:04:53.385 23:38:41 env -- common/autotest_common.sh@10 -- # set +x 00:04:53.385 23:38:41 -- spdk/autotest.sh@156 -- # run_test rpc /home/vagrant/spdk_repo/spdk/test/rpc/rpc.sh 00:04:53.385 23:38:41 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:04:53.385 23:38:41 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:04:53.385 23:38:41 -- common/autotest_common.sh@10 -- # set +x 00:04:53.385 ************************************ 00:04:53.385 START TEST rpc 00:04:53.385 ************************************ 00:04:53.385 23:38:41 rpc -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/rpc/rpc.sh 00:04:53.385 * Looking for test storage... 00:04:53.385 * Found test storage at /home/vagrant/spdk_repo/spdk/test/rpc 00:04:53.385 23:38:41 rpc -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:04:53.385 23:38:41 rpc -- common/autotest_common.sh@1693 -- # lcov --version 00:04:53.385 23:38:41 rpc -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:04:53.385 23:38:41 rpc -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:04:53.385 23:38:41 rpc -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:04:53.385 23:38:41 rpc -- scripts/common.sh@333 -- # local ver1 ver1_l 00:04:53.385 23:38:41 rpc -- scripts/common.sh@334 -- # local ver2 ver2_l 00:04:53.385 23:38:41 rpc -- scripts/common.sh@336 -- # IFS=.-: 00:04:53.385 23:38:41 rpc -- scripts/common.sh@336 -- # read -ra ver1 00:04:53.385 23:38:41 rpc -- scripts/common.sh@337 -- # IFS=.-: 00:04:53.385 23:38:41 rpc -- scripts/common.sh@337 -- # read -ra ver2 00:04:53.385 23:38:41 rpc -- scripts/common.sh@338 -- # local 'op=<' 00:04:53.385 23:38:41 rpc -- scripts/common.sh@340 -- # ver1_l=2 00:04:53.385 23:38:41 rpc -- scripts/common.sh@341 -- # ver2_l=1 00:04:53.385 23:38:41 rpc -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:04:53.385 23:38:41 rpc -- scripts/common.sh@344 -- # case "$op" in 00:04:53.385 23:38:41 rpc -- scripts/common.sh@345 -- # : 1 00:04:53.385 23:38:41 rpc -- scripts/common.sh@364 -- # (( v = 0 )) 00:04:53.385 23:38:41 rpc -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:04:53.385 23:38:41 rpc -- scripts/common.sh@365 -- # decimal 1 00:04:53.385 23:38:41 rpc -- scripts/common.sh@353 -- # local d=1 00:04:53.385 23:38:41 rpc -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:04:53.385 23:38:41 rpc -- scripts/common.sh@355 -- # echo 1 00:04:53.385 23:38:41 rpc -- scripts/common.sh@365 -- # ver1[v]=1 00:04:53.385 23:38:41 rpc -- scripts/common.sh@366 -- # decimal 2 00:04:53.385 23:38:41 rpc -- scripts/common.sh@353 -- # local d=2 00:04:53.385 23:38:41 rpc -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:04:53.385 23:38:41 rpc -- scripts/common.sh@355 -- # echo 2 00:04:53.385 23:38:41 rpc -- scripts/common.sh@366 -- # ver2[v]=2 00:04:53.385 23:38:41 rpc -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:04:53.385 23:38:41 rpc -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:04:53.385 23:38:41 rpc -- scripts/common.sh@368 -- # return 0 00:04:53.385 23:38:41 rpc -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:04:53.385 23:38:41 rpc -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:04:53.385 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:04:53.385 --rc genhtml_branch_coverage=1 00:04:53.385 --rc genhtml_function_coverage=1 00:04:53.385 --rc genhtml_legend=1 00:04:53.385 --rc geninfo_all_blocks=1 00:04:53.385 --rc geninfo_unexecuted_blocks=1 00:04:53.385 00:04:53.385 ' 00:04:53.385 23:38:41 rpc -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:04:53.385 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:04:53.385 --rc genhtml_branch_coverage=1 00:04:53.385 --rc genhtml_function_coverage=1 00:04:53.385 --rc genhtml_legend=1 00:04:53.385 --rc geninfo_all_blocks=1 00:04:53.385 --rc geninfo_unexecuted_blocks=1 00:04:53.385 00:04:53.385 ' 00:04:53.385 23:38:41 rpc -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:04:53.386 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:04:53.386 --rc genhtml_branch_coverage=1 00:04:53.386 --rc genhtml_function_coverage=1 00:04:53.386 --rc genhtml_legend=1 00:04:53.386 --rc geninfo_all_blocks=1 00:04:53.386 --rc geninfo_unexecuted_blocks=1 00:04:53.386 00:04:53.386 ' 00:04:53.386 23:38:41 rpc -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:04:53.386 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:04:53.386 --rc genhtml_branch_coverage=1 00:04:53.386 --rc genhtml_function_coverage=1 00:04:53.386 --rc genhtml_legend=1 00:04:53.386 --rc geninfo_all_blocks=1 00:04:53.386 --rc geninfo_unexecuted_blocks=1 00:04:53.386 00:04:53.386 ' 00:04:53.386 23:38:41 rpc -- rpc/rpc.sh@65 -- # spdk_pid=68854 00:04:53.386 23:38:41 rpc -- rpc/rpc.sh@64 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -e bdev 00:04:53.386 23:38:41 rpc -- rpc/rpc.sh@66 -- # trap 'killprocess $spdk_pid; exit 1' SIGINT SIGTERM EXIT 00:04:53.386 23:38:41 rpc -- rpc/rpc.sh@67 -- # waitforlisten 68854 00:04:53.386 23:38:41 rpc -- common/autotest_common.sh@835 -- # '[' -z 68854 ']' 00:04:53.386 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:04:53.386 23:38:41 rpc -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:04:53.386 23:38:41 rpc -- common/autotest_common.sh@840 -- # local max_retries=100 00:04:53.386 23:38:41 rpc -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:04:53.386 23:38:41 rpc -- common/autotest_common.sh@844 -- # xtrace_disable 00:04:53.386 23:38:41 rpc -- common/autotest_common.sh@10 -- # set +x 00:04:53.645 [2024-11-26 23:38:41.606476] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:04:53.645 [2024-11-26 23:38:41.606640] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid68854 ] 00:04:53.645 [2024-11-26 23:38:41.763059] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:04:53.905 [2024-11-26 23:38:41.791847] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask bdev specified. 00:04:53.905 [2024-11-26 23:38:41.791901] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s spdk_tgt -p 68854' to capture a snapshot of events at runtime. 00:04:53.905 [2024-11-26 23:38:41.791926] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:04:53.905 [2024-11-26 23:38:41.791935] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:04:53.905 [2024-11-26 23:38:41.791944] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/spdk_tgt_trace.pid68854 for offline analysis/debug. 00:04:53.905 [2024-11-26 23:38:41.792294] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:04:54.474 23:38:42 rpc -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:04:54.474 23:38:42 rpc -- common/autotest_common.sh@868 -- # return 0 00:04:54.474 23:38:42 rpc -- rpc/rpc.sh@69 -- # export PYTHONPATH=:/home/vagrant/spdk_repo/spdk/python:/home/vagrant/spdk_repo/spdk/test/rpc_plugins:/home/vagrant/spdk_repo/spdk/python:/home/vagrant/spdk_repo/spdk/test/rpc_plugins:/home/vagrant/spdk_repo/spdk/test/rpc 00:04:54.474 23:38:42 rpc -- rpc/rpc.sh@69 -- # PYTHONPATH=:/home/vagrant/spdk_repo/spdk/python:/home/vagrant/spdk_repo/spdk/test/rpc_plugins:/home/vagrant/spdk_repo/spdk/python:/home/vagrant/spdk_repo/spdk/test/rpc_plugins:/home/vagrant/spdk_repo/spdk/test/rpc 00:04:54.474 23:38:42 rpc -- rpc/rpc.sh@72 -- # rpc=rpc_cmd 00:04:54.474 23:38:42 rpc -- rpc/rpc.sh@73 -- # run_test rpc_integrity rpc_integrity 00:04:54.474 23:38:42 rpc -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:04:54.474 23:38:42 rpc -- common/autotest_common.sh@1111 -- # xtrace_disable 00:04:54.474 23:38:42 rpc -- common/autotest_common.sh@10 -- # set +x 00:04:54.474 ************************************ 00:04:54.474 START TEST rpc_integrity 00:04:54.474 ************************************ 00:04:54.474 23:38:42 rpc.rpc_integrity -- common/autotest_common.sh@1129 -- # rpc_integrity 00:04:54.474 23:38:42 rpc.rpc_integrity -- rpc/rpc.sh@12 -- # rpc_cmd bdev_get_bdevs 00:04:54.474 23:38:42 rpc.rpc_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:04:54.474 23:38:42 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:04:54.474 23:38:42 rpc.rpc_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:04:54.474 23:38:42 rpc.rpc_integrity -- rpc/rpc.sh@12 -- # bdevs='[]' 00:04:54.474 23:38:42 rpc.rpc_integrity -- rpc/rpc.sh@13 -- # jq length 00:04:54.474 23:38:42 rpc.rpc_integrity -- rpc/rpc.sh@13 -- # '[' 0 == 0 ']' 00:04:54.474 23:38:42 rpc.rpc_integrity -- rpc/rpc.sh@15 -- # rpc_cmd bdev_malloc_create 8 512 00:04:54.474 23:38:42 rpc.rpc_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:04:54.474 23:38:42 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:04:54.474 23:38:42 rpc.rpc_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:04:54.474 23:38:42 rpc.rpc_integrity -- rpc/rpc.sh@15 -- # malloc=Malloc0 00:04:54.474 23:38:42 rpc.rpc_integrity -- rpc/rpc.sh@16 -- # rpc_cmd bdev_get_bdevs 00:04:54.474 23:38:42 rpc.rpc_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:04:54.474 23:38:42 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:04:54.474 23:38:42 rpc.rpc_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:04:54.474 23:38:42 rpc.rpc_integrity -- rpc/rpc.sh@16 -- # bdevs='[ 00:04:54.474 { 00:04:54.474 "name": "Malloc0", 00:04:54.474 "aliases": [ 00:04:54.474 "39d87f22-76b3-472e-b1df-9fc76b8e6b7f" 00:04:54.474 ], 00:04:54.474 "product_name": "Malloc disk", 00:04:54.474 "block_size": 512, 00:04:54.474 "num_blocks": 16384, 00:04:54.474 "uuid": "39d87f22-76b3-472e-b1df-9fc76b8e6b7f", 00:04:54.474 "assigned_rate_limits": { 00:04:54.474 "rw_ios_per_sec": 0, 00:04:54.474 "rw_mbytes_per_sec": 0, 00:04:54.474 "r_mbytes_per_sec": 0, 00:04:54.474 "w_mbytes_per_sec": 0 00:04:54.474 }, 00:04:54.474 "claimed": false, 00:04:54.474 "zoned": false, 00:04:54.474 "supported_io_types": { 00:04:54.474 "read": true, 00:04:54.474 "write": true, 00:04:54.474 "unmap": true, 00:04:54.474 "flush": true, 00:04:54.474 "reset": true, 00:04:54.474 "nvme_admin": false, 00:04:54.474 "nvme_io": false, 00:04:54.474 "nvme_io_md": false, 00:04:54.474 "write_zeroes": true, 00:04:54.474 "zcopy": true, 00:04:54.474 "get_zone_info": false, 00:04:54.474 "zone_management": false, 00:04:54.474 "zone_append": false, 00:04:54.474 "compare": false, 00:04:54.474 "compare_and_write": false, 00:04:54.474 "abort": true, 00:04:54.474 "seek_hole": false, 00:04:54.474 "seek_data": false, 00:04:54.474 "copy": true, 00:04:54.474 "nvme_iov_md": false 00:04:54.474 }, 00:04:54.474 "memory_domains": [ 00:04:54.474 { 00:04:54.474 "dma_device_id": "system", 00:04:54.474 "dma_device_type": 1 00:04:54.474 }, 00:04:54.474 { 00:04:54.474 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:04:54.474 "dma_device_type": 2 00:04:54.474 } 00:04:54.474 ], 00:04:54.474 "driver_specific": {} 00:04:54.474 } 00:04:54.474 ]' 00:04:54.474 23:38:42 rpc.rpc_integrity -- rpc/rpc.sh@17 -- # jq length 00:04:54.475 23:38:42 rpc.rpc_integrity -- rpc/rpc.sh@17 -- # '[' 1 == 1 ']' 00:04:54.475 23:38:42 rpc.rpc_integrity -- rpc/rpc.sh@19 -- # rpc_cmd bdev_passthru_create -b Malloc0 -p Passthru0 00:04:54.475 23:38:42 rpc.rpc_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:04:54.475 23:38:42 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:04:54.475 [2024-11-26 23:38:42.595528] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on Malloc0 00:04:54.475 [2024-11-26 23:38:42.595688] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:04:54.475 [2024-11-26 23:38:42.595727] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006c80 00:04:54.475 [2024-11-26 23:38:42.595737] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:04:54.475 [2024-11-26 23:38:42.598478] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:04:54.475 [2024-11-26 23:38:42.598520] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: Passthru0 00:04:54.475 Passthru0 00:04:54.475 23:38:42 rpc.rpc_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:04:54.755 23:38:42 rpc.rpc_integrity -- rpc/rpc.sh@20 -- # rpc_cmd bdev_get_bdevs 00:04:54.755 23:38:42 rpc.rpc_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:04:54.755 23:38:42 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:04:54.756 23:38:42 rpc.rpc_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:04:54.756 23:38:42 rpc.rpc_integrity -- rpc/rpc.sh@20 -- # bdevs='[ 00:04:54.756 { 00:04:54.756 "name": "Malloc0", 00:04:54.756 "aliases": [ 00:04:54.756 "39d87f22-76b3-472e-b1df-9fc76b8e6b7f" 00:04:54.756 ], 00:04:54.756 "product_name": "Malloc disk", 00:04:54.756 "block_size": 512, 00:04:54.756 "num_blocks": 16384, 00:04:54.756 "uuid": "39d87f22-76b3-472e-b1df-9fc76b8e6b7f", 00:04:54.756 "assigned_rate_limits": { 00:04:54.756 "rw_ios_per_sec": 0, 00:04:54.756 "rw_mbytes_per_sec": 0, 00:04:54.756 "r_mbytes_per_sec": 0, 00:04:54.756 "w_mbytes_per_sec": 0 00:04:54.756 }, 00:04:54.756 "claimed": true, 00:04:54.756 "claim_type": "exclusive_write", 00:04:54.756 "zoned": false, 00:04:54.756 "supported_io_types": { 00:04:54.756 "read": true, 00:04:54.756 "write": true, 00:04:54.756 "unmap": true, 00:04:54.756 "flush": true, 00:04:54.756 "reset": true, 00:04:54.756 "nvme_admin": false, 00:04:54.756 "nvme_io": false, 00:04:54.756 "nvme_io_md": false, 00:04:54.756 "write_zeroes": true, 00:04:54.756 "zcopy": true, 00:04:54.756 "get_zone_info": false, 00:04:54.756 "zone_management": false, 00:04:54.756 "zone_append": false, 00:04:54.756 "compare": false, 00:04:54.756 "compare_and_write": false, 00:04:54.756 "abort": true, 00:04:54.756 "seek_hole": false, 00:04:54.756 "seek_data": false, 00:04:54.757 "copy": true, 00:04:54.757 "nvme_iov_md": false 00:04:54.757 }, 00:04:54.757 "memory_domains": [ 00:04:54.757 { 00:04:54.757 "dma_device_id": "system", 00:04:54.757 "dma_device_type": 1 00:04:54.757 }, 00:04:54.757 { 00:04:54.757 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:04:54.757 "dma_device_type": 2 00:04:54.757 } 00:04:54.757 ], 00:04:54.757 "driver_specific": {} 00:04:54.757 }, 00:04:54.757 { 00:04:54.757 "name": "Passthru0", 00:04:54.757 "aliases": [ 00:04:54.757 "de815858-5826-5647-b7d2-930351edd60a" 00:04:54.757 ], 00:04:54.757 "product_name": "passthru", 00:04:54.757 "block_size": 512, 00:04:54.757 "num_blocks": 16384, 00:04:54.757 "uuid": "de815858-5826-5647-b7d2-930351edd60a", 00:04:54.757 "assigned_rate_limits": { 00:04:54.757 "rw_ios_per_sec": 0, 00:04:54.757 "rw_mbytes_per_sec": 0, 00:04:54.757 "r_mbytes_per_sec": 0, 00:04:54.757 "w_mbytes_per_sec": 0 00:04:54.757 }, 00:04:54.757 "claimed": false, 00:04:54.757 "zoned": false, 00:04:54.757 "supported_io_types": { 00:04:54.757 "read": true, 00:04:54.757 "write": true, 00:04:54.757 "unmap": true, 00:04:54.757 "flush": true, 00:04:54.757 "reset": true, 00:04:54.757 "nvme_admin": false, 00:04:54.757 "nvme_io": false, 00:04:54.757 "nvme_io_md": false, 00:04:54.757 "write_zeroes": true, 00:04:54.757 "zcopy": true, 00:04:54.757 "get_zone_info": false, 00:04:54.757 "zone_management": false, 00:04:54.757 "zone_append": false, 00:04:54.757 "compare": false, 00:04:54.757 "compare_and_write": false, 00:04:54.757 "abort": true, 00:04:54.757 "seek_hole": false, 00:04:54.757 "seek_data": false, 00:04:54.757 "copy": true, 00:04:54.757 "nvme_iov_md": false 00:04:54.757 }, 00:04:54.757 "memory_domains": [ 00:04:54.757 { 00:04:54.757 "dma_device_id": "system", 00:04:54.757 "dma_device_type": 1 00:04:54.757 }, 00:04:54.757 { 00:04:54.757 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:04:54.757 "dma_device_type": 2 00:04:54.757 } 00:04:54.757 ], 00:04:54.757 "driver_specific": { 00:04:54.757 "passthru": { 00:04:54.758 "name": "Passthru0", 00:04:54.758 "base_bdev_name": "Malloc0" 00:04:54.758 } 00:04:54.758 } 00:04:54.758 } 00:04:54.758 ]' 00:04:54.758 23:38:42 rpc.rpc_integrity -- rpc/rpc.sh@21 -- # jq length 00:04:54.758 23:38:42 rpc.rpc_integrity -- rpc/rpc.sh@21 -- # '[' 2 == 2 ']' 00:04:54.758 23:38:42 rpc.rpc_integrity -- rpc/rpc.sh@23 -- # rpc_cmd bdev_passthru_delete Passthru0 00:04:54.758 23:38:42 rpc.rpc_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:04:54.758 23:38:42 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:04:54.758 23:38:42 rpc.rpc_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:04:54.758 23:38:42 rpc.rpc_integrity -- rpc/rpc.sh@24 -- # rpc_cmd bdev_malloc_delete Malloc0 00:04:54.758 23:38:42 rpc.rpc_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:04:54.758 23:38:42 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:04:54.758 23:38:42 rpc.rpc_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:04:54.758 23:38:42 rpc.rpc_integrity -- rpc/rpc.sh@25 -- # rpc_cmd bdev_get_bdevs 00:04:54.758 23:38:42 rpc.rpc_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:04:54.758 23:38:42 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:04:54.758 23:38:42 rpc.rpc_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:04:54.758 23:38:42 rpc.rpc_integrity -- rpc/rpc.sh@25 -- # bdevs='[]' 00:04:54.758 23:38:42 rpc.rpc_integrity -- rpc/rpc.sh@26 -- # jq length 00:04:54.758 ************************************ 00:04:54.758 END TEST rpc_integrity 00:04:54.758 ************************************ 00:04:54.758 23:38:42 rpc.rpc_integrity -- rpc/rpc.sh@26 -- # '[' 0 == 0 ']' 00:04:54.758 00:04:54.758 real 0m0.312s 00:04:54.758 user 0m0.187s 00:04:54.758 sys 0m0.051s 00:04:54.758 23:38:42 rpc.rpc_integrity -- common/autotest_common.sh@1130 -- # xtrace_disable 00:04:54.758 23:38:42 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:04:54.758 23:38:42 rpc -- rpc/rpc.sh@74 -- # run_test rpc_plugins rpc_plugins 00:04:54.758 23:38:42 rpc -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:04:54.758 23:38:42 rpc -- common/autotest_common.sh@1111 -- # xtrace_disable 00:04:54.758 23:38:42 rpc -- common/autotest_common.sh@10 -- # set +x 00:04:54.758 ************************************ 00:04:54.758 START TEST rpc_plugins 00:04:54.758 ************************************ 00:04:54.758 23:38:42 rpc.rpc_plugins -- common/autotest_common.sh@1129 -- # rpc_plugins 00:04:54.758 23:38:42 rpc.rpc_plugins -- rpc/rpc.sh@30 -- # rpc_cmd --plugin rpc_plugin create_malloc 00:04:54.758 23:38:42 rpc.rpc_plugins -- common/autotest_common.sh@563 -- # xtrace_disable 00:04:54.758 23:38:42 rpc.rpc_plugins -- common/autotest_common.sh@10 -- # set +x 00:04:54.758 23:38:42 rpc.rpc_plugins -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:04:54.758 23:38:42 rpc.rpc_plugins -- rpc/rpc.sh@30 -- # malloc=Malloc1 00:04:54.758 23:38:42 rpc.rpc_plugins -- rpc/rpc.sh@31 -- # rpc_cmd bdev_get_bdevs 00:04:54.759 23:38:42 rpc.rpc_plugins -- common/autotest_common.sh@563 -- # xtrace_disable 00:04:54.759 23:38:42 rpc.rpc_plugins -- common/autotest_common.sh@10 -- # set +x 00:04:54.759 23:38:42 rpc.rpc_plugins -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:04:54.759 23:38:42 rpc.rpc_plugins -- rpc/rpc.sh@31 -- # bdevs='[ 00:04:54.759 { 00:04:54.759 "name": "Malloc1", 00:04:54.759 "aliases": [ 00:04:54.759 "4a89696b-a95b-48c4-a0f2-be413644e96d" 00:04:54.759 ], 00:04:54.759 "product_name": "Malloc disk", 00:04:54.759 "block_size": 4096, 00:04:54.759 "num_blocks": 256, 00:04:54.759 "uuid": "4a89696b-a95b-48c4-a0f2-be413644e96d", 00:04:54.759 "assigned_rate_limits": { 00:04:54.759 "rw_ios_per_sec": 0, 00:04:54.759 "rw_mbytes_per_sec": 0, 00:04:54.759 "r_mbytes_per_sec": 0, 00:04:54.759 "w_mbytes_per_sec": 0 00:04:54.759 }, 00:04:54.759 "claimed": false, 00:04:54.759 "zoned": false, 00:04:54.759 "supported_io_types": { 00:04:54.759 "read": true, 00:04:54.759 "write": true, 00:04:54.759 "unmap": true, 00:04:54.759 "flush": true, 00:04:54.759 "reset": true, 00:04:54.759 "nvme_admin": false, 00:04:54.759 "nvme_io": false, 00:04:54.759 "nvme_io_md": false, 00:04:54.759 "write_zeroes": true, 00:04:54.759 "zcopy": true, 00:04:54.759 "get_zone_info": false, 00:04:54.759 "zone_management": false, 00:04:54.759 "zone_append": false, 00:04:54.759 "compare": false, 00:04:54.759 "compare_and_write": false, 00:04:54.759 "abort": true, 00:04:54.759 "seek_hole": false, 00:04:54.759 "seek_data": false, 00:04:54.759 "copy": true, 00:04:54.759 "nvme_iov_md": false 00:04:54.759 }, 00:04:54.759 "memory_domains": [ 00:04:54.759 { 00:04:54.759 "dma_device_id": "system", 00:04:54.759 "dma_device_type": 1 00:04:54.759 }, 00:04:54.759 { 00:04:54.759 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:04:54.759 "dma_device_type": 2 00:04:54.759 } 00:04:54.759 ], 00:04:54.759 "driver_specific": {} 00:04:54.759 } 00:04:54.759 ]' 00:04:54.759 23:38:42 rpc.rpc_plugins -- rpc/rpc.sh@32 -- # jq length 00:04:55.023 23:38:42 rpc.rpc_plugins -- rpc/rpc.sh@32 -- # '[' 1 == 1 ']' 00:04:55.023 23:38:42 rpc.rpc_plugins -- rpc/rpc.sh@34 -- # rpc_cmd --plugin rpc_plugin delete_malloc Malloc1 00:04:55.023 23:38:42 rpc.rpc_plugins -- common/autotest_common.sh@563 -- # xtrace_disable 00:04:55.023 23:38:42 rpc.rpc_plugins -- common/autotest_common.sh@10 -- # set +x 00:04:55.023 23:38:42 rpc.rpc_plugins -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:04:55.023 23:38:42 rpc.rpc_plugins -- rpc/rpc.sh@35 -- # rpc_cmd bdev_get_bdevs 00:04:55.023 23:38:42 rpc.rpc_plugins -- common/autotest_common.sh@563 -- # xtrace_disable 00:04:55.023 23:38:42 rpc.rpc_plugins -- common/autotest_common.sh@10 -- # set +x 00:04:55.023 23:38:42 rpc.rpc_plugins -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:04:55.023 23:38:42 rpc.rpc_plugins -- rpc/rpc.sh@35 -- # bdevs='[]' 00:04:55.023 23:38:42 rpc.rpc_plugins -- rpc/rpc.sh@36 -- # jq length 00:04:55.023 23:38:42 rpc.rpc_plugins -- rpc/rpc.sh@36 -- # '[' 0 == 0 ']' 00:04:55.023 00:04:55.023 real 0m0.170s 00:04:55.023 user 0m0.097s 00:04:55.023 sys 0m0.027s 00:04:55.023 23:38:42 rpc.rpc_plugins -- common/autotest_common.sh@1130 -- # xtrace_disable 00:04:55.023 23:38:42 rpc.rpc_plugins -- common/autotest_common.sh@10 -- # set +x 00:04:55.023 ************************************ 00:04:55.023 END TEST rpc_plugins 00:04:55.023 ************************************ 00:04:55.023 23:38:43 rpc -- rpc/rpc.sh@75 -- # run_test rpc_trace_cmd_test rpc_trace_cmd_test 00:04:55.023 23:38:43 rpc -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:04:55.023 23:38:43 rpc -- common/autotest_common.sh@1111 -- # xtrace_disable 00:04:55.023 23:38:43 rpc -- common/autotest_common.sh@10 -- # set +x 00:04:55.023 ************************************ 00:04:55.023 START TEST rpc_trace_cmd_test 00:04:55.023 ************************************ 00:04:55.023 23:38:43 rpc.rpc_trace_cmd_test -- common/autotest_common.sh@1129 -- # rpc_trace_cmd_test 00:04:55.023 23:38:43 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@40 -- # local info 00:04:55.023 23:38:43 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@42 -- # rpc_cmd trace_get_info 00:04:55.023 23:38:43 rpc.rpc_trace_cmd_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:04:55.023 23:38:43 rpc.rpc_trace_cmd_test -- common/autotest_common.sh@10 -- # set +x 00:04:55.023 23:38:43 rpc.rpc_trace_cmd_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:04:55.023 23:38:43 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@42 -- # info='{ 00:04:55.023 "tpoint_shm_path": "/dev/shm/spdk_tgt_trace.pid68854", 00:04:55.023 "tpoint_group_mask": "0x8", 00:04:55.023 "iscsi_conn": { 00:04:55.023 "mask": "0x2", 00:04:55.023 "tpoint_mask": "0x0" 00:04:55.023 }, 00:04:55.023 "scsi": { 00:04:55.023 "mask": "0x4", 00:04:55.023 "tpoint_mask": "0x0" 00:04:55.023 }, 00:04:55.023 "bdev": { 00:04:55.024 "mask": "0x8", 00:04:55.024 "tpoint_mask": "0xffffffffffffffff" 00:04:55.024 }, 00:04:55.024 "nvmf_rdma": { 00:04:55.024 "mask": "0x10", 00:04:55.024 "tpoint_mask": "0x0" 00:04:55.024 }, 00:04:55.024 "nvmf_tcp": { 00:04:55.024 "mask": "0x20", 00:04:55.024 "tpoint_mask": "0x0" 00:04:55.024 }, 00:04:55.024 "ftl": { 00:04:55.024 "mask": "0x40", 00:04:55.024 "tpoint_mask": "0x0" 00:04:55.024 }, 00:04:55.024 "blobfs": { 00:04:55.024 "mask": "0x80", 00:04:55.024 "tpoint_mask": "0x0" 00:04:55.024 }, 00:04:55.024 "dsa": { 00:04:55.024 "mask": "0x200", 00:04:55.024 "tpoint_mask": "0x0" 00:04:55.024 }, 00:04:55.024 "thread": { 00:04:55.024 "mask": "0x400", 00:04:55.024 "tpoint_mask": "0x0" 00:04:55.024 }, 00:04:55.024 "nvme_pcie": { 00:04:55.024 "mask": "0x800", 00:04:55.024 "tpoint_mask": "0x0" 00:04:55.024 }, 00:04:55.024 "iaa": { 00:04:55.024 "mask": "0x1000", 00:04:55.024 "tpoint_mask": "0x0" 00:04:55.024 }, 00:04:55.024 "nvme_tcp": { 00:04:55.024 "mask": "0x2000", 00:04:55.024 "tpoint_mask": "0x0" 00:04:55.024 }, 00:04:55.024 "bdev_nvme": { 00:04:55.024 "mask": "0x4000", 00:04:55.024 "tpoint_mask": "0x0" 00:04:55.024 }, 00:04:55.024 "sock": { 00:04:55.024 "mask": "0x8000", 00:04:55.024 "tpoint_mask": "0x0" 00:04:55.024 }, 00:04:55.024 "blob": { 00:04:55.024 "mask": "0x10000", 00:04:55.024 "tpoint_mask": "0x0" 00:04:55.024 }, 00:04:55.024 "bdev_raid": { 00:04:55.024 "mask": "0x20000", 00:04:55.024 "tpoint_mask": "0x0" 00:04:55.024 }, 00:04:55.024 "scheduler": { 00:04:55.024 "mask": "0x40000", 00:04:55.024 "tpoint_mask": "0x0" 00:04:55.024 } 00:04:55.024 }' 00:04:55.024 23:38:43 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@43 -- # jq length 00:04:55.024 23:38:43 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@43 -- # '[' 19 -gt 2 ']' 00:04:55.024 23:38:43 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@44 -- # jq 'has("tpoint_group_mask")' 00:04:55.284 23:38:43 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@44 -- # '[' true = true ']' 00:04:55.284 23:38:43 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@45 -- # jq 'has("tpoint_shm_path")' 00:04:55.284 23:38:43 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@45 -- # '[' true = true ']' 00:04:55.284 23:38:43 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@46 -- # jq 'has("bdev")' 00:04:55.284 23:38:43 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@46 -- # '[' true = true ']' 00:04:55.284 23:38:43 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@47 -- # jq -r .bdev.tpoint_mask 00:04:55.284 ************************************ 00:04:55.284 END TEST rpc_trace_cmd_test 00:04:55.284 ************************************ 00:04:55.284 23:38:43 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@47 -- # '[' 0xffffffffffffffff '!=' 0x0 ']' 00:04:55.284 00:04:55.284 real 0m0.276s 00:04:55.284 user 0m0.216s 00:04:55.284 sys 0m0.048s 00:04:55.284 23:38:43 rpc.rpc_trace_cmd_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:04:55.284 23:38:43 rpc.rpc_trace_cmd_test -- common/autotest_common.sh@10 -- # set +x 00:04:55.284 23:38:43 rpc -- rpc/rpc.sh@76 -- # [[ 0 -eq 1 ]] 00:04:55.284 23:38:43 rpc -- rpc/rpc.sh@80 -- # rpc=rpc_cmd 00:04:55.284 23:38:43 rpc -- rpc/rpc.sh@81 -- # run_test rpc_daemon_integrity rpc_integrity 00:04:55.284 23:38:43 rpc -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:04:55.284 23:38:43 rpc -- common/autotest_common.sh@1111 -- # xtrace_disable 00:04:55.284 23:38:43 rpc -- common/autotest_common.sh@10 -- # set +x 00:04:55.284 ************************************ 00:04:55.284 START TEST rpc_daemon_integrity 00:04:55.284 ************************************ 00:04:55.284 23:38:43 rpc.rpc_daemon_integrity -- common/autotest_common.sh@1129 -- # rpc_integrity 00:04:55.284 23:38:43 rpc.rpc_daemon_integrity -- rpc/rpc.sh@12 -- # rpc_cmd bdev_get_bdevs 00:04:55.284 23:38:43 rpc.rpc_daemon_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:04:55.284 23:38:43 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:04:55.284 23:38:43 rpc.rpc_daemon_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:04:55.284 23:38:43 rpc.rpc_daemon_integrity -- rpc/rpc.sh@12 -- # bdevs='[]' 00:04:55.284 23:38:43 rpc.rpc_daemon_integrity -- rpc/rpc.sh@13 -- # jq length 00:04:55.544 23:38:43 rpc.rpc_daemon_integrity -- rpc/rpc.sh@13 -- # '[' 0 == 0 ']' 00:04:55.544 23:38:43 rpc.rpc_daemon_integrity -- rpc/rpc.sh@15 -- # rpc_cmd bdev_malloc_create 8 512 00:04:55.544 23:38:43 rpc.rpc_daemon_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:04:55.544 23:38:43 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:04:55.544 23:38:43 rpc.rpc_daemon_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:04:55.544 23:38:43 rpc.rpc_daemon_integrity -- rpc/rpc.sh@15 -- # malloc=Malloc2 00:04:55.544 23:38:43 rpc.rpc_daemon_integrity -- rpc/rpc.sh@16 -- # rpc_cmd bdev_get_bdevs 00:04:55.544 23:38:43 rpc.rpc_daemon_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:04:55.544 23:38:43 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:04:55.544 23:38:43 rpc.rpc_daemon_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:04:55.544 23:38:43 rpc.rpc_daemon_integrity -- rpc/rpc.sh@16 -- # bdevs='[ 00:04:55.544 { 00:04:55.544 "name": "Malloc2", 00:04:55.544 "aliases": [ 00:04:55.544 "9b283a46-2e36-4e41-8571-3b3886434de0" 00:04:55.544 ], 00:04:55.544 "product_name": "Malloc disk", 00:04:55.544 "block_size": 512, 00:04:55.544 "num_blocks": 16384, 00:04:55.544 "uuid": "9b283a46-2e36-4e41-8571-3b3886434de0", 00:04:55.544 "assigned_rate_limits": { 00:04:55.544 "rw_ios_per_sec": 0, 00:04:55.544 "rw_mbytes_per_sec": 0, 00:04:55.544 "r_mbytes_per_sec": 0, 00:04:55.544 "w_mbytes_per_sec": 0 00:04:55.544 }, 00:04:55.544 "claimed": false, 00:04:55.544 "zoned": false, 00:04:55.544 "supported_io_types": { 00:04:55.544 "read": true, 00:04:55.544 "write": true, 00:04:55.544 "unmap": true, 00:04:55.544 "flush": true, 00:04:55.544 "reset": true, 00:04:55.544 "nvme_admin": false, 00:04:55.544 "nvme_io": false, 00:04:55.544 "nvme_io_md": false, 00:04:55.544 "write_zeroes": true, 00:04:55.544 "zcopy": true, 00:04:55.544 "get_zone_info": false, 00:04:55.544 "zone_management": false, 00:04:55.544 "zone_append": false, 00:04:55.544 "compare": false, 00:04:55.544 "compare_and_write": false, 00:04:55.544 "abort": true, 00:04:55.544 "seek_hole": false, 00:04:55.544 "seek_data": false, 00:04:55.544 "copy": true, 00:04:55.544 "nvme_iov_md": false 00:04:55.544 }, 00:04:55.544 "memory_domains": [ 00:04:55.544 { 00:04:55.544 "dma_device_id": "system", 00:04:55.544 "dma_device_type": 1 00:04:55.544 }, 00:04:55.544 { 00:04:55.544 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:04:55.544 "dma_device_type": 2 00:04:55.544 } 00:04:55.544 ], 00:04:55.544 "driver_specific": {} 00:04:55.544 } 00:04:55.544 ]' 00:04:55.544 23:38:43 rpc.rpc_daemon_integrity -- rpc/rpc.sh@17 -- # jq length 00:04:55.544 23:38:43 rpc.rpc_daemon_integrity -- rpc/rpc.sh@17 -- # '[' 1 == 1 ']' 00:04:55.544 23:38:43 rpc.rpc_daemon_integrity -- rpc/rpc.sh@19 -- # rpc_cmd bdev_passthru_create -b Malloc2 -p Passthru0 00:04:55.544 23:38:43 rpc.rpc_daemon_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:04:55.544 23:38:43 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:04:55.544 [2024-11-26 23:38:43.542537] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on Malloc2 00:04:55.544 [2024-11-26 23:38:43.542697] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:04:55.544 [2024-11-26 23:38:43.542728] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:04:55.544 [2024-11-26 23:38:43.542738] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:04:55.544 [2024-11-26 23:38:43.545018] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:04:55.544 [2024-11-26 23:38:43.545058] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: Passthru0 00:04:55.544 Passthru0 00:04:55.544 23:38:43 rpc.rpc_daemon_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:04:55.544 23:38:43 rpc.rpc_daemon_integrity -- rpc/rpc.sh@20 -- # rpc_cmd bdev_get_bdevs 00:04:55.544 23:38:43 rpc.rpc_daemon_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:04:55.544 23:38:43 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:04:55.544 23:38:43 rpc.rpc_daemon_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:04:55.544 23:38:43 rpc.rpc_daemon_integrity -- rpc/rpc.sh@20 -- # bdevs='[ 00:04:55.544 { 00:04:55.544 "name": "Malloc2", 00:04:55.544 "aliases": [ 00:04:55.544 "9b283a46-2e36-4e41-8571-3b3886434de0" 00:04:55.544 ], 00:04:55.544 "product_name": "Malloc disk", 00:04:55.544 "block_size": 512, 00:04:55.544 "num_blocks": 16384, 00:04:55.544 "uuid": "9b283a46-2e36-4e41-8571-3b3886434de0", 00:04:55.544 "assigned_rate_limits": { 00:04:55.544 "rw_ios_per_sec": 0, 00:04:55.544 "rw_mbytes_per_sec": 0, 00:04:55.544 "r_mbytes_per_sec": 0, 00:04:55.544 "w_mbytes_per_sec": 0 00:04:55.544 }, 00:04:55.544 "claimed": true, 00:04:55.544 "claim_type": "exclusive_write", 00:04:55.544 "zoned": false, 00:04:55.544 "supported_io_types": { 00:04:55.544 "read": true, 00:04:55.544 "write": true, 00:04:55.544 "unmap": true, 00:04:55.544 "flush": true, 00:04:55.544 "reset": true, 00:04:55.544 "nvme_admin": false, 00:04:55.544 "nvme_io": false, 00:04:55.544 "nvme_io_md": false, 00:04:55.544 "write_zeroes": true, 00:04:55.544 "zcopy": true, 00:04:55.544 "get_zone_info": false, 00:04:55.544 "zone_management": false, 00:04:55.544 "zone_append": false, 00:04:55.544 "compare": false, 00:04:55.544 "compare_and_write": false, 00:04:55.544 "abort": true, 00:04:55.544 "seek_hole": false, 00:04:55.544 "seek_data": false, 00:04:55.544 "copy": true, 00:04:55.544 "nvme_iov_md": false 00:04:55.544 }, 00:04:55.544 "memory_domains": [ 00:04:55.544 { 00:04:55.544 "dma_device_id": "system", 00:04:55.544 "dma_device_type": 1 00:04:55.544 }, 00:04:55.544 { 00:04:55.544 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:04:55.544 "dma_device_type": 2 00:04:55.544 } 00:04:55.544 ], 00:04:55.544 "driver_specific": {} 00:04:55.544 }, 00:04:55.544 { 00:04:55.544 "name": "Passthru0", 00:04:55.544 "aliases": [ 00:04:55.544 "41faa32d-f102-5970-aa0d-edb5413172af" 00:04:55.544 ], 00:04:55.544 "product_name": "passthru", 00:04:55.544 "block_size": 512, 00:04:55.544 "num_blocks": 16384, 00:04:55.544 "uuid": "41faa32d-f102-5970-aa0d-edb5413172af", 00:04:55.544 "assigned_rate_limits": { 00:04:55.544 "rw_ios_per_sec": 0, 00:04:55.544 "rw_mbytes_per_sec": 0, 00:04:55.544 "r_mbytes_per_sec": 0, 00:04:55.544 "w_mbytes_per_sec": 0 00:04:55.544 }, 00:04:55.544 "claimed": false, 00:04:55.544 "zoned": false, 00:04:55.544 "supported_io_types": { 00:04:55.544 "read": true, 00:04:55.544 "write": true, 00:04:55.544 "unmap": true, 00:04:55.544 "flush": true, 00:04:55.544 "reset": true, 00:04:55.545 "nvme_admin": false, 00:04:55.545 "nvme_io": false, 00:04:55.545 "nvme_io_md": false, 00:04:55.545 "write_zeroes": true, 00:04:55.545 "zcopy": true, 00:04:55.545 "get_zone_info": false, 00:04:55.545 "zone_management": false, 00:04:55.545 "zone_append": false, 00:04:55.545 "compare": false, 00:04:55.545 "compare_and_write": false, 00:04:55.545 "abort": true, 00:04:55.545 "seek_hole": false, 00:04:55.545 "seek_data": false, 00:04:55.545 "copy": true, 00:04:55.545 "nvme_iov_md": false 00:04:55.545 }, 00:04:55.545 "memory_domains": [ 00:04:55.545 { 00:04:55.545 "dma_device_id": "system", 00:04:55.545 "dma_device_type": 1 00:04:55.545 }, 00:04:55.545 { 00:04:55.545 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:04:55.545 "dma_device_type": 2 00:04:55.545 } 00:04:55.545 ], 00:04:55.545 "driver_specific": { 00:04:55.545 "passthru": { 00:04:55.545 "name": "Passthru0", 00:04:55.545 "base_bdev_name": "Malloc2" 00:04:55.545 } 00:04:55.545 } 00:04:55.545 } 00:04:55.545 ]' 00:04:55.545 23:38:43 rpc.rpc_daemon_integrity -- rpc/rpc.sh@21 -- # jq length 00:04:55.545 23:38:43 rpc.rpc_daemon_integrity -- rpc/rpc.sh@21 -- # '[' 2 == 2 ']' 00:04:55.545 23:38:43 rpc.rpc_daemon_integrity -- rpc/rpc.sh@23 -- # rpc_cmd bdev_passthru_delete Passthru0 00:04:55.545 23:38:43 rpc.rpc_daemon_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:04:55.545 23:38:43 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:04:55.545 23:38:43 rpc.rpc_daemon_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:04:55.545 23:38:43 rpc.rpc_daemon_integrity -- rpc/rpc.sh@24 -- # rpc_cmd bdev_malloc_delete Malloc2 00:04:55.545 23:38:43 rpc.rpc_daemon_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:04:55.545 23:38:43 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:04:55.545 23:38:43 rpc.rpc_daemon_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:04:55.545 23:38:43 rpc.rpc_daemon_integrity -- rpc/rpc.sh@25 -- # rpc_cmd bdev_get_bdevs 00:04:55.545 23:38:43 rpc.rpc_daemon_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:04:55.545 23:38:43 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:04:55.545 23:38:43 rpc.rpc_daemon_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:04:55.545 23:38:43 rpc.rpc_daemon_integrity -- rpc/rpc.sh@25 -- # bdevs='[]' 00:04:55.545 23:38:43 rpc.rpc_daemon_integrity -- rpc/rpc.sh@26 -- # jq length 00:04:55.805 23:38:43 rpc.rpc_daemon_integrity -- rpc/rpc.sh@26 -- # '[' 0 == 0 ']' 00:04:55.805 00:04:55.805 real 0m0.323s 00:04:55.805 user 0m0.199s 00:04:55.805 sys 0m0.058s 00:04:55.805 23:38:43 rpc.rpc_daemon_integrity -- common/autotest_common.sh@1130 -- # xtrace_disable 00:04:55.805 23:38:43 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:04:55.805 ************************************ 00:04:55.805 END TEST rpc_daemon_integrity 00:04:55.805 ************************************ 00:04:55.805 23:38:43 rpc -- rpc/rpc.sh@83 -- # trap - SIGINT SIGTERM EXIT 00:04:55.805 23:38:43 rpc -- rpc/rpc.sh@84 -- # killprocess 68854 00:04:55.805 23:38:43 rpc -- common/autotest_common.sh@954 -- # '[' -z 68854 ']' 00:04:55.805 23:38:43 rpc -- common/autotest_common.sh@958 -- # kill -0 68854 00:04:55.805 23:38:43 rpc -- common/autotest_common.sh@959 -- # uname 00:04:55.805 23:38:43 rpc -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:04:55.805 23:38:43 rpc -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 68854 00:04:55.805 killing process with pid 68854 00:04:55.805 23:38:43 rpc -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:04:55.805 23:38:43 rpc -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:04:55.805 23:38:43 rpc -- common/autotest_common.sh@972 -- # echo 'killing process with pid 68854' 00:04:55.805 23:38:43 rpc -- common/autotest_common.sh@973 -- # kill 68854 00:04:55.805 23:38:43 rpc -- common/autotest_common.sh@978 -- # wait 68854 00:04:56.063 00:04:56.063 real 0m2.910s 00:04:56.063 user 0m3.550s 00:04:56.063 sys 0m0.852s 00:04:56.063 ************************************ 00:04:56.063 END TEST rpc 00:04:56.063 23:38:44 rpc -- common/autotest_common.sh@1130 -- # xtrace_disable 00:04:56.063 23:38:44 rpc -- common/autotest_common.sh@10 -- # set +x 00:04:56.063 ************************************ 00:04:56.321 23:38:44 -- spdk/autotest.sh@157 -- # run_test skip_rpc /home/vagrant/spdk_repo/spdk/test/rpc/skip_rpc.sh 00:04:56.321 23:38:44 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:04:56.321 23:38:44 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:04:56.321 23:38:44 -- common/autotest_common.sh@10 -- # set +x 00:04:56.321 ************************************ 00:04:56.321 START TEST skip_rpc 00:04:56.321 ************************************ 00:04:56.321 23:38:44 skip_rpc -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/rpc/skip_rpc.sh 00:04:56.321 * Looking for test storage... 00:04:56.321 * Found test storage at /home/vagrant/spdk_repo/spdk/test/rpc 00:04:56.321 23:38:44 skip_rpc -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:04:56.321 23:38:44 skip_rpc -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:04:56.321 23:38:44 skip_rpc -- common/autotest_common.sh@1693 -- # lcov --version 00:04:56.321 23:38:44 skip_rpc -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:04:56.321 23:38:44 skip_rpc -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:04:56.321 23:38:44 skip_rpc -- scripts/common.sh@333 -- # local ver1 ver1_l 00:04:56.321 23:38:44 skip_rpc -- scripts/common.sh@334 -- # local ver2 ver2_l 00:04:56.321 23:38:44 skip_rpc -- scripts/common.sh@336 -- # IFS=.-: 00:04:56.321 23:38:44 skip_rpc -- scripts/common.sh@336 -- # read -ra ver1 00:04:56.321 23:38:44 skip_rpc -- scripts/common.sh@337 -- # IFS=.-: 00:04:56.321 23:38:44 skip_rpc -- scripts/common.sh@337 -- # read -ra ver2 00:04:56.321 23:38:44 skip_rpc -- scripts/common.sh@338 -- # local 'op=<' 00:04:56.321 23:38:44 skip_rpc -- scripts/common.sh@340 -- # ver1_l=2 00:04:56.321 23:38:44 skip_rpc -- scripts/common.sh@341 -- # ver2_l=1 00:04:56.321 23:38:44 skip_rpc -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:04:56.321 23:38:44 skip_rpc -- scripts/common.sh@344 -- # case "$op" in 00:04:56.321 23:38:44 skip_rpc -- scripts/common.sh@345 -- # : 1 00:04:56.321 23:38:44 skip_rpc -- scripts/common.sh@364 -- # (( v = 0 )) 00:04:56.322 23:38:44 skip_rpc -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:04:56.322 23:38:44 skip_rpc -- scripts/common.sh@365 -- # decimal 1 00:04:56.581 23:38:44 skip_rpc -- scripts/common.sh@353 -- # local d=1 00:04:56.581 23:38:44 skip_rpc -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:04:56.581 23:38:44 skip_rpc -- scripts/common.sh@355 -- # echo 1 00:04:56.581 23:38:44 skip_rpc -- scripts/common.sh@365 -- # ver1[v]=1 00:04:56.581 23:38:44 skip_rpc -- scripts/common.sh@366 -- # decimal 2 00:04:56.581 23:38:44 skip_rpc -- scripts/common.sh@353 -- # local d=2 00:04:56.581 23:38:44 skip_rpc -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:04:56.581 23:38:44 skip_rpc -- scripts/common.sh@355 -- # echo 2 00:04:56.581 23:38:44 skip_rpc -- scripts/common.sh@366 -- # ver2[v]=2 00:04:56.581 23:38:44 skip_rpc -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:04:56.581 23:38:44 skip_rpc -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:04:56.581 23:38:44 skip_rpc -- scripts/common.sh@368 -- # return 0 00:04:56.581 23:38:44 skip_rpc -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:04:56.581 23:38:44 skip_rpc -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:04:56.581 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:04:56.581 --rc genhtml_branch_coverage=1 00:04:56.581 --rc genhtml_function_coverage=1 00:04:56.581 --rc genhtml_legend=1 00:04:56.581 --rc geninfo_all_blocks=1 00:04:56.581 --rc geninfo_unexecuted_blocks=1 00:04:56.581 00:04:56.581 ' 00:04:56.581 23:38:44 skip_rpc -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:04:56.581 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:04:56.581 --rc genhtml_branch_coverage=1 00:04:56.581 --rc genhtml_function_coverage=1 00:04:56.581 --rc genhtml_legend=1 00:04:56.581 --rc geninfo_all_blocks=1 00:04:56.581 --rc geninfo_unexecuted_blocks=1 00:04:56.581 00:04:56.581 ' 00:04:56.581 23:38:44 skip_rpc -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:04:56.581 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:04:56.581 --rc genhtml_branch_coverage=1 00:04:56.581 --rc genhtml_function_coverage=1 00:04:56.581 --rc genhtml_legend=1 00:04:56.581 --rc geninfo_all_blocks=1 00:04:56.581 --rc geninfo_unexecuted_blocks=1 00:04:56.581 00:04:56.581 ' 00:04:56.581 23:38:44 skip_rpc -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:04:56.581 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:04:56.581 --rc genhtml_branch_coverage=1 00:04:56.581 --rc genhtml_function_coverage=1 00:04:56.581 --rc genhtml_legend=1 00:04:56.581 --rc geninfo_all_blocks=1 00:04:56.581 --rc geninfo_unexecuted_blocks=1 00:04:56.581 00:04:56.581 ' 00:04:56.581 23:38:44 skip_rpc -- rpc/skip_rpc.sh@11 -- # CONFIG_PATH=/home/vagrant/spdk_repo/spdk/test/rpc/config.json 00:04:56.581 23:38:44 skip_rpc -- rpc/skip_rpc.sh@12 -- # LOG_PATH=/home/vagrant/spdk_repo/spdk/test/rpc/log.txt 00:04:56.581 23:38:44 skip_rpc -- rpc/skip_rpc.sh@73 -- # run_test skip_rpc test_skip_rpc 00:04:56.581 23:38:44 skip_rpc -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:04:56.581 23:38:44 skip_rpc -- common/autotest_common.sh@1111 -- # xtrace_disable 00:04:56.581 23:38:44 skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:04:56.581 ************************************ 00:04:56.581 START TEST skip_rpc 00:04:56.581 ************************************ 00:04:56.581 23:38:44 skip_rpc.skip_rpc -- common/autotest_common.sh@1129 -- # test_skip_rpc 00:04:56.581 23:38:44 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@16 -- # local spdk_pid=69061 00:04:56.581 23:38:44 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@18 -- # trap 'killprocess $spdk_pid; exit 1' SIGINT SIGTERM EXIT 00:04:56.581 23:38:44 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@15 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt --no-rpc-server -m 0x1 00:04:56.581 23:38:44 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@19 -- # sleep 5 00:04:56.581 [2024-11-26 23:38:44.575870] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:04:56.581 [2024-11-26 23:38:44.576116] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid69061 ] 00:04:56.840 [2024-11-26 23:38:44.729931] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:04:56.840 [2024-11-26 23:38:44.758823] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:05:02.109 23:38:49 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@21 -- # NOT rpc_cmd spdk_get_version 00:05:02.109 23:38:49 skip_rpc.skip_rpc -- common/autotest_common.sh@652 -- # local es=0 00:05:02.109 23:38:49 skip_rpc.skip_rpc -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd spdk_get_version 00:05:02.109 23:38:49 skip_rpc.skip_rpc -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:05:02.109 23:38:49 skip_rpc.skip_rpc -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:05:02.109 23:38:49 skip_rpc.skip_rpc -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:05:02.109 23:38:49 skip_rpc.skip_rpc -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:05:02.109 23:38:49 skip_rpc.skip_rpc -- common/autotest_common.sh@655 -- # rpc_cmd spdk_get_version 00:05:02.109 23:38:49 skip_rpc.skip_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:02.109 23:38:49 skip_rpc.skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:05:02.109 23:38:49 skip_rpc.skip_rpc -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:05:02.109 23:38:49 skip_rpc.skip_rpc -- common/autotest_common.sh@655 -- # es=1 00:05:02.109 23:38:49 skip_rpc.skip_rpc -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:05:02.109 23:38:49 skip_rpc.skip_rpc -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:05:02.109 23:38:49 skip_rpc.skip_rpc -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:05:02.109 23:38:49 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@22 -- # trap - SIGINT SIGTERM EXIT 00:05:02.109 23:38:49 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@23 -- # killprocess 69061 00:05:02.109 23:38:49 skip_rpc.skip_rpc -- common/autotest_common.sh@954 -- # '[' -z 69061 ']' 00:05:02.109 23:38:49 skip_rpc.skip_rpc -- common/autotest_common.sh@958 -- # kill -0 69061 00:05:02.109 23:38:49 skip_rpc.skip_rpc -- common/autotest_common.sh@959 -- # uname 00:05:02.109 23:38:49 skip_rpc.skip_rpc -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:05:02.109 23:38:49 skip_rpc.skip_rpc -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 69061 00:05:02.109 killing process with pid 69061 00:05:02.109 23:38:49 skip_rpc.skip_rpc -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:05:02.109 23:38:49 skip_rpc.skip_rpc -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:05:02.109 23:38:49 skip_rpc.skip_rpc -- common/autotest_common.sh@972 -- # echo 'killing process with pid 69061' 00:05:02.109 23:38:49 skip_rpc.skip_rpc -- common/autotest_common.sh@973 -- # kill 69061 00:05:02.109 23:38:49 skip_rpc.skip_rpc -- common/autotest_common.sh@978 -- # wait 69061 00:05:02.109 00:05:02.109 real 0m5.425s 00:05:02.109 user 0m5.039s 00:05:02.109 sys 0m0.310s 00:05:02.109 23:38:49 skip_rpc.skip_rpc -- common/autotest_common.sh@1130 -- # xtrace_disable 00:05:02.109 ************************************ 00:05:02.109 END TEST skip_rpc 00:05:02.109 ************************************ 00:05:02.109 23:38:49 skip_rpc.skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:05:02.109 23:38:49 skip_rpc -- rpc/skip_rpc.sh@74 -- # run_test skip_rpc_with_json test_skip_rpc_with_json 00:05:02.109 23:38:49 skip_rpc -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:05:02.109 23:38:49 skip_rpc -- common/autotest_common.sh@1111 -- # xtrace_disable 00:05:02.109 23:38:49 skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:05:02.109 ************************************ 00:05:02.109 START TEST skip_rpc_with_json 00:05:02.109 ************************************ 00:05:02.109 23:38:49 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@1129 -- # test_skip_rpc_with_json 00:05:02.109 23:38:49 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@44 -- # gen_json_config 00:05:02.109 23:38:49 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@28 -- # local spdk_pid=69143 00:05:02.109 23:38:49 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@27 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 00:05:02.109 23:38:49 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@30 -- # trap 'killprocess $spdk_pid; exit 1' SIGINT SIGTERM EXIT 00:05:02.109 23:38:49 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@31 -- # waitforlisten 69143 00:05:02.109 23:38:49 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@835 -- # '[' -z 69143 ']' 00:05:02.109 23:38:49 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:05:02.109 23:38:49 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@840 -- # local max_retries=100 00:05:02.109 23:38:49 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:05:02.109 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:05:02.109 23:38:49 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@844 -- # xtrace_disable 00:05:02.109 23:38:49 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@10 -- # set +x 00:05:02.109 [2024-11-26 23:38:50.064153] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:05:02.109 [2024-11-26 23:38:50.064402] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid69143 ] 00:05:02.109 [2024-11-26 23:38:50.222183] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:05:02.368 [2024-11-26 23:38:50.251220] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:05:02.964 23:38:50 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:05:02.964 23:38:50 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@868 -- # return 0 00:05:02.964 23:38:50 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@34 -- # rpc_cmd nvmf_get_transports --trtype tcp 00:05:02.965 23:38:50 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:02.965 23:38:50 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@10 -- # set +x 00:05:02.965 [2024-11-26 23:38:50.936183] nvmf_rpc.c:2706:rpc_nvmf_get_transports: *ERROR*: transport 'tcp' does not exist 00:05:02.965 request: 00:05:02.965 { 00:05:02.965 "trtype": "tcp", 00:05:02.965 "method": "nvmf_get_transports", 00:05:02.965 "req_id": 1 00:05:02.965 } 00:05:02.965 Got JSON-RPC error response 00:05:02.965 response: 00:05:02.965 { 00:05:02.965 "code": -19, 00:05:02.965 "message": "No such device" 00:05:02.965 } 00:05:02.965 23:38:50 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:05:02.965 23:38:50 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@34 -- # rpc_cmd nvmf_create_transport -t tcp 00:05:02.965 23:38:50 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:02.965 23:38:50 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@10 -- # set +x 00:05:02.965 [2024-11-26 23:38:50.948322] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:05:02.965 23:38:50 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:02.965 23:38:50 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@36 -- # rpc_cmd save_config 00:05:02.965 23:38:50 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:02.965 23:38:50 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@10 -- # set +x 00:05:03.225 23:38:51 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:03.225 23:38:51 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@37 -- # cat /home/vagrant/spdk_repo/spdk/test/rpc/config.json 00:05:03.225 { 00:05:03.225 "subsystems": [ 00:05:03.225 { 00:05:03.225 "subsystem": "fsdev", 00:05:03.225 "config": [ 00:05:03.225 { 00:05:03.225 "method": "fsdev_set_opts", 00:05:03.225 "params": { 00:05:03.225 "fsdev_io_pool_size": 65535, 00:05:03.225 "fsdev_io_cache_size": 256 00:05:03.225 } 00:05:03.225 } 00:05:03.225 ] 00:05:03.225 }, 00:05:03.225 { 00:05:03.225 "subsystem": "keyring", 00:05:03.225 "config": [] 00:05:03.225 }, 00:05:03.225 { 00:05:03.225 "subsystem": "iobuf", 00:05:03.225 "config": [ 00:05:03.225 { 00:05:03.225 "method": "iobuf_set_options", 00:05:03.225 "params": { 00:05:03.225 "small_pool_count": 8192, 00:05:03.225 "large_pool_count": 1024, 00:05:03.225 "small_bufsize": 8192, 00:05:03.225 "large_bufsize": 135168, 00:05:03.225 "enable_numa": false 00:05:03.225 } 00:05:03.225 } 00:05:03.225 ] 00:05:03.225 }, 00:05:03.225 { 00:05:03.225 "subsystem": "sock", 00:05:03.225 "config": [ 00:05:03.225 { 00:05:03.225 "method": "sock_set_default_impl", 00:05:03.225 "params": { 00:05:03.225 "impl_name": "posix" 00:05:03.225 } 00:05:03.225 }, 00:05:03.225 { 00:05:03.225 "method": "sock_impl_set_options", 00:05:03.225 "params": { 00:05:03.225 "impl_name": "ssl", 00:05:03.225 "recv_buf_size": 4096, 00:05:03.225 "send_buf_size": 4096, 00:05:03.225 "enable_recv_pipe": true, 00:05:03.225 "enable_quickack": false, 00:05:03.225 "enable_placement_id": 0, 00:05:03.225 "enable_zerocopy_send_server": true, 00:05:03.225 "enable_zerocopy_send_client": false, 00:05:03.225 "zerocopy_threshold": 0, 00:05:03.225 "tls_version": 0, 00:05:03.225 "enable_ktls": false 00:05:03.225 } 00:05:03.225 }, 00:05:03.225 { 00:05:03.225 "method": "sock_impl_set_options", 00:05:03.225 "params": { 00:05:03.225 "impl_name": "posix", 00:05:03.225 "recv_buf_size": 2097152, 00:05:03.225 "send_buf_size": 2097152, 00:05:03.225 "enable_recv_pipe": true, 00:05:03.225 "enable_quickack": false, 00:05:03.225 "enable_placement_id": 0, 00:05:03.225 "enable_zerocopy_send_server": true, 00:05:03.225 "enable_zerocopy_send_client": false, 00:05:03.225 "zerocopy_threshold": 0, 00:05:03.225 "tls_version": 0, 00:05:03.225 "enable_ktls": false 00:05:03.225 } 00:05:03.225 } 00:05:03.225 ] 00:05:03.225 }, 00:05:03.225 { 00:05:03.225 "subsystem": "vmd", 00:05:03.225 "config": [] 00:05:03.225 }, 00:05:03.225 { 00:05:03.225 "subsystem": "accel", 00:05:03.225 "config": [ 00:05:03.225 { 00:05:03.225 "method": "accel_set_options", 00:05:03.225 "params": { 00:05:03.225 "small_cache_size": 128, 00:05:03.225 "large_cache_size": 16, 00:05:03.225 "task_count": 2048, 00:05:03.225 "sequence_count": 2048, 00:05:03.225 "buf_count": 2048 00:05:03.225 } 00:05:03.225 } 00:05:03.225 ] 00:05:03.225 }, 00:05:03.225 { 00:05:03.225 "subsystem": "bdev", 00:05:03.225 "config": [ 00:05:03.225 { 00:05:03.225 "method": "bdev_set_options", 00:05:03.225 "params": { 00:05:03.225 "bdev_io_pool_size": 65535, 00:05:03.225 "bdev_io_cache_size": 256, 00:05:03.225 "bdev_auto_examine": true, 00:05:03.225 "iobuf_small_cache_size": 128, 00:05:03.225 "iobuf_large_cache_size": 16 00:05:03.225 } 00:05:03.226 }, 00:05:03.226 { 00:05:03.226 "method": "bdev_raid_set_options", 00:05:03.226 "params": { 00:05:03.226 "process_window_size_kb": 1024, 00:05:03.226 "process_max_bandwidth_mb_sec": 0 00:05:03.226 } 00:05:03.226 }, 00:05:03.226 { 00:05:03.226 "method": "bdev_iscsi_set_options", 00:05:03.226 "params": { 00:05:03.226 "timeout_sec": 30 00:05:03.226 } 00:05:03.226 }, 00:05:03.226 { 00:05:03.226 "method": "bdev_nvme_set_options", 00:05:03.226 "params": { 00:05:03.226 "action_on_timeout": "none", 00:05:03.226 "timeout_us": 0, 00:05:03.226 "timeout_admin_us": 0, 00:05:03.226 "keep_alive_timeout_ms": 10000, 00:05:03.226 "arbitration_burst": 0, 00:05:03.226 "low_priority_weight": 0, 00:05:03.226 "medium_priority_weight": 0, 00:05:03.226 "high_priority_weight": 0, 00:05:03.226 "nvme_adminq_poll_period_us": 10000, 00:05:03.226 "nvme_ioq_poll_period_us": 0, 00:05:03.226 "io_queue_requests": 0, 00:05:03.226 "delay_cmd_submit": true, 00:05:03.226 "transport_retry_count": 4, 00:05:03.226 "bdev_retry_count": 3, 00:05:03.226 "transport_ack_timeout": 0, 00:05:03.226 "ctrlr_loss_timeout_sec": 0, 00:05:03.226 "reconnect_delay_sec": 0, 00:05:03.226 "fast_io_fail_timeout_sec": 0, 00:05:03.226 "disable_auto_failback": false, 00:05:03.226 "generate_uuids": false, 00:05:03.226 "transport_tos": 0, 00:05:03.226 "nvme_error_stat": false, 00:05:03.226 "rdma_srq_size": 0, 00:05:03.226 "io_path_stat": false, 00:05:03.226 "allow_accel_sequence": false, 00:05:03.226 "rdma_max_cq_size": 0, 00:05:03.226 "rdma_cm_event_timeout_ms": 0, 00:05:03.226 "dhchap_digests": [ 00:05:03.226 "sha256", 00:05:03.226 "sha384", 00:05:03.226 "sha512" 00:05:03.226 ], 00:05:03.226 "dhchap_dhgroups": [ 00:05:03.226 "null", 00:05:03.226 "ffdhe2048", 00:05:03.226 "ffdhe3072", 00:05:03.226 "ffdhe4096", 00:05:03.226 "ffdhe6144", 00:05:03.226 "ffdhe8192" 00:05:03.226 ] 00:05:03.226 } 00:05:03.226 }, 00:05:03.226 { 00:05:03.226 "method": "bdev_nvme_set_hotplug", 00:05:03.226 "params": { 00:05:03.226 "period_us": 100000, 00:05:03.226 "enable": false 00:05:03.226 } 00:05:03.226 }, 00:05:03.226 { 00:05:03.226 "method": "bdev_wait_for_examine" 00:05:03.226 } 00:05:03.226 ] 00:05:03.226 }, 00:05:03.226 { 00:05:03.226 "subsystem": "scsi", 00:05:03.226 "config": null 00:05:03.226 }, 00:05:03.226 { 00:05:03.226 "subsystem": "scheduler", 00:05:03.226 "config": [ 00:05:03.226 { 00:05:03.226 "method": "framework_set_scheduler", 00:05:03.226 "params": { 00:05:03.226 "name": "static" 00:05:03.226 } 00:05:03.226 } 00:05:03.226 ] 00:05:03.226 }, 00:05:03.226 { 00:05:03.226 "subsystem": "vhost_scsi", 00:05:03.226 "config": [] 00:05:03.226 }, 00:05:03.226 { 00:05:03.226 "subsystem": "vhost_blk", 00:05:03.226 "config": [] 00:05:03.226 }, 00:05:03.226 { 00:05:03.226 "subsystem": "ublk", 00:05:03.226 "config": [] 00:05:03.226 }, 00:05:03.226 { 00:05:03.226 "subsystem": "nbd", 00:05:03.226 "config": [] 00:05:03.226 }, 00:05:03.226 { 00:05:03.226 "subsystem": "nvmf", 00:05:03.226 "config": [ 00:05:03.226 { 00:05:03.226 "method": "nvmf_set_config", 00:05:03.226 "params": { 00:05:03.226 "discovery_filter": "match_any", 00:05:03.226 "admin_cmd_passthru": { 00:05:03.226 "identify_ctrlr": false 00:05:03.226 }, 00:05:03.226 "dhchap_digests": [ 00:05:03.226 "sha256", 00:05:03.226 "sha384", 00:05:03.226 "sha512" 00:05:03.226 ], 00:05:03.226 "dhchap_dhgroups": [ 00:05:03.226 "null", 00:05:03.226 "ffdhe2048", 00:05:03.226 "ffdhe3072", 00:05:03.226 "ffdhe4096", 00:05:03.226 "ffdhe6144", 00:05:03.226 "ffdhe8192" 00:05:03.226 ] 00:05:03.226 } 00:05:03.226 }, 00:05:03.226 { 00:05:03.226 "method": "nvmf_set_max_subsystems", 00:05:03.226 "params": { 00:05:03.226 "max_subsystems": 1024 00:05:03.226 } 00:05:03.226 }, 00:05:03.226 { 00:05:03.226 "method": "nvmf_set_crdt", 00:05:03.226 "params": { 00:05:03.226 "crdt1": 0, 00:05:03.226 "crdt2": 0, 00:05:03.226 "crdt3": 0 00:05:03.226 } 00:05:03.226 }, 00:05:03.226 { 00:05:03.226 "method": "nvmf_create_transport", 00:05:03.226 "params": { 00:05:03.226 "trtype": "TCP", 00:05:03.226 "max_queue_depth": 128, 00:05:03.226 "max_io_qpairs_per_ctrlr": 127, 00:05:03.226 "in_capsule_data_size": 4096, 00:05:03.226 "max_io_size": 131072, 00:05:03.226 "io_unit_size": 131072, 00:05:03.226 "max_aq_depth": 128, 00:05:03.226 "num_shared_buffers": 511, 00:05:03.226 "buf_cache_size": 4294967295, 00:05:03.226 "dif_insert_or_strip": false, 00:05:03.226 "zcopy": false, 00:05:03.226 "c2h_success": true, 00:05:03.226 "sock_priority": 0, 00:05:03.226 "abort_timeout_sec": 1, 00:05:03.226 "ack_timeout": 0, 00:05:03.226 "data_wr_pool_size": 0 00:05:03.226 } 00:05:03.226 } 00:05:03.226 ] 00:05:03.226 }, 00:05:03.226 { 00:05:03.226 "subsystem": "iscsi", 00:05:03.226 "config": [ 00:05:03.226 { 00:05:03.226 "method": "iscsi_set_options", 00:05:03.226 "params": { 00:05:03.226 "node_base": "iqn.2016-06.io.spdk", 00:05:03.226 "max_sessions": 128, 00:05:03.226 "max_connections_per_session": 2, 00:05:03.226 "max_queue_depth": 64, 00:05:03.226 "default_time2wait": 2, 00:05:03.226 "default_time2retain": 20, 00:05:03.226 "first_burst_length": 8192, 00:05:03.226 "immediate_data": true, 00:05:03.226 "allow_duplicated_isid": false, 00:05:03.226 "error_recovery_level": 0, 00:05:03.226 "nop_timeout": 60, 00:05:03.226 "nop_in_interval": 30, 00:05:03.226 "disable_chap": false, 00:05:03.226 "require_chap": false, 00:05:03.226 "mutual_chap": false, 00:05:03.226 "chap_group": 0, 00:05:03.226 "max_large_datain_per_connection": 64, 00:05:03.226 "max_r2t_per_connection": 4, 00:05:03.226 "pdu_pool_size": 36864, 00:05:03.226 "immediate_data_pool_size": 16384, 00:05:03.226 "data_out_pool_size": 2048 00:05:03.226 } 00:05:03.226 } 00:05:03.226 ] 00:05:03.226 } 00:05:03.226 ] 00:05:03.226 } 00:05:03.226 23:38:51 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@39 -- # trap - SIGINT SIGTERM EXIT 00:05:03.226 23:38:51 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@40 -- # killprocess 69143 00:05:03.226 23:38:51 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@954 -- # '[' -z 69143 ']' 00:05:03.226 23:38:51 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@958 -- # kill -0 69143 00:05:03.226 23:38:51 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@959 -- # uname 00:05:03.226 23:38:51 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:05:03.226 23:38:51 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 69143 00:05:03.226 23:38:51 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:05:03.226 23:38:51 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:05:03.226 23:38:51 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@972 -- # echo 'killing process with pid 69143' 00:05:03.226 killing process with pid 69143 00:05:03.226 23:38:51 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@973 -- # kill 69143 00:05:03.226 23:38:51 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@978 -- # wait 69143 00:05:03.486 23:38:51 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@47 -- # local spdk_pid=69177 00:05:03.486 23:38:51 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@46 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt --no-rpc-server -m 0x1 --json /home/vagrant/spdk_repo/spdk/test/rpc/config.json 00:05:03.486 23:38:51 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@48 -- # sleep 5 00:05:08.762 23:38:56 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@50 -- # killprocess 69177 00:05:08.762 23:38:56 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@954 -- # '[' -z 69177 ']' 00:05:08.762 23:38:56 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@958 -- # kill -0 69177 00:05:08.762 23:38:56 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@959 -- # uname 00:05:08.762 23:38:56 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:05:08.762 23:38:56 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 69177 00:05:08.762 23:38:56 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:05:08.763 23:38:56 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:05:08.763 23:38:56 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@972 -- # echo 'killing process with pid 69177' 00:05:08.763 killing process with pid 69177 00:05:08.763 23:38:56 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@973 -- # kill 69177 00:05:08.763 23:38:56 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@978 -- # wait 69177 00:05:09.022 23:38:56 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@51 -- # grep -q 'TCP Transport Init' /home/vagrant/spdk_repo/spdk/test/rpc/log.txt 00:05:09.022 23:38:56 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@52 -- # rm /home/vagrant/spdk_repo/spdk/test/rpc/log.txt 00:05:09.022 00:05:09.022 real 0m6.972s 00:05:09.022 user 0m6.595s 00:05:09.022 sys 0m0.702s 00:05:09.022 ************************************ 00:05:09.022 END TEST skip_rpc_with_json 00:05:09.022 ************************************ 00:05:09.022 23:38:56 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@1130 -- # xtrace_disable 00:05:09.022 23:38:56 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@10 -- # set +x 00:05:09.022 23:38:56 skip_rpc -- rpc/skip_rpc.sh@75 -- # run_test skip_rpc_with_delay test_skip_rpc_with_delay 00:05:09.023 23:38:56 skip_rpc -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:05:09.023 23:38:56 skip_rpc -- common/autotest_common.sh@1111 -- # xtrace_disable 00:05:09.023 23:38:56 skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:05:09.023 ************************************ 00:05:09.023 START TEST skip_rpc_with_delay 00:05:09.023 ************************************ 00:05:09.023 23:38:57 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@1129 -- # test_skip_rpc_with_delay 00:05:09.023 23:38:57 skip_rpc.skip_rpc_with_delay -- rpc/skip_rpc.sh@57 -- # NOT /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt --no-rpc-server -m 0x1 --wait-for-rpc 00:05:09.023 23:38:57 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@652 -- # local es=0 00:05:09.023 23:38:57 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@654 -- # valid_exec_arg /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt --no-rpc-server -m 0x1 --wait-for-rpc 00:05:09.023 23:38:57 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@640 -- # local arg=/home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:05:09.023 23:38:57 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:05:09.023 23:38:57 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@644 -- # type -t /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:05:09.023 23:38:57 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:05:09.023 23:38:57 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@646 -- # type -P /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:05:09.023 23:38:57 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:05:09.023 23:38:57 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@646 -- # arg=/home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:05:09.023 23:38:57 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@646 -- # [[ -x /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt ]] 00:05:09.023 23:38:57 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@655 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt --no-rpc-server -m 0x1 --wait-for-rpc 00:05:09.023 [2024-11-26 23:38:57.104970] app.c: 842:spdk_app_start: *ERROR*: Cannot use '--wait-for-rpc' if no RPC server is going to be started. 00:05:09.283 23:38:57 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@655 -- # es=1 00:05:09.283 23:38:57 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:05:09.283 23:38:57 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:05:09.283 23:38:57 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:05:09.283 00:05:09.283 real 0m0.164s 00:05:09.283 user 0m0.084s 00:05:09.283 sys 0m0.078s 00:05:09.283 23:38:57 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@1130 -- # xtrace_disable 00:05:09.283 ************************************ 00:05:09.283 END TEST skip_rpc_with_delay 00:05:09.283 ************************************ 00:05:09.283 23:38:57 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@10 -- # set +x 00:05:09.283 23:38:57 skip_rpc -- rpc/skip_rpc.sh@77 -- # uname 00:05:09.283 23:38:57 skip_rpc -- rpc/skip_rpc.sh@77 -- # '[' Linux '!=' FreeBSD ']' 00:05:09.283 23:38:57 skip_rpc -- rpc/skip_rpc.sh@78 -- # run_test exit_on_failed_rpc_init test_exit_on_failed_rpc_init 00:05:09.283 23:38:57 skip_rpc -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:05:09.283 23:38:57 skip_rpc -- common/autotest_common.sh@1111 -- # xtrace_disable 00:05:09.283 23:38:57 skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:05:09.283 ************************************ 00:05:09.283 START TEST exit_on_failed_rpc_init 00:05:09.283 ************************************ 00:05:09.283 23:38:57 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@1129 -- # test_exit_on_failed_rpc_init 00:05:09.283 23:38:57 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@62 -- # local spdk_pid=69283 00:05:09.283 23:38:57 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@61 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 00:05:09.283 23:38:57 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@63 -- # waitforlisten 69283 00:05:09.283 23:38:57 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@835 -- # '[' -z 69283 ']' 00:05:09.283 23:38:57 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:05:09.283 23:38:57 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@840 -- # local max_retries=100 00:05:09.283 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:05:09.283 23:38:57 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:05:09.283 23:38:57 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@844 -- # xtrace_disable 00:05:09.283 23:38:57 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@10 -- # set +x 00:05:09.283 [2024-11-26 23:38:57.335887] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:05:09.283 [2024-11-26 23:38:57.336022] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid69283 ] 00:05:09.543 [2024-11-26 23:38:57.492374] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:05:09.543 [2024-11-26 23:38:57.520972] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:05:10.137 23:38:58 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:05:10.137 23:38:58 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@868 -- # return 0 00:05:10.137 23:38:58 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@65 -- # trap 'killprocess $spdk_pid; exit 1' SIGINT SIGTERM EXIT 00:05:10.137 23:38:58 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@67 -- # NOT /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x2 00:05:10.137 23:38:58 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@652 -- # local es=0 00:05:10.137 23:38:58 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@654 -- # valid_exec_arg /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x2 00:05:10.137 23:38:58 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@640 -- # local arg=/home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:05:10.137 23:38:58 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:05:10.137 23:38:58 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@644 -- # type -t /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:05:10.137 23:38:58 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:05:10.137 23:38:58 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@646 -- # type -P /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:05:10.137 23:38:58 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:05:10.137 23:38:58 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@646 -- # arg=/home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:05:10.137 23:38:58 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@646 -- # [[ -x /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt ]] 00:05:10.137 23:38:58 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@655 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x2 00:05:10.396 [2024-11-26 23:38:58.281132] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:05:10.397 [2024-11-26 23:38:58.281376] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid69301 ] 00:05:10.397 [2024-11-26 23:38:58.437269] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:05:10.397 [2024-11-26 23:38:58.466828] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:05:10.397 [2024-11-26 23:38:58.467025] rpc.c: 180:_spdk_rpc_listen: *ERROR*: RPC Unix domain socket path /var/tmp/spdk.sock in use. Specify another. 00:05:10.397 [2024-11-26 23:38:58.467079] rpc.c: 166:spdk_rpc_initialize: *ERROR*: Unable to start RPC service at /var/tmp/spdk.sock 00:05:10.397 [2024-11-26 23:38:58.467102] app.c:1064:spdk_app_stop: *WARNING*: spdk_app_stop'd on non-zero 00:05:10.656 23:38:58 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@655 -- # es=234 00:05:10.656 23:38:58 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:05:10.656 23:38:58 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@664 -- # es=106 00:05:10.656 23:38:58 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@665 -- # case "$es" in 00:05:10.656 23:38:58 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@672 -- # es=1 00:05:10.656 23:38:58 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:05:10.656 23:38:58 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@69 -- # trap - SIGINT SIGTERM EXIT 00:05:10.656 23:38:58 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@70 -- # killprocess 69283 00:05:10.656 23:38:58 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@954 -- # '[' -z 69283 ']' 00:05:10.656 23:38:58 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@958 -- # kill -0 69283 00:05:10.656 23:38:58 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@959 -- # uname 00:05:10.656 23:38:58 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:05:10.656 23:38:58 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 69283 00:05:10.656 23:38:58 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:05:10.656 23:38:58 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:05:10.656 killing process with pid 69283 00:05:10.656 23:38:58 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@972 -- # echo 'killing process with pid 69283' 00:05:10.656 23:38:58 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@973 -- # kill 69283 00:05:10.656 23:38:58 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@978 -- # wait 69283 00:05:10.915 00:05:10.915 real 0m1.737s 00:05:10.915 user 0m1.883s 00:05:10.915 sys 0m0.492s 00:05:10.915 ************************************ 00:05:10.915 END TEST exit_on_failed_rpc_init 00:05:10.915 ************************************ 00:05:10.915 23:38:58 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@1130 -- # xtrace_disable 00:05:10.915 23:38:58 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@10 -- # set +x 00:05:10.915 23:38:59 skip_rpc -- rpc/skip_rpc.sh@81 -- # rm /home/vagrant/spdk_repo/spdk/test/rpc/config.json 00:05:10.915 00:05:10.915 real 0m14.792s 00:05:10.915 user 0m13.815s 00:05:10.915 sys 0m1.877s 00:05:10.915 23:38:59 skip_rpc -- common/autotest_common.sh@1130 -- # xtrace_disable 00:05:10.915 23:38:59 skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:05:10.915 ************************************ 00:05:10.915 END TEST skip_rpc 00:05:10.915 ************************************ 00:05:11.174 23:38:59 -- spdk/autotest.sh@158 -- # run_test rpc_client /home/vagrant/spdk_repo/spdk/test/rpc_client/rpc_client.sh 00:05:11.174 23:38:59 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:05:11.174 23:38:59 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:05:11.174 23:38:59 -- common/autotest_common.sh@10 -- # set +x 00:05:11.174 ************************************ 00:05:11.174 START TEST rpc_client 00:05:11.174 ************************************ 00:05:11.174 23:38:59 rpc_client -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/rpc_client/rpc_client.sh 00:05:11.174 * Looking for test storage... 00:05:11.174 * Found test storage at /home/vagrant/spdk_repo/spdk/test/rpc_client 00:05:11.174 23:38:59 rpc_client -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:05:11.174 23:38:59 rpc_client -- common/autotest_common.sh@1693 -- # lcov --version 00:05:11.174 23:38:59 rpc_client -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:05:11.434 23:38:59 rpc_client -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:05:11.434 23:38:59 rpc_client -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:05:11.434 23:38:59 rpc_client -- scripts/common.sh@333 -- # local ver1 ver1_l 00:05:11.434 23:38:59 rpc_client -- scripts/common.sh@334 -- # local ver2 ver2_l 00:05:11.434 23:38:59 rpc_client -- scripts/common.sh@336 -- # IFS=.-: 00:05:11.434 23:38:59 rpc_client -- scripts/common.sh@336 -- # read -ra ver1 00:05:11.434 23:38:59 rpc_client -- scripts/common.sh@337 -- # IFS=.-: 00:05:11.434 23:38:59 rpc_client -- scripts/common.sh@337 -- # read -ra ver2 00:05:11.434 23:38:59 rpc_client -- scripts/common.sh@338 -- # local 'op=<' 00:05:11.434 23:38:59 rpc_client -- scripts/common.sh@340 -- # ver1_l=2 00:05:11.434 23:38:59 rpc_client -- scripts/common.sh@341 -- # ver2_l=1 00:05:11.434 23:38:59 rpc_client -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:05:11.434 23:38:59 rpc_client -- scripts/common.sh@344 -- # case "$op" in 00:05:11.434 23:38:59 rpc_client -- scripts/common.sh@345 -- # : 1 00:05:11.434 23:38:59 rpc_client -- scripts/common.sh@364 -- # (( v = 0 )) 00:05:11.434 23:38:59 rpc_client -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:05:11.434 23:38:59 rpc_client -- scripts/common.sh@365 -- # decimal 1 00:05:11.434 23:38:59 rpc_client -- scripts/common.sh@353 -- # local d=1 00:05:11.434 23:38:59 rpc_client -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:05:11.434 23:38:59 rpc_client -- scripts/common.sh@355 -- # echo 1 00:05:11.434 23:38:59 rpc_client -- scripts/common.sh@365 -- # ver1[v]=1 00:05:11.434 23:38:59 rpc_client -- scripts/common.sh@366 -- # decimal 2 00:05:11.434 23:38:59 rpc_client -- scripts/common.sh@353 -- # local d=2 00:05:11.434 23:38:59 rpc_client -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:05:11.434 23:38:59 rpc_client -- scripts/common.sh@355 -- # echo 2 00:05:11.434 23:38:59 rpc_client -- scripts/common.sh@366 -- # ver2[v]=2 00:05:11.434 23:38:59 rpc_client -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:05:11.434 23:38:59 rpc_client -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:05:11.434 23:38:59 rpc_client -- scripts/common.sh@368 -- # return 0 00:05:11.434 23:38:59 rpc_client -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:05:11.434 23:38:59 rpc_client -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:05:11.434 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:11.434 --rc genhtml_branch_coverage=1 00:05:11.434 --rc genhtml_function_coverage=1 00:05:11.434 --rc genhtml_legend=1 00:05:11.434 --rc geninfo_all_blocks=1 00:05:11.434 --rc geninfo_unexecuted_blocks=1 00:05:11.434 00:05:11.434 ' 00:05:11.434 23:38:59 rpc_client -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:05:11.434 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:11.434 --rc genhtml_branch_coverage=1 00:05:11.434 --rc genhtml_function_coverage=1 00:05:11.434 --rc genhtml_legend=1 00:05:11.434 --rc geninfo_all_blocks=1 00:05:11.434 --rc geninfo_unexecuted_blocks=1 00:05:11.434 00:05:11.434 ' 00:05:11.434 23:38:59 rpc_client -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:05:11.434 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:11.434 --rc genhtml_branch_coverage=1 00:05:11.434 --rc genhtml_function_coverage=1 00:05:11.434 --rc genhtml_legend=1 00:05:11.434 --rc geninfo_all_blocks=1 00:05:11.434 --rc geninfo_unexecuted_blocks=1 00:05:11.434 00:05:11.434 ' 00:05:11.434 23:38:59 rpc_client -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:05:11.434 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:11.434 --rc genhtml_branch_coverage=1 00:05:11.434 --rc genhtml_function_coverage=1 00:05:11.434 --rc genhtml_legend=1 00:05:11.434 --rc geninfo_all_blocks=1 00:05:11.434 --rc geninfo_unexecuted_blocks=1 00:05:11.434 00:05:11.434 ' 00:05:11.434 23:38:59 rpc_client -- rpc_client/rpc_client.sh@10 -- # /home/vagrant/spdk_repo/spdk/test/rpc_client/rpc_client_test 00:05:11.434 OK 00:05:11.434 23:38:59 rpc_client -- rpc_client/rpc_client.sh@12 -- # trap - SIGINT SIGTERM EXIT 00:05:11.434 00:05:11.434 real 0m0.302s 00:05:11.434 user 0m0.152s 00:05:11.434 sys 0m0.166s 00:05:11.434 23:38:59 rpc_client -- common/autotest_common.sh@1130 -- # xtrace_disable 00:05:11.434 23:38:59 rpc_client -- common/autotest_common.sh@10 -- # set +x 00:05:11.434 ************************************ 00:05:11.434 END TEST rpc_client 00:05:11.434 ************************************ 00:05:11.434 23:38:59 -- spdk/autotest.sh@159 -- # run_test json_config /home/vagrant/spdk_repo/spdk/test/json_config/json_config.sh 00:05:11.434 23:38:59 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:05:11.434 23:38:59 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:05:11.434 23:38:59 -- common/autotest_common.sh@10 -- # set +x 00:05:11.434 ************************************ 00:05:11.434 START TEST json_config 00:05:11.434 ************************************ 00:05:11.434 23:38:59 json_config -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/json_config/json_config.sh 00:05:11.434 23:38:59 json_config -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:05:11.434 23:38:59 json_config -- common/autotest_common.sh@1693 -- # lcov --version 00:05:11.434 23:38:59 json_config -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:05:11.694 23:38:59 json_config -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:05:11.694 23:38:59 json_config -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:05:11.694 23:38:59 json_config -- scripts/common.sh@333 -- # local ver1 ver1_l 00:05:11.694 23:38:59 json_config -- scripts/common.sh@334 -- # local ver2 ver2_l 00:05:11.694 23:38:59 json_config -- scripts/common.sh@336 -- # IFS=.-: 00:05:11.694 23:38:59 json_config -- scripts/common.sh@336 -- # read -ra ver1 00:05:11.694 23:38:59 json_config -- scripts/common.sh@337 -- # IFS=.-: 00:05:11.694 23:38:59 json_config -- scripts/common.sh@337 -- # read -ra ver2 00:05:11.694 23:38:59 json_config -- scripts/common.sh@338 -- # local 'op=<' 00:05:11.694 23:38:59 json_config -- scripts/common.sh@340 -- # ver1_l=2 00:05:11.694 23:38:59 json_config -- scripts/common.sh@341 -- # ver2_l=1 00:05:11.694 23:38:59 json_config -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:05:11.694 23:38:59 json_config -- scripts/common.sh@344 -- # case "$op" in 00:05:11.694 23:38:59 json_config -- scripts/common.sh@345 -- # : 1 00:05:11.694 23:38:59 json_config -- scripts/common.sh@364 -- # (( v = 0 )) 00:05:11.694 23:38:59 json_config -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:05:11.694 23:38:59 json_config -- scripts/common.sh@365 -- # decimal 1 00:05:11.694 23:38:59 json_config -- scripts/common.sh@353 -- # local d=1 00:05:11.694 23:38:59 json_config -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:05:11.694 23:38:59 json_config -- scripts/common.sh@355 -- # echo 1 00:05:11.694 23:38:59 json_config -- scripts/common.sh@365 -- # ver1[v]=1 00:05:11.694 23:38:59 json_config -- scripts/common.sh@366 -- # decimal 2 00:05:11.694 23:38:59 json_config -- scripts/common.sh@353 -- # local d=2 00:05:11.694 23:38:59 json_config -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:05:11.694 23:38:59 json_config -- scripts/common.sh@355 -- # echo 2 00:05:11.694 23:38:59 json_config -- scripts/common.sh@366 -- # ver2[v]=2 00:05:11.694 23:38:59 json_config -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:05:11.694 23:38:59 json_config -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:05:11.694 23:38:59 json_config -- scripts/common.sh@368 -- # return 0 00:05:11.694 23:38:59 json_config -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:05:11.694 23:38:59 json_config -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:05:11.694 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:11.694 --rc genhtml_branch_coverage=1 00:05:11.694 --rc genhtml_function_coverage=1 00:05:11.694 --rc genhtml_legend=1 00:05:11.694 --rc geninfo_all_blocks=1 00:05:11.694 --rc geninfo_unexecuted_blocks=1 00:05:11.694 00:05:11.694 ' 00:05:11.694 23:38:59 json_config -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:05:11.694 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:11.694 --rc genhtml_branch_coverage=1 00:05:11.694 --rc genhtml_function_coverage=1 00:05:11.694 --rc genhtml_legend=1 00:05:11.694 --rc geninfo_all_blocks=1 00:05:11.694 --rc geninfo_unexecuted_blocks=1 00:05:11.694 00:05:11.694 ' 00:05:11.694 23:38:59 json_config -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:05:11.694 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:11.694 --rc genhtml_branch_coverage=1 00:05:11.694 --rc genhtml_function_coverage=1 00:05:11.694 --rc genhtml_legend=1 00:05:11.694 --rc geninfo_all_blocks=1 00:05:11.694 --rc geninfo_unexecuted_blocks=1 00:05:11.694 00:05:11.694 ' 00:05:11.694 23:38:59 json_config -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:05:11.694 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:11.694 --rc genhtml_branch_coverage=1 00:05:11.694 --rc genhtml_function_coverage=1 00:05:11.694 --rc genhtml_legend=1 00:05:11.694 --rc geninfo_all_blocks=1 00:05:11.694 --rc geninfo_unexecuted_blocks=1 00:05:11.694 00:05:11.694 ' 00:05:11.694 23:38:59 json_config -- json_config/json_config.sh@8 -- # source /home/vagrant/spdk_repo/spdk/test/nvmf/common.sh 00:05:11.694 23:38:59 json_config -- nvmf/common.sh@7 -- # uname -s 00:05:11.694 23:38:59 json_config -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:05:11.694 23:38:59 json_config -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:05:11.694 23:38:59 json_config -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:05:11.694 23:38:59 json_config -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:05:11.694 23:38:59 json_config -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:05:11.694 23:38:59 json_config -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:05:11.694 23:38:59 json_config -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:05:11.694 23:38:59 json_config -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:05:11.694 23:38:59 json_config -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:05:11.694 23:38:59 json_config -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:05:11.694 23:38:59 json_config -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:0d30ab93-e078-437c-8ce5-2573a58e7d32 00:05:11.694 23:38:59 json_config -- nvmf/common.sh@18 -- # NVME_HOSTID=0d30ab93-e078-437c-8ce5-2573a58e7d32 00:05:11.694 23:38:59 json_config -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:05:11.694 23:38:59 json_config -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:05:11.694 23:38:59 json_config -- nvmf/common.sh@21 -- # NET_TYPE=phy-fallback 00:05:11.694 23:38:59 json_config -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:05:11.694 23:38:59 json_config -- nvmf/common.sh@49 -- # source /home/vagrant/spdk_repo/spdk/scripts/common.sh 00:05:11.694 23:38:59 json_config -- scripts/common.sh@15 -- # shopt -s extglob 00:05:11.694 23:38:59 json_config -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:05:11.694 23:38:59 json_config -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:05:11.695 23:38:59 json_config -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:05:11.695 23:38:59 json_config -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:05:11.695 23:38:59 json_config -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:05:11.695 23:38:59 json_config -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:05:11.695 23:38:59 json_config -- paths/export.sh@5 -- # export PATH 00:05:11.695 23:38:59 json_config -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:05:11.695 23:38:59 json_config -- nvmf/common.sh@51 -- # : 0 00:05:11.695 23:38:59 json_config -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:05:11.695 23:38:59 json_config -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:05:11.695 23:38:59 json_config -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:05:11.695 23:38:59 json_config -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:05:11.695 23:38:59 json_config -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:05:11.695 23:38:59 json_config -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:05:11.695 /home/vagrant/spdk_repo/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:05:11.695 23:38:59 json_config -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:05:11.695 23:38:59 json_config -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:05:11.695 23:38:59 json_config -- nvmf/common.sh@55 -- # have_pci_nics=0 00:05:11.695 23:38:59 json_config -- json_config/json_config.sh@9 -- # source /home/vagrant/spdk_repo/spdk/test/json_config/common.sh 00:05:11.695 23:38:59 json_config -- json_config/json_config.sh@11 -- # [[ 0 -eq 1 ]] 00:05:11.695 23:38:59 json_config -- json_config/json_config.sh@15 -- # [[ 0 -ne 1 ]] 00:05:11.695 23:38:59 json_config -- json_config/json_config.sh@15 -- # [[ 0 -eq 1 ]] 00:05:11.695 23:38:59 json_config -- json_config/json_config.sh@26 -- # (( SPDK_TEST_BLOCKDEV + SPDK_TEST_ISCSI + SPDK_TEST_NVMF + SPDK_TEST_VHOST + SPDK_TEST_VHOST_INIT + SPDK_TEST_RBD == 0 )) 00:05:11.695 23:38:59 json_config -- json_config/json_config.sh@27 -- # echo 'WARNING: No tests are enabled so not running JSON configuration tests' 00:05:11.695 WARNING: No tests are enabled so not running JSON configuration tests 00:05:11.695 23:38:59 json_config -- json_config/json_config.sh@28 -- # exit 0 00:05:11.695 00:05:11.695 real 0m0.236s 00:05:11.695 user 0m0.147s 00:05:11.695 sys 0m0.093s 00:05:11.695 23:38:59 json_config -- common/autotest_common.sh@1130 -- # xtrace_disable 00:05:11.695 23:38:59 json_config -- common/autotest_common.sh@10 -- # set +x 00:05:11.695 ************************************ 00:05:11.695 END TEST json_config 00:05:11.695 ************************************ 00:05:11.695 23:38:59 -- spdk/autotest.sh@160 -- # run_test json_config_extra_key /home/vagrant/spdk_repo/spdk/test/json_config/json_config_extra_key.sh 00:05:11.695 23:38:59 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:05:11.695 23:38:59 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:05:11.695 23:38:59 -- common/autotest_common.sh@10 -- # set +x 00:05:11.695 ************************************ 00:05:11.695 START TEST json_config_extra_key 00:05:11.695 ************************************ 00:05:11.695 23:38:59 json_config_extra_key -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/json_config/json_config_extra_key.sh 00:05:11.955 23:38:59 json_config_extra_key -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:05:11.955 23:38:59 json_config_extra_key -- common/autotest_common.sh@1693 -- # lcov --version 00:05:11.955 23:38:59 json_config_extra_key -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:05:11.955 23:38:59 json_config_extra_key -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:05:11.955 23:38:59 json_config_extra_key -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:05:11.955 23:38:59 json_config_extra_key -- scripts/common.sh@333 -- # local ver1 ver1_l 00:05:11.955 23:38:59 json_config_extra_key -- scripts/common.sh@334 -- # local ver2 ver2_l 00:05:11.955 23:38:59 json_config_extra_key -- scripts/common.sh@336 -- # IFS=.-: 00:05:11.955 23:38:59 json_config_extra_key -- scripts/common.sh@336 -- # read -ra ver1 00:05:11.955 23:38:59 json_config_extra_key -- scripts/common.sh@337 -- # IFS=.-: 00:05:11.955 23:38:59 json_config_extra_key -- scripts/common.sh@337 -- # read -ra ver2 00:05:11.955 23:38:59 json_config_extra_key -- scripts/common.sh@338 -- # local 'op=<' 00:05:11.956 23:38:59 json_config_extra_key -- scripts/common.sh@340 -- # ver1_l=2 00:05:11.956 23:38:59 json_config_extra_key -- scripts/common.sh@341 -- # ver2_l=1 00:05:11.956 23:38:59 json_config_extra_key -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:05:11.956 23:38:59 json_config_extra_key -- scripts/common.sh@344 -- # case "$op" in 00:05:11.956 23:38:59 json_config_extra_key -- scripts/common.sh@345 -- # : 1 00:05:11.956 23:38:59 json_config_extra_key -- scripts/common.sh@364 -- # (( v = 0 )) 00:05:11.956 23:38:59 json_config_extra_key -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:05:11.956 23:38:59 json_config_extra_key -- scripts/common.sh@365 -- # decimal 1 00:05:11.956 23:38:59 json_config_extra_key -- scripts/common.sh@353 -- # local d=1 00:05:11.956 23:38:59 json_config_extra_key -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:05:11.956 23:38:59 json_config_extra_key -- scripts/common.sh@355 -- # echo 1 00:05:11.956 23:38:59 json_config_extra_key -- scripts/common.sh@365 -- # ver1[v]=1 00:05:11.956 23:38:59 json_config_extra_key -- scripts/common.sh@366 -- # decimal 2 00:05:11.956 23:38:59 json_config_extra_key -- scripts/common.sh@353 -- # local d=2 00:05:11.956 23:38:59 json_config_extra_key -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:05:11.956 23:38:59 json_config_extra_key -- scripts/common.sh@355 -- # echo 2 00:05:11.956 23:38:59 json_config_extra_key -- scripts/common.sh@366 -- # ver2[v]=2 00:05:11.956 23:38:59 json_config_extra_key -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:05:11.956 23:38:59 json_config_extra_key -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:05:11.956 23:38:59 json_config_extra_key -- scripts/common.sh@368 -- # return 0 00:05:11.956 23:38:59 json_config_extra_key -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:05:11.956 23:38:59 json_config_extra_key -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:05:11.956 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:11.956 --rc genhtml_branch_coverage=1 00:05:11.956 --rc genhtml_function_coverage=1 00:05:11.956 --rc genhtml_legend=1 00:05:11.956 --rc geninfo_all_blocks=1 00:05:11.956 --rc geninfo_unexecuted_blocks=1 00:05:11.956 00:05:11.956 ' 00:05:11.956 23:38:59 json_config_extra_key -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:05:11.956 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:11.956 --rc genhtml_branch_coverage=1 00:05:11.956 --rc genhtml_function_coverage=1 00:05:11.956 --rc genhtml_legend=1 00:05:11.956 --rc geninfo_all_blocks=1 00:05:11.956 --rc geninfo_unexecuted_blocks=1 00:05:11.956 00:05:11.956 ' 00:05:11.956 23:38:59 json_config_extra_key -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:05:11.956 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:11.956 --rc genhtml_branch_coverage=1 00:05:11.956 --rc genhtml_function_coverage=1 00:05:11.956 --rc genhtml_legend=1 00:05:11.956 --rc geninfo_all_blocks=1 00:05:11.956 --rc geninfo_unexecuted_blocks=1 00:05:11.956 00:05:11.956 ' 00:05:11.956 23:38:59 json_config_extra_key -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:05:11.956 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:11.956 --rc genhtml_branch_coverage=1 00:05:11.956 --rc genhtml_function_coverage=1 00:05:11.956 --rc genhtml_legend=1 00:05:11.956 --rc geninfo_all_blocks=1 00:05:11.956 --rc geninfo_unexecuted_blocks=1 00:05:11.956 00:05:11.956 ' 00:05:11.956 23:38:59 json_config_extra_key -- json_config/json_config_extra_key.sh@9 -- # source /home/vagrant/spdk_repo/spdk/test/nvmf/common.sh 00:05:11.956 23:38:59 json_config_extra_key -- nvmf/common.sh@7 -- # uname -s 00:05:11.956 23:38:59 json_config_extra_key -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:05:11.956 23:38:59 json_config_extra_key -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:05:11.956 23:38:59 json_config_extra_key -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:05:11.956 23:38:59 json_config_extra_key -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:05:11.956 23:38:59 json_config_extra_key -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:05:11.956 23:38:59 json_config_extra_key -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:05:11.956 23:38:59 json_config_extra_key -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:05:11.956 23:38:59 json_config_extra_key -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:05:11.956 23:38:59 json_config_extra_key -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:05:11.956 23:38:59 json_config_extra_key -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:05:11.956 23:38:59 json_config_extra_key -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:0d30ab93-e078-437c-8ce5-2573a58e7d32 00:05:11.956 23:38:59 json_config_extra_key -- nvmf/common.sh@18 -- # NVME_HOSTID=0d30ab93-e078-437c-8ce5-2573a58e7d32 00:05:11.956 23:38:59 json_config_extra_key -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:05:11.956 23:38:59 json_config_extra_key -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:05:11.956 23:38:59 json_config_extra_key -- nvmf/common.sh@21 -- # NET_TYPE=phy-fallback 00:05:11.956 23:38:59 json_config_extra_key -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:05:11.956 23:38:59 json_config_extra_key -- nvmf/common.sh@49 -- # source /home/vagrant/spdk_repo/spdk/scripts/common.sh 00:05:11.956 23:38:59 json_config_extra_key -- scripts/common.sh@15 -- # shopt -s extglob 00:05:11.956 23:38:59 json_config_extra_key -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:05:11.956 23:38:59 json_config_extra_key -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:05:11.956 23:38:59 json_config_extra_key -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:05:11.956 23:38:59 json_config_extra_key -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:05:11.956 23:38:59 json_config_extra_key -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:05:11.956 23:38:59 json_config_extra_key -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:05:11.956 23:38:59 json_config_extra_key -- paths/export.sh@5 -- # export PATH 00:05:11.956 23:38:59 json_config_extra_key -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:05:11.956 23:38:59 json_config_extra_key -- nvmf/common.sh@51 -- # : 0 00:05:11.956 23:38:59 json_config_extra_key -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:05:11.956 23:38:59 json_config_extra_key -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:05:11.956 23:38:59 json_config_extra_key -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:05:11.956 23:38:59 json_config_extra_key -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:05:11.956 23:38:59 json_config_extra_key -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:05:11.956 23:38:59 json_config_extra_key -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:05:11.956 /home/vagrant/spdk_repo/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:05:11.956 23:38:59 json_config_extra_key -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:05:11.956 23:38:59 json_config_extra_key -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:05:11.956 23:38:59 json_config_extra_key -- nvmf/common.sh@55 -- # have_pci_nics=0 00:05:11.956 23:38:59 json_config_extra_key -- json_config/json_config_extra_key.sh@10 -- # source /home/vagrant/spdk_repo/spdk/test/json_config/common.sh 00:05:11.956 23:38:59 json_config_extra_key -- json_config/json_config_extra_key.sh@17 -- # app_pid=(['target']='') 00:05:11.956 23:38:59 json_config_extra_key -- json_config/json_config_extra_key.sh@17 -- # declare -A app_pid 00:05:11.956 23:38:59 json_config_extra_key -- json_config/json_config_extra_key.sh@18 -- # app_socket=(['target']='/var/tmp/spdk_tgt.sock') 00:05:11.956 23:38:59 json_config_extra_key -- json_config/json_config_extra_key.sh@18 -- # declare -A app_socket 00:05:11.957 23:38:59 json_config_extra_key -- json_config/json_config_extra_key.sh@19 -- # app_params=(['target']='-m 0x1 -s 1024') 00:05:11.957 23:38:59 json_config_extra_key -- json_config/json_config_extra_key.sh@19 -- # declare -A app_params 00:05:11.957 23:38:59 json_config_extra_key -- json_config/json_config_extra_key.sh@20 -- # configs_path=(['target']='/home/vagrant/spdk_repo/spdk/test/json_config/extra_key.json') 00:05:11.957 23:38:59 json_config_extra_key -- json_config/json_config_extra_key.sh@20 -- # declare -A configs_path 00:05:11.957 23:38:59 json_config_extra_key -- json_config/json_config_extra_key.sh@22 -- # trap 'on_error_exit "${FUNCNAME}" "${LINENO}"' ERR 00:05:11.957 23:38:59 json_config_extra_key -- json_config/json_config_extra_key.sh@24 -- # echo 'INFO: launching applications...' 00:05:11.957 INFO: launching applications... 00:05:11.957 23:38:59 json_config_extra_key -- json_config/json_config_extra_key.sh@25 -- # json_config_test_start_app target --json /home/vagrant/spdk_repo/spdk/test/json_config/extra_key.json 00:05:11.957 23:38:59 json_config_extra_key -- json_config/common.sh@9 -- # local app=target 00:05:11.957 23:38:59 json_config_extra_key -- json_config/common.sh@10 -- # shift 00:05:11.957 23:38:59 json_config_extra_key -- json_config/common.sh@12 -- # [[ -n 22 ]] 00:05:11.957 23:38:59 json_config_extra_key -- json_config/common.sh@13 -- # [[ -z '' ]] 00:05:11.957 23:38:59 json_config_extra_key -- json_config/common.sh@15 -- # local app_extra_params= 00:05:11.957 23:38:59 json_config_extra_key -- json_config/common.sh@16 -- # [[ 0 -eq 1 ]] 00:05:11.957 23:38:59 json_config_extra_key -- json_config/common.sh@16 -- # [[ 0 -eq 1 ]] 00:05:11.957 23:38:59 json_config_extra_key -- json_config/common.sh@22 -- # app_pid["$app"]=69489 00:05:11.957 23:38:59 json_config_extra_key -- json_config/common.sh@24 -- # echo 'Waiting for target to run...' 00:05:11.957 Waiting for target to run... 00:05:11.957 23:38:59 json_config_extra_key -- json_config/common.sh@25 -- # waitforlisten 69489 /var/tmp/spdk_tgt.sock 00:05:11.957 23:38:59 json_config_extra_key -- json_config/common.sh@21 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 -s 1024 -r /var/tmp/spdk_tgt.sock --json /home/vagrant/spdk_repo/spdk/test/json_config/extra_key.json 00:05:11.957 23:38:59 json_config_extra_key -- common/autotest_common.sh@835 -- # '[' -z 69489 ']' 00:05:11.957 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk_tgt.sock... 00:05:11.957 23:38:59 json_config_extra_key -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk_tgt.sock 00:05:11.957 23:38:59 json_config_extra_key -- common/autotest_common.sh@840 -- # local max_retries=100 00:05:11.957 23:38:59 json_config_extra_key -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk_tgt.sock...' 00:05:11.957 23:38:59 json_config_extra_key -- common/autotest_common.sh@844 -- # xtrace_disable 00:05:11.957 23:38:59 json_config_extra_key -- common/autotest_common.sh@10 -- # set +x 00:05:12.243 [2024-11-26 23:39:00.087634] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:05:12.243 [2024-11-26 23:39:00.087778] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 -m 1024 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid69489 ] 00:05:12.543 [2024-11-26 23:39:00.453934] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:05:12.543 [2024-11-26 23:39:00.472686] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:05:13.112 00:05:13.112 INFO: shutting down applications... 00:05:13.112 23:39:00 json_config_extra_key -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:05:13.112 23:39:00 json_config_extra_key -- common/autotest_common.sh@868 -- # return 0 00:05:13.112 23:39:00 json_config_extra_key -- json_config/common.sh@26 -- # echo '' 00:05:13.112 23:39:00 json_config_extra_key -- json_config/json_config_extra_key.sh@27 -- # echo 'INFO: shutting down applications...' 00:05:13.112 23:39:00 json_config_extra_key -- json_config/json_config_extra_key.sh@28 -- # json_config_test_shutdown_app target 00:05:13.112 23:39:00 json_config_extra_key -- json_config/common.sh@31 -- # local app=target 00:05:13.112 23:39:00 json_config_extra_key -- json_config/common.sh@34 -- # [[ -n 22 ]] 00:05:13.112 23:39:00 json_config_extra_key -- json_config/common.sh@35 -- # [[ -n 69489 ]] 00:05:13.112 23:39:00 json_config_extra_key -- json_config/common.sh@38 -- # kill -SIGINT 69489 00:05:13.112 23:39:00 json_config_extra_key -- json_config/common.sh@40 -- # (( i = 0 )) 00:05:13.112 23:39:00 json_config_extra_key -- json_config/common.sh@40 -- # (( i < 30 )) 00:05:13.112 23:39:00 json_config_extra_key -- json_config/common.sh@41 -- # kill -0 69489 00:05:13.112 23:39:00 json_config_extra_key -- json_config/common.sh@45 -- # sleep 0.5 00:05:13.371 23:39:01 json_config_extra_key -- json_config/common.sh@40 -- # (( i++ )) 00:05:13.371 23:39:01 json_config_extra_key -- json_config/common.sh@40 -- # (( i < 30 )) 00:05:13.371 23:39:01 json_config_extra_key -- json_config/common.sh@41 -- # kill -0 69489 00:05:13.371 23:39:01 json_config_extra_key -- json_config/common.sh@42 -- # app_pid["$app"]= 00:05:13.371 23:39:01 json_config_extra_key -- json_config/common.sh@43 -- # break 00:05:13.371 23:39:01 json_config_extra_key -- json_config/common.sh@48 -- # [[ -n '' ]] 00:05:13.371 23:39:01 json_config_extra_key -- json_config/common.sh@53 -- # echo 'SPDK target shutdown done' 00:05:13.371 SPDK target shutdown done 00:05:13.371 23:39:01 json_config_extra_key -- json_config/json_config_extra_key.sh@30 -- # echo Success 00:05:13.371 Success 00:05:13.371 00:05:13.371 real 0m1.725s 00:05:13.371 user 0m1.491s 00:05:13.371 sys 0m0.485s 00:05:13.371 23:39:01 json_config_extra_key -- common/autotest_common.sh@1130 -- # xtrace_disable 00:05:13.371 23:39:01 json_config_extra_key -- common/autotest_common.sh@10 -- # set +x 00:05:13.371 ************************************ 00:05:13.371 END TEST json_config_extra_key 00:05:13.371 ************************************ 00:05:13.631 23:39:01 -- spdk/autotest.sh@161 -- # run_test alias_rpc /home/vagrant/spdk_repo/spdk/test/json_config/alias_rpc/alias_rpc.sh 00:05:13.631 23:39:01 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:05:13.631 23:39:01 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:05:13.631 23:39:01 -- common/autotest_common.sh@10 -- # set +x 00:05:13.631 ************************************ 00:05:13.631 START TEST alias_rpc 00:05:13.631 ************************************ 00:05:13.631 23:39:01 alias_rpc -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/json_config/alias_rpc/alias_rpc.sh 00:05:13.631 * Looking for test storage... 00:05:13.631 * Found test storage at /home/vagrant/spdk_repo/spdk/test/json_config/alias_rpc 00:05:13.631 23:39:01 alias_rpc -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:05:13.631 23:39:01 alias_rpc -- common/autotest_common.sh@1693 -- # lcov --version 00:05:13.631 23:39:01 alias_rpc -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:05:13.631 23:39:01 alias_rpc -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:05:13.631 23:39:01 alias_rpc -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:05:13.631 23:39:01 alias_rpc -- scripts/common.sh@333 -- # local ver1 ver1_l 00:05:13.631 23:39:01 alias_rpc -- scripts/common.sh@334 -- # local ver2 ver2_l 00:05:13.631 23:39:01 alias_rpc -- scripts/common.sh@336 -- # IFS=.-: 00:05:13.631 23:39:01 alias_rpc -- scripts/common.sh@336 -- # read -ra ver1 00:05:13.631 23:39:01 alias_rpc -- scripts/common.sh@337 -- # IFS=.-: 00:05:13.631 23:39:01 alias_rpc -- scripts/common.sh@337 -- # read -ra ver2 00:05:13.631 23:39:01 alias_rpc -- scripts/common.sh@338 -- # local 'op=<' 00:05:13.631 23:39:01 alias_rpc -- scripts/common.sh@340 -- # ver1_l=2 00:05:13.631 23:39:01 alias_rpc -- scripts/common.sh@341 -- # ver2_l=1 00:05:13.631 23:39:01 alias_rpc -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:05:13.631 23:39:01 alias_rpc -- scripts/common.sh@344 -- # case "$op" in 00:05:13.631 23:39:01 alias_rpc -- scripts/common.sh@345 -- # : 1 00:05:13.631 23:39:01 alias_rpc -- scripts/common.sh@364 -- # (( v = 0 )) 00:05:13.631 23:39:01 alias_rpc -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:05:13.631 23:39:01 alias_rpc -- scripts/common.sh@365 -- # decimal 1 00:05:13.891 23:39:01 alias_rpc -- scripts/common.sh@353 -- # local d=1 00:05:13.891 23:39:01 alias_rpc -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:05:13.891 23:39:01 alias_rpc -- scripts/common.sh@355 -- # echo 1 00:05:13.891 23:39:01 alias_rpc -- scripts/common.sh@365 -- # ver1[v]=1 00:05:13.891 23:39:01 alias_rpc -- scripts/common.sh@366 -- # decimal 2 00:05:13.891 23:39:01 alias_rpc -- scripts/common.sh@353 -- # local d=2 00:05:13.891 23:39:01 alias_rpc -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:05:13.891 23:39:01 alias_rpc -- scripts/common.sh@355 -- # echo 2 00:05:13.891 23:39:01 alias_rpc -- scripts/common.sh@366 -- # ver2[v]=2 00:05:13.891 23:39:01 alias_rpc -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:05:13.891 23:39:01 alias_rpc -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:05:13.891 23:39:01 alias_rpc -- scripts/common.sh@368 -- # return 0 00:05:13.891 23:39:01 alias_rpc -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:05:13.891 23:39:01 alias_rpc -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:05:13.891 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:13.891 --rc genhtml_branch_coverage=1 00:05:13.891 --rc genhtml_function_coverage=1 00:05:13.891 --rc genhtml_legend=1 00:05:13.891 --rc geninfo_all_blocks=1 00:05:13.891 --rc geninfo_unexecuted_blocks=1 00:05:13.891 00:05:13.891 ' 00:05:13.891 23:39:01 alias_rpc -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:05:13.891 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:13.891 --rc genhtml_branch_coverage=1 00:05:13.891 --rc genhtml_function_coverage=1 00:05:13.891 --rc genhtml_legend=1 00:05:13.891 --rc geninfo_all_blocks=1 00:05:13.891 --rc geninfo_unexecuted_blocks=1 00:05:13.891 00:05:13.891 ' 00:05:13.891 23:39:01 alias_rpc -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:05:13.891 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:13.891 --rc genhtml_branch_coverage=1 00:05:13.891 --rc genhtml_function_coverage=1 00:05:13.891 --rc genhtml_legend=1 00:05:13.891 --rc geninfo_all_blocks=1 00:05:13.891 --rc geninfo_unexecuted_blocks=1 00:05:13.891 00:05:13.891 ' 00:05:13.891 23:39:01 alias_rpc -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:05:13.891 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:13.891 --rc genhtml_branch_coverage=1 00:05:13.891 --rc genhtml_function_coverage=1 00:05:13.891 --rc genhtml_legend=1 00:05:13.891 --rc geninfo_all_blocks=1 00:05:13.891 --rc geninfo_unexecuted_blocks=1 00:05:13.891 00:05:13.891 ' 00:05:13.891 23:39:01 alias_rpc -- alias_rpc/alias_rpc.sh@10 -- # trap 'killprocess $spdk_tgt_pid; exit 1' ERR 00:05:13.891 23:39:01 alias_rpc -- alias_rpc/alias_rpc.sh@12 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:05:13.891 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:05:13.891 23:39:01 alias_rpc -- alias_rpc/alias_rpc.sh@13 -- # spdk_tgt_pid=69564 00:05:13.891 23:39:01 alias_rpc -- alias_rpc/alias_rpc.sh@14 -- # waitforlisten 69564 00:05:13.892 23:39:01 alias_rpc -- common/autotest_common.sh@835 -- # '[' -z 69564 ']' 00:05:13.892 23:39:01 alias_rpc -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:05:13.892 23:39:01 alias_rpc -- common/autotest_common.sh@840 -- # local max_retries=100 00:05:13.892 23:39:01 alias_rpc -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:05:13.892 23:39:01 alias_rpc -- common/autotest_common.sh@844 -- # xtrace_disable 00:05:13.892 23:39:01 alias_rpc -- common/autotest_common.sh@10 -- # set +x 00:05:13.892 [2024-11-26 23:39:01.867000] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:05:13.892 [2024-11-26 23:39:01.867232] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid69564 ] 00:05:14.155 [2024-11-26 23:39:02.023848] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:05:14.155 [2024-11-26 23:39:02.052823] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:05:14.725 23:39:02 alias_rpc -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:05:14.725 23:39:02 alias_rpc -- common/autotest_common.sh@868 -- # return 0 00:05:14.725 23:39:02 alias_rpc -- alias_rpc/alias_rpc.sh@17 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py load_config -i 00:05:14.984 23:39:03 alias_rpc -- alias_rpc/alias_rpc.sh@19 -- # killprocess 69564 00:05:14.984 23:39:03 alias_rpc -- common/autotest_common.sh@954 -- # '[' -z 69564 ']' 00:05:14.984 23:39:03 alias_rpc -- common/autotest_common.sh@958 -- # kill -0 69564 00:05:14.984 23:39:03 alias_rpc -- common/autotest_common.sh@959 -- # uname 00:05:14.984 23:39:03 alias_rpc -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:05:14.984 23:39:03 alias_rpc -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 69564 00:05:14.984 23:39:03 alias_rpc -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:05:14.984 23:39:03 alias_rpc -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:05:14.984 23:39:03 alias_rpc -- common/autotest_common.sh@972 -- # echo 'killing process with pid 69564' 00:05:14.984 killing process with pid 69564 00:05:14.984 23:39:03 alias_rpc -- common/autotest_common.sh@973 -- # kill 69564 00:05:14.984 23:39:03 alias_rpc -- common/autotest_common.sh@978 -- # wait 69564 00:05:15.554 00:05:15.554 real 0m1.868s 00:05:15.554 user 0m2.002s 00:05:15.554 sys 0m0.505s 00:05:15.554 23:39:03 alias_rpc -- common/autotest_common.sh@1130 -- # xtrace_disable 00:05:15.554 23:39:03 alias_rpc -- common/autotest_common.sh@10 -- # set +x 00:05:15.554 ************************************ 00:05:15.554 END TEST alias_rpc 00:05:15.554 ************************************ 00:05:15.554 23:39:03 -- spdk/autotest.sh@163 -- # [[ 0 -eq 0 ]] 00:05:15.554 23:39:03 -- spdk/autotest.sh@164 -- # run_test spdkcli_tcp /home/vagrant/spdk_repo/spdk/test/spdkcli/tcp.sh 00:05:15.554 23:39:03 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:05:15.554 23:39:03 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:05:15.554 23:39:03 -- common/autotest_common.sh@10 -- # set +x 00:05:15.554 ************************************ 00:05:15.554 START TEST spdkcli_tcp 00:05:15.554 ************************************ 00:05:15.554 23:39:03 spdkcli_tcp -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/spdkcli/tcp.sh 00:05:15.554 * Looking for test storage... 00:05:15.554 * Found test storage at /home/vagrant/spdk_repo/spdk/test/spdkcli 00:05:15.554 23:39:03 spdkcli_tcp -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:05:15.554 23:39:03 spdkcli_tcp -- common/autotest_common.sh@1693 -- # lcov --version 00:05:15.554 23:39:03 spdkcli_tcp -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:05:15.814 23:39:03 spdkcli_tcp -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:05:15.814 23:39:03 spdkcli_tcp -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:05:15.814 23:39:03 spdkcli_tcp -- scripts/common.sh@333 -- # local ver1 ver1_l 00:05:15.814 23:39:03 spdkcli_tcp -- scripts/common.sh@334 -- # local ver2 ver2_l 00:05:15.814 23:39:03 spdkcli_tcp -- scripts/common.sh@336 -- # IFS=.-: 00:05:15.814 23:39:03 spdkcli_tcp -- scripts/common.sh@336 -- # read -ra ver1 00:05:15.814 23:39:03 spdkcli_tcp -- scripts/common.sh@337 -- # IFS=.-: 00:05:15.814 23:39:03 spdkcli_tcp -- scripts/common.sh@337 -- # read -ra ver2 00:05:15.814 23:39:03 spdkcli_tcp -- scripts/common.sh@338 -- # local 'op=<' 00:05:15.814 23:39:03 spdkcli_tcp -- scripts/common.sh@340 -- # ver1_l=2 00:05:15.814 23:39:03 spdkcli_tcp -- scripts/common.sh@341 -- # ver2_l=1 00:05:15.814 23:39:03 spdkcli_tcp -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:05:15.814 23:39:03 spdkcli_tcp -- scripts/common.sh@344 -- # case "$op" in 00:05:15.814 23:39:03 spdkcli_tcp -- scripts/common.sh@345 -- # : 1 00:05:15.814 23:39:03 spdkcli_tcp -- scripts/common.sh@364 -- # (( v = 0 )) 00:05:15.814 23:39:03 spdkcli_tcp -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:05:15.814 23:39:03 spdkcli_tcp -- scripts/common.sh@365 -- # decimal 1 00:05:15.814 23:39:03 spdkcli_tcp -- scripts/common.sh@353 -- # local d=1 00:05:15.814 23:39:03 spdkcli_tcp -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:05:15.814 23:39:03 spdkcli_tcp -- scripts/common.sh@355 -- # echo 1 00:05:15.814 23:39:03 spdkcli_tcp -- scripts/common.sh@365 -- # ver1[v]=1 00:05:15.814 23:39:03 spdkcli_tcp -- scripts/common.sh@366 -- # decimal 2 00:05:15.814 23:39:03 spdkcli_tcp -- scripts/common.sh@353 -- # local d=2 00:05:15.814 23:39:03 spdkcli_tcp -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:05:15.814 23:39:03 spdkcli_tcp -- scripts/common.sh@355 -- # echo 2 00:05:15.814 23:39:03 spdkcli_tcp -- scripts/common.sh@366 -- # ver2[v]=2 00:05:15.814 23:39:03 spdkcli_tcp -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:05:15.814 23:39:03 spdkcli_tcp -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:05:15.814 23:39:03 spdkcli_tcp -- scripts/common.sh@368 -- # return 0 00:05:15.814 23:39:03 spdkcli_tcp -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:05:15.814 23:39:03 spdkcli_tcp -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:05:15.814 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:15.814 --rc genhtml_branch_coverage=1 00:05:15.814 --rc genhtml_function_coverage=1 00:05:15.814 --rc genhtml_legend=1 00:05:15.814 --rc geninfo_all_blocks=1 00:05:15.814 --rc geninfo_unexecuted_blocks=1 00:05:15.814 00:05:15.814 ' 00:05:15.814 23:39:03 spdkcli_tcp -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:05:15.814 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:15.814 --rc genhtml_branch_coverage=1 00:05:15.814 --rc genhtml_function_coverage=1 00:05:15.814 --rc genhtml_legend=1 00:05:15.814 --rc geninfo_all_blocks=1 00:05:15.814 --rc geninfo_unexecuted_blocks=1 00:05:15.814 00:05:15.814 ' 00:05:15.814 23:39:03 spdkcli_tcp -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:05:15.814 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:15.814 --rc genhtml_branch_coverage=1 00:05:15.814 --rc genhtml_function_coverage=1 00:05:15.814 --rc genhtml_legend=1 00:05:15.814 --rc geninfo_all_blocks=1 00:05:15.814 --rc geninfo_unexecuted_blocks=1 00:05:15.814 00:05:15.814 ' 00:05:15.814 23:39:03 spdkcli_tcp -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:05:15.814 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:15.814 --rc genhtml_branch_coverage=1 00:05:15.814 --rc genhtml_function_coverage=1 00:05:15.814 --rc genhtml_legend=1 00:05:15.814 --rc geninfo_all_blocks=1 00:05:15.814 --rc geninfo_unexecuted_blocks=1 00:05:15.814 00:05:15.814 ' 00:05:15.814 23:39:03 spdkcli_tcp -- spdkcli/tcp.sh@9 -- # source /home/vagrant/spdk_repo/spdk/test/spdkcli/common.sh 00:05:15.814 23:39:03 spdkcli_tcp -- spdkcli/common.sh@6 -- # spdkcli_job=/home/vagrant/spdk_repo/spdk/test/spdkcli/spdkcli_job.py 00:05:15.814 23:39:03 spdkcli_tcp -- spdkcli/common.sh@7 -- # spdk_clear_config_py=/home/vagrant/spdk_repo/spdk/test/json_config/clear_config.py 00:05:15.814 23:39:03 spdkcli_tcp -- spdkcli/tcp.sh@18 -- # IP_ADDRESS=127.0.0.1 00:05:15.814 23:39:03 spdkcli_tcp -- spdkcli/tcp.sh@19 -- # PORT=9998 00:05:15.814 23:39:03 spdkcli_tcp -- spdkcli/tcp.sh@21 -- # trap 'err_cleanup; exit 1' SIGINT SIGTERM EXIT 00:05:15.814 23:39:03 spdkcli_tcp -- spdkcli/tcp.sh@23 -- # timing_enter run_spdk_tgt_tcp 00:05:15.814 23:39:03 spdkcli_tcp -- common/autotest_common.sh@726 -- # xtrace_disable 00:05:15.814 23:39:03 spdkcli_tcp -- common/autotest_common.sh@10 -- # set +x 00:05:15.814 23:39:03 spdkcli_tcp -- spdkcli/tcp.sh@25 -- # spdk_tgt_pid=69642 00:05:15.814 23:39:03 spdkcli_tcp -- spdkcli/tcp.sh@24 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x3 -p 0 00:05:15.814 23:39:03 spdkcli_tcp -- spdkcli/tcp.sh@27 -- # waitforlisten 69642 00:05:15.814 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:05:15.814 23:39:03 spdkcli_tcp -- common/autotest_common.sh@835 -- # '[' -z 69642 ']' 00:05:15.814 23:39:03 spdkcli_tcp -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:05:15.814 23:39:03 spdkcli_tcp -- common/autotest_common.sh@840 -- # local max_retries=100 00:05:15.815 23:39:03 spdkcli_tcp -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:05:15.815 23:39:03 spdkcli_tcp -- common/autotest_common.sh@844 -- # xtrace_disable 00:05:15.815 23:39:03 spdkcli_tcp -- common/autotest_common.sh@10 -- # set +x 00:05:15.815 [2024-11-26 23:39:03.819416] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:05:15.815 [2024-11-26 23:39:03.820047] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x3 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid69642 ] 00:05:16.074 [2024-11-26 23:39:03.976446] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 2 00:05:16.074 [2024-11-26 23:39:04.006372] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:05:16.074 [2024-11-26 23:39:04.006446] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:05:16.642 23:39:04 spdkcli_tcp -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:05:16.642 23:39:04 spdkcli_tcp -- common/autotest_common.sh@868 -- # return 0 00:05:16.642 23:39:04 spdkcli_tcp -- spdkcli/tcp.sh@31 -- # socat_pid=69659 00:05:16.642 23:39:04 spdkcli_tcp -- spdkcli/tcp.sh@33 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -r 100 -t 2 -s 127.0.0.1 -p 9998 rpc_get_methods 00:05:16.642 23:39:04 spdkcli_tcp -- spdkcli/tcp.sh@30 -- # socat TCP-LISTEN:9998 UNIX-CONNECT:/var/tmp/spdk.sock 00:05:16.903 [ 00:05:16.903 "bdev_malloc_delete", 00:05:16.903 "bdev_malloc_create", 00:05:16.903 "bdev_null_resize", 00:05:16.903 "bdev_null_delete", 00:05:16.903 "bdev_null_create", 00:05:16.903 "bdev_nvme_cuse_unregister", 00:05:16.903 "bdev_nvme_cuse_register", 00:05:16.903 "bdev_opal_new_user", 00:05:16.903 "bdev_opal_set_lock_state", 00:05:16.903 "bdev_opal_delete", 00:05:16.903 "bdev_opal_get_info", 00:05:16.903 "bdev_opal_create", 00:05:16.903 "bdev_nvme_opal_revert", 00:05:16.903 "bdev_nvme_opal_init", 00:05:16.903 "bdev_nvme_send_cmd", 00:05:16.903 "bdev_nvme_set_keys", 00:05:16.903 "bdev_nvme_get_path_iostat", 00:05:16.903 "bdev_nvme_get_mdns_discovery_info", 00:05:16.903 "bdev_nvme_stop_mdns_discovery", 00:05:16.903 "bdev_nvme_start_mdns_discovery", 00:05:16.903 "bdev_nvme_set_multipath_policy", 00:05:16.903 "bdev_nvme_set_preferred_path", 00:05:16.903 "bdev_nvme_get_io_paths", 00:05:16.903 "bdev_nvme_remove_error_injection", 00:05:16.903 "bdev_nvme_add_error_injection", 00:05:16.903 "bdev_nvme_get_discovery_info", 00:05:16.903 "bdev_nvme_stop_discovery", 00:05:16.903 "bdev_nvme_start_discovery", 00:05:16.903 "bdev_nvme_get_controller_health_info", 00:05:16.903 "bdev_nvme_disable_controller", 00:05:16.903 "bdev_nvme_enable_controller", 00:05:16.903 "bdev_nvme_reset_controller", 00:05:16.903 "bdev_nvme_get_transport_statistics", 00:05:16.903 "bdev_nvme_apply_firmware", 00:05:16.903 "bdev_nvme_detach_controller", 00:05:16.903 "bdev_nvme_get_controllers", 00:05:16.903 "bdev_nvme_attach_controller", 00:05:16.903 "bdev_nvme_set_hotplug", 00:05:16.903 "bdev_nvme_set_options", 00:05:16.903 "bdev_passthru_delete", 00:05:16.903 "bdev_passthru_create", 00:05:16.903 "bdev_lvol_set_parent_bdev", 00:05:16.903 "bdev_lvol_set_parent", 00:05:16.903 "bdev_lvol_check_shallow_copy", 00:05:16.903 "bdev_lvol_start_shallow_copy", 00:05:16.903 "bdev_lvol_grow_lvstore", 00:05:16.903 "bdev_lvol_get_lvols", 00:05:16.903 "bdev_lvol_get_lvstores", 00:05:16.903 "bdev_lvol_delete", 00:05:16.903 "bdev_lvol_set_read_only", 00:05:16.903 "bdev_lvol_resize", 00:05:16.903 "bdev_lvol_decouple_parent", 00:05:16.903 "bdev_lvol_inflate", 00:05:16.903 "bdev_lvol_rename", 00:05:16.903 "bdev_lvol_clone_bdev", 00:05:16.903 "bdev_lvol_clone", 00:05:16.903 "bdev_lvol_snapshot", 00:05:16.903 "bdev_lvol_create", 00:05:16.903 "bdev_lvol_delete_lvstore", 00:05:16.903 "bdev_lvol_rename_lvstore", 00:05:16.903 "bdev_lvol_create_lvstore", 00:05:16.903 "bdev_raid_set_options", 00:05:16.903 "bdev_raid_remove_base_bdev", 00:05:16.903 "bdev_raid_add_base_bdev", 00:05:16.903 "bdev_raid_delete", 00:05:16.903 "bdev_raid_create", 00:05:16.903 "bdev_raid_get_bdevs", 00:05:16.903 "bdev_error_inject_error", 00:05:16.903 "bdev_error_delete", 00:05:16.903 "bdev_error_create", 00:05:16.903 "bdev_split_delete", 00:05:16.903 "bdev_split_create", 00:05:16.903 "bdev_delay_delete", 00:05:16.903 "bdev_delay_create", 00:05:16.903 "bdev_delay_update_latency", 00:05:16.903 "bdev_zone_block_delete", 00:05:16.903 "bdev_zone_block_create", 00:05:16.903 "blobfs_create", 00:05:16.903 "blobfs_detect", 00:05:16.903 "blobfs_set_cache_size", 00:05:16.903 "bdev_aio_delete", 00:05:16.903 "bdev_aio_rescan", 00:05:16.903 "bdev_aio_create", 00:05:16.903 "bdev_ftl_set_property", 00:05:16.903 "bdev_ftl_get_properties", 00:05:16.903 "bdev_ftl_get_stats", 00:05:16.903 "bdev_ftl_unmap", 00:05:16.903 "bdev_ftl_unload", 00:05:16.903 "bdev_ftl_delete", 00:05:16.903 "bdev_ftl_load", 00:05:16.903 "bdev_ftl_create", 00:05:16.903 "bdev_virtio_attach_controller", 00:05:16.903 "bdev_virtio_scsi_get_devices", 00:05:16.903 "bdev_virtio_detach_controller", 00:05:16.903 "bdev_virtio_blk_set_hotplug", 00:05:16.903 "bdev_iscsi_delete", 00:05:16.903 "bdev_iscsi_create", 00:05:16.903 "bdev_iscsi_set_options", 00:05:16.903 "accel_error_inject_error", 00:05:16.903 "ioat_scan_accel_module", 00:05:16.903 "dsa_scan_accel_module", 00:05:16.903 "iaa_scan_accel_module", 00:05:16.903 "keyring_file_remove_key", 00:05:16.903 "keyring_file_add_key", 00:05:16.903 "keyring_linux_set_options", 00:05:16.903 "fsdev_aio_delete", 00:05:16.903 "fsdev_aio_create", 00:05:16.903 "iscsi_get_histogram", 00:05:16.903 "iscsi_enable_histogram", 00:05:16.903 "iscsi_set_options", 00:05:16.903 "iscsi_get_auth_groups", 00:05:16.903 "iscsi_auth_group_remove_secret", 00:05:16.903 "iscsi_auth_group_add_secret", 00:05:16.903 "iscsi_delete_auth_group", 00:05:16.904 "iscsi_create_auth_group", 00:05:16.904 "iscsi_set_discovery_auth", 00:05:16.904 "iscsi_get_options", 00:05:16.904 "iscsi_target_node_request_logout", 00:05:16.904 "iscsi_target_node_set_redirect", 00:05:16.904 "iscsi_target_node_set_auth", 00:05:16.904 "iscsi_target_node_add_lun", 00:05:16.904 "iscsi_get_stats", 00:05:16.904 "iscsi_get_connections", 00:05:16.904 "iscsi_portal_group_set_auth", 00:05:16.904 "iscsi_start_portal_group", 00:05:16.904 "iscsi_delete_portal_group", 00:05:16.904 "iscsi_create_portal_group", 00:05:16.904 "iscsi_get_portal_groups", 00:05:16.904 "iscsi_delete_target_node", 00:05:16.904 "iscsi_target_node_remove_pg_ig_maps", 00:05:16.904 "iscsi_target_node_add_pg_ig_maps", 00:05:16.904 "iscsi_create_target_node", 00:05:16.904 "iscsi_get_target_nodes", 00:05:16.904 "iscsi_delete_initiator_group", 00:05:16.904 "iscsi_initiator_group_remove_initiators", 00:05:16.904 "iscsi_initiator_group_add_initiators", 00:05:16.904 "iscsi_create_initiator_group", 00:05:16.904 "iscsi_get_initiator_groups", 00:05:16.904 "nvmf_set_crdt", 00:05:16.904 "nvmf_set_config", 00:05:16.904 "nvmf_set_max_subsystems", 00:05:16.904 "nvmf_stop_mdns_prr", 00:05:16.904 "nvmf_publish_mdns_prr", 00:05:16.904 "nvmf_subsystem_get_listeners", 00:05:16.904 "nvmf_subsystem_get_qpairs", 00:05:16.904 "nvmf_subsystem_get_controllers", 00:05:16.904 "nvmf_get_stats", 00:05:16.904 "nvmf_get_transports", 00:05:16.904 "nvmf_create_transport", 00:05:16.904 "nvmf_get_targets", 00:05:16.904 "nvmf_delete_target", 00:05:16.904 "nvmf_create_target", 00:05:16.904 "nvmf_subsystem_allow_any_host", 00:05:16.904 "nvmf_subsystem_set_keys", 00:05:16.904 "nvmf_subsystem_remove_host", 00:05:16.904 "nvmf_subsystem_add_host", 00:05:16.904 "nvmf_ns_remove_host", 00:05:16.904 "nvmf_ns_add_host", 00:05:16.904 "nvmf_subsystem_remove_ns", 00:05:16.904 "nvmf_subsystem_set_ns_ana_group", 00:05:16.904 "nvmf_subsystem_add_ns", 00:05:16.904 "nvmf_subsystem_listener_set_ana_state", 00:05:16.904 "nvmf_discovery_get_referrals", 00:05:16.904 "nvmf_discovery_remove_referral", 00:05:16.904 "nvmf_discovery_add_referral", 00:05:16.904 "nvmf_subsystem_remove_listener", 00:05:16.904 "nvmf_subsystem_add_listener", 00:05:16.904 "nvmf_delete_subsystem", 00:05:16.904 "nvmf_create_subsystem", 00:05:16.904 "nvmf_get_subsystems", 00:05:16.904 "env_dpdk_get_mem_stats", 00:05:16.904 "nbd_get_disks", 00:05:16.904 "nbd_stop_disk", 00:05:16.904 "nbd_start_disk", 00:05:16.904 "ublk_recover_disk", 00:05:16.904 "ublk_get_disks", 00:05:16.904 "ublk_stop_disk", 00:05:16.904 "ublk_start_disk", 00:05:16.904 "ublk_destroy_target", 00:05:16.904 "ublk_create_target", 00:05:16.904 "virtio_blk_create_transport", 00:05:16.904 "virtio_blk_get_transports", 00:05:16.904 "vhost_controller_set_coalescing", 00:05:16.904 "vhost_get_controllers", 00:05:16.904 "vhost_delete_controller", 00:05:16.904 "vhost_create_blk_controller", 00:05:16.904 "vhost_scsi_controller_remove_target", 00:05:16.904 "vhost_scsi_controller_add_target", 00:05:16.904 "vhost_start_scsi_controller", 00:05:16.904 "vhost_create_scsi_controller", 00:05:16.904 "thread_set_cpumask", 00:05:16.904 "scheduler_set_options", 00:05:16.904 "framework_get_governor", 00:05:16.904 "framework_get_scheduler", 00:05:16.904 "framework_set_scheduler", 00:05:16.904 "framework_get_reactors", 00:05:16.904 "thread_get_io_channels", 00:05:16.904 "thread_get_pollers", 00:05:16.904 "thread_get_stats", 00:05:16.904 "framework_monitor_context_switch", 00:05:16.904 "spdk_kill_instance", 00:05:16.904 "log_enable_timestamps", 00:05:16.904 "log_get_flags", 00:05:16.904 "log_clear_flag", 00:05:16.904 "log_set_flag", 00:05:16.904 "log_get_level", 00:05:16.904 "log_set_level", 00:05:16.904 "log_get_print_level", 00:05:16.904 "log_set_print_level", 00:05:16.904 "framework_enable_cpumask_locks", 00:05:16.904 "framework_disable_cpumask_locks", 00:05:16.904 "framework_wait_init", 00:05:16.904 "framework_start_init", 00:05:16.904 "scsi_get_devices", 00:05:16.904 "bdev_get_histogram", 00:05:16.904 "bdev_enable_histogram", 00:05:16.904 "bdev_set_qos_limit", 00:05:16.904 "bdev_set_qd_sampling_period", 00:05:16.904 "bdev_get_bdevs", 00:05:16.904 "bdev_reset_iostat", 00:05:16.904 "bdev_get_iostat", 00:05:16.904 "bdev_examine", 00:05:16.904 "bdev_wait_for_examine", 00:05:16.904 "bdev_set_options", 00:05:16.904 "accel_get_stats", 00:05:16.904 "accel_set_options", 00:05:16.904 "accel_set_driver", 00:05:16.904 "accel_crypto_key_destroy", 00:05:16.904 "accel_crypto_keys_get", 00:05:16.904 "accel_crypto_key_create", 00:05:16.904 "accel_assign_opc", 00:05:16.904 "accel_get_module_info", 00:05:16.904 "accel_get_opc_assignments", 00:05:16.904 "vmd_rescan", 00:05:16.904 "vmd_remove_device", 00:05:16.904 "vmd_enable", 00:05:16.904 "sock_get_default_impl", 00:05:16.904 "sock_set_default_impl", 00:05:16.904 "sock_impl_set_options", 00:05:16.904 "sock_impl_get_options", 00:05:16.904 "iobuf_get_stats", 00:05:16.904 "iobuf_set_options", 00:05:16.904 "keyring_get_keys", 00:05:16.904 "framework_get_pci_devices", 00:05:16.904 "framework_get_config", 00:05:16.904 "framework_get_subsystems", 00:05:16.904 "fsdev_set_opts", 00:05:16.904 "fsdev_get_opts", 00:05:16.904 "trace_get_info", 00:05:16.904 "trace_get_tpoint_group_mask", 00:05:16.904 "trace_disable_tpoint_group", 00:05:16.904 "trace_enable_tpoint_group", 00:05:16.904 "trace_clear_tpoint_mask", 00:05:16.904 "trace_set_tpoint_mask", 00:05:16.904 "notify_get_notifications", 00:05:16.904 "notify_get_types", 00:05:16.904 "spdk_get_version", 00:05:16.904 "rpc_get_methods" 00:05:16.904 ] 00:05:16.904 23:39:04 spdkcli_tcp -- spdkcli/tcp.sh@35 -- # timing_exit run_spdk_tgt_tcp 00:05:16.904 23:39:04 spdkcli_tcp -- common/autotest_common.sh@732 -- # xtrace_disable 00:05:16.904 23:39:04 spdkcli_tcp -- common/autotest_common.sh@10 -- # set +x 00:05:16.904 23:39:04 spdkcli_tcp -- spdkcli/tcp.sh@37 -- # trap - SIGINT SIGTERM EXIT 00:05:16.904 23:39:04 spdkcli_tcp -- spdkcli/tcp.sh@38 -- # killprocess 69642 00:05:16.904 23:39:04 spdkcli_tcp -- common/autotest_common.sh@954 -- # '[' -z 69642 ']' 00:05:16.904 23:39:04 spdkcli_tcp -- common/autotest_common.sh@958 -- # kill -0 69642 00:05:16.904 23:39:04 spdkcli_tcp -- common/autotest_common.sh@959 -- # uname 00:05:16.904 23:39:04 spdkcli_tcp -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:05:16.904 23:39:04 spdkcli_tcp -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 69642 00:05:16.904 killing process with pid 69642 00:05:16.904 23:39:04 spdkcli_tcp -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:05:16.904 23:39:04 spdkcli_tcp -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:05:16.904 23:39:04 spdkcli_tcp -- common/autotest_common.sh@972 -- # echo 'killing process with pid 69642' 00:05:16.904 23:39:04 spdkcli_tcp -- common/autotest_common.sh@973 -- # kill 69642 00:05:16.904 23:39:04 spdkcli_tcp -- common/autotest_common.sh@978 -- # wait 69642 00:05:17.477 ************************************ 00:05:17.477 END TEST spdkcli_tcp 00:05:17.477 ************************************ 00:05:17.477 00:05:17.477 real 0m1.868s 00:05:17.477 user 0m3.206s 00:05:17.477 sys 0m0.565s 00:05:17.477 23:39:05 spdkcli_tcp -- common/autotest_common.sh@1130 -- # xtrace_disable 00:05:17.477 23:39:05 spdkcli_tcp -- common/autotest_common.sh@10 -- # set +x 00:05:17.477 23:39:05 -- spdk/autotest.sh@167 -- # run_test dpdk_mem_utility /home/vagrant/spdk_repo/spdk/test/dpdk_memory_utility/test_dpdk_mem_info.sh 00:05:17.477 23:39:05 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:05:17.477 23:39:05 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:05:17.477 23:39:05 -- common/autotest_common.sh@10 -- # set +x 00:05:17.477 ************************************ 00:05:17.477 START TEST dpdk_mem_utility 00:05:17.477 ************************************ 00:05:17.477 23:39:05 dpdk_mem_utility -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/dpdk_memory_utility/test_dpdk_mem_info.sh 00:05:17.477 * Looking for test storage... 00:05:17.477 * Found test storage at /home/vagrant/spdk_repo/spdk/test/dpdk_memory_utility 00:05:17.477 23:39:05 dpdk_mem_utility -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:05:17.477 23:39:05 dpdk_mem_utility -- common/autotest_common.sh@1693 -- # lcov --version 00:05:17.477 23:39:05 dpdk_mem_utility -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:05:17.736 23:39:05 dpdk_mem_utility -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:05:17.736 23:39:05 dpdk_mem_utility -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:05:17.736 23:39:05 dpdk_mem_utility -- scripts/common.sh@333 -- # local ver1 ver1_l 00:05:17.736 23:39:05 dpdk_mem_utility -- scripts/common.sh@334 -- # local ver2 ver2_l 00:05:17.736 23:39:05 dpdk_mem_utility -- scripts/common.sh@336 -- # IFS=.-: 00:05:17.736 23:39:05 dpdk_mem_utility -- scripts/common.sh@336 -- # read -ra ver1 00:05:17.736 23:39:05 dpdk_mem_utility -- scripts/common.sh@337 -- # IFS=.-: 00:05:17.736 23:39:05 dpdk_mem_utility -- scripts/common.sh@337 -- # read -ra ver2 00:05:17.736 23:39:05 dpdk_mem_utility -- scripts/common.sh@338 -- # local 'op=<' 00:05:17.736 23:39:05 dpdk_mem_utility -- scripts/common.sh@340 -- # ver1_l=2 00:05:17.736 23:39:05 dpdk_mem_utility -- scripts/common.sh@341 -- # ver2_l=1 00:05:17.736 23:39:05 dpdk_mem_utility -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:05:17.736 23:39:05 dpdk_mem_utility -- scripts/common.sh@344 -- # case "$op" in 00:05:17.736 23:39:05 dpdk_mem_utility -- scripts/common.sh@345 -- # : 1 00:05:17.736 23:39:05 dpdk_mem_utility -- scripts/common.sh@364 -- # (( v = 0 )) 00:05:17.736 23:39:05 dpdk_mem_utility -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:05:17.736 23:39:05 dpdk_mem_utility -- scripts/common.sh@365 -- # decimal 1 00:05:17.736 23:39:05 dpdk_mem_utility -- scripts/common.sh@353 -- # local d=1 00:05:17.736 23:39:05 dpdk_mem_utility -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:05:17.736 23:39:05 dpdk_mem_utility -- scripts/common.sh@355 -- # echo 1 00:05:17.736 23:39:05 dpdk_mem_utility -- scripts/common.sh@365 -- # ver1[v]=1 00:05:17.736 23:39:05 dpdk_mem_utility -- scripts/common.sh@366 -- # decimal 2 00:05:17.736 23:39:05 dpdk_mem_utility -- scripts/common.sh@353 -- # local d=2 00:05:17.736 23:39:05 dpdk_mem_utility -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:05:17.736 23:39:05 dpdk_mem_utility -- scripts/common.sh@355 -- # echo 2 00:05:17.736 23:39:05 dpdk_mem_utility -- scripts/common.sh@366 -- # ver2[v]=2 00:05:17.736 23:39:05 dpdk_mem_utility -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:05:17.736 23:39:05 dpdk_mem_utility -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:05:17.736 23:39:05 dpdk_mem_utility -- scripts/common.sh@368 -- # return 0 00:05:17.736 23:39:05 dpdk_mem_utility -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:05:17.736 23:39:05 dpdk_mem_utility -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:05:17.736 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:17.736 --rc genhtml_branch_coverage=1 00:05:17.736 --rc genhtml_function_coverage=1 00:05:17.736 --rc genhtml_legend=1 00:05:17.736 --rc geninfo_all_blocks=1 00:05:17.736 --rc geninfo_unexecuted_blocks=1 00:05:17.736 00:05:17.736 ' 00:05:17.736 23:39:05 dpdk_mem_utility -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:05:17.736 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:17.736 --rc genhtml_branch_coverage=1 00:05:17.736 --rc genhtml_function_coverage=1 00:05:17.736 --rc genhtml_legend=1 00:05:17.736 --rc geninfo_all_blocks=1 00:05:17.736 --rc geninfo_unexecuted_blocks=1 00:05:17.736 00:05:17.736 ' 00:05:17.736 23:39:05 dpdk_mem_utility -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:05:17.736 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:17.736 --rc genhtml_branch_coverage=1 00:05:17.736 --rc genhtml_function_coverage=1 00:05:17.736 --rc genhtml_legend=1 00:05:17.736 --rc geninfo_all_blocks=1 00:05:17.736 --rc geninfo_unexecuted_blocks=1 00:05:17.736 00:05:17.736 ' 00:05:17.736 23:39:05 dpdk_mem_utility -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:05:17.736 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:17.736 --rc genhtml_branch_coverage=1 00:05:17.736 --rc genhtml_function_coverage=1 00:05:17.736 --rc genhtml_legend=1 00:05:17.736 --rc geninfo_all_blocks=1 00:05:17.736 --rc geninfo_unexecuted_blocks=1 00:05:17.736 00:05:17.736 ' 00:05:17.736 23:39:05 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@10 -- # MEM_SCRIPT=/home/vagrant/spdk_repo/spdk/scripts/dpdk_mem_info.py 00:05:17.737 23:39:05 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@13 -- # spdkpid=69742 00:05:17.737 23:39:05 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@12 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:05:17.737 23:39:05 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@15 -- # waitforlisten 69742 00:05:17.737 23:39:05 dpdk_mem_utility -- common/autotest_common.sh@835 -- # '[' -z 69742 ']' 00:05:17.737 23:39:05 dpdk_mem_utility -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:05:17.737 23:39:05 dpdk_mem_utility -- common/autotest_common.sh@840 -- # local max_retries=100 00:05:17.737 23:39:05 dpdk_mem_utility -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:05:17.737 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:05:17.737 23:39:05 dpdk_mem_utility -- common/autotest_common.sh@844 -- # xtrace_disable 00:05:17.737 23:39:05 dpdk_mem_utility -- common/autotest_common.sh@10 -- # set +x 00:05:17.737 [2024-11-26 23:39:05.746874] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:05:17.737 [2024-11-26 23:39:05.747108] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid69742 ] 00:05:17.996 [2024-11-26 23:39:05.903127] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:05:17.996 [2024-11-26 23:39:05.932114] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:05:18.565 23:39:06 dpdk_mem_utility -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:05:18.565 23:39:06 dpdk_mem_utility -- common/autotest_common.sh@868 -- # return 0 00:05:18.565 23:39:06 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@17 -- # trap 'killprocess $spdkpid' SIGINT SIGTERM EXIT 00:05:18.565 23:39:06 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@19 -- # rpc_cmd env_dpdk_get_mem_stats 00:05:18.565 23:39:06 dpdk_mem_utility -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:18.565 23:39:06 dpdk_mem_utility -- common/autotest_common.sh@10 -- # set +x 00:05:18.565 { 00:05:18.565 "filename": "/tmp/spdk_mem_dump.txt" 00:05:18.565 } 00:05:18.565 23:39:06 dpdk_mem_utility -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:18.565 23:39:06 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@21 -- # /home/vagrant/spdk_repo/spdk/scripts/dpdk_mem_info.py 00:05:18.565 DPDK memory size 818.000000 MiB in 1 heap(s) 00:05:18.565 1 heaps totaling size 818.000000 MiB 00:05:18.565 size: 818.000000 MiB heap id: 0 00:05:18.565 end heaps---------- 00:05:18.565 9 mempools totaling size 603.782043 MiB 00:05:18.565 size: 212.674988 MiB name: PDU_immediate_data_Pool 00:05:18.565 size: 158.602051 MiB name: PDU_data_out_Pool 00:05:18.565 size: 100.555481 MiB name: bdev_io_69742 00:05:18.565 size: 50.003479 MiB name: msgpool_69742 00:05:18.565 size: 36.509338 MiB name: fsdev_io_69742 00:05:18.565 size: 21.763794 MiB name: PDU_Pool 00:05:18.565 size: 19.513306 MiB name: SCSI_TASK_Pool 00:05:18.565 size: 4.133484 MiB name: evtpool_69742 00:05:18.565 size: 0.026123 MiB name: Session_Pool 00:05:18.565 end mempools------- 00:05:18.565 6 memzones totaling size 4.142822 MiB 00:05:18.565 size: 1.000366 MiB name: RG_ring_0_69742 00:05:18.565 size: 1.000366 MiB name: RG_ring_1_69742 00:05:18.566 size: 1.000366 MiB name: RG_ring_4_69742 00:05:18.566 size: 1.000366 MiB name: RG_ring_5_69742 00:05:18.566 size: 0.125366 MiB name: RG_ring_2_69742 00:05:18.566 size: 0.015991 MiB name: RG_ring_3_69742 00:05:18.566 end memzones------- 00:05:18.566 23:39:06 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@23 -- # /home/vagrant/spdk_repo/spdk/scripts/dpdk_mem_info.py -m 0 00:05:18.827 heap id: 0 total size: 818.000000 MiB number of busy elements: 308 number of free elements: 15 00:05:18.827 list of free elements. size: 10.804138 MiB 00:05:18.827 element at address: 0x200019200000 with size: 0.999878 MiB 00:05:18.827 element at address: 0x200019400000 with size: 0.999878 MiB 00:05:18.827 element at address: 0x200032000000 with size: 0.994446 MiB 00:05:18.827 element at address: 0x200000400000 with size: 0.993958 MiB 00:05:18.827 element at address: 0x200006400000 with size: 0.959839 MiB 00:05:18.827 element at address: 0x200012c00000 with size: 0.944275 MiB 00:05:18.827 element at address: 0x200019600000 with size: 0.936584 MiB 00:05:18.827 element at address: 0x200000200000 with size: 0.717346 MiB 00:05:18.827 element at address: 0x20001ae00000 with size: 0.568787 MiB 00:05:18.827 element at address: 0x20000a600000 with size: 0.488892 MiB 00:05:18.827 element at address: 0x200000c00000 with size: 0.486267 MiB 00:05:18.827 element at address: 0x200019800000 with size: 0.485657 MiB 00:05:18.827 element at address: 0x200003e00000 with size: 0.480286 MiB 00:05:18.828 element at address: 0x200028200000 with size: 0.396301 MiB 00:05:18.828 element at address: 0x200000800000 with size: 0.351746 MiB 00:05:18.828 list of standard malloc elements. size: 199.266968 MiB 00:05:18.828 element at address: 0x20000a7fff80 with size: 132.000122 MiB 00:05:18.828 element at address: 0x2000065fff80 with size: 64.000122 MiB 00:05:18.828 element at address: 0x2000192fff80 with size: 1.000122 MiB 00:05:18.828 element at address: 0x2000194fff80 with size: 1.000122 MiB 00:05:18.828 element at address: 0x2000196fff80 with size: 1.000122 MiB 00:05:18.828 element at address: 0x2000003d9f00 with size: 0.140747 MiB 00:05:18.828 element at address: 0x2000196eff00 with size: 0.062622 MiB 00:05:18.828 element at address: 0x2000003fdf80 with size: 0.007935 MiB 00:05:18.828 element at address: 0x2000196efdc0 with size: 0.000305 MiB 00:05:18.828 element at address: 0x2000002d7c40 with size: 0.000183 MiB 00:05:18.828 element at address: 0x2000003d9e40 with size: 0.000183 MiB 00:05:18.828 element at address: 0x2000004fe740 with size: 0.000183 MiB 00:05:18.828 element at address: 0x2000004fe800 with size: 0.000183 MiB 00:05:18.828 element at address: 0x2000004fe8c0 with size: 0.000183 MiB 00:05:18.828 element at address: 0x2000004fe980 with size: 0.000183 MiB 00:05:18.828 element at address: 0x2000004fea40 with size: 0.000183 MiB 00:05:18.828 element at address: 0x2000004feb00 with size: 0.000183 MiB 00:05:18.828 element at address: 0x2000004febc0 with size: 0.000183 MiB 00:05:18.828 element at address: 0x2000004fec80 with size: 0.000183 MiB 00:05:18.828 element at address: 0x2000004fed40 with size: 0.000183 MiB 00:05:18.828 element at address: 0x2000004fee00 with size: 0.000183 MiB 00:05:18.828 element at address: 0x2000004feec0 with size: 0.000183 MiB 00:05:18.828 element at address: 0x2000004fef80 with size: 0.000183 MiB 00:05:18.828 element at address: 0x2000004ff040 with size: 0.000183 MiB 00:05:18.828 element at address: 0x2000004ff100 with size: 0.000183 MiB 00:05:18.828 element at address: 0x2000004ff1c0 with size: 0.000183 MiB 00:05:18.828 element at address: 0x2000004ff280 with size: 0.000183 MiB 00:05:18.828 element at address: 0x2000004ff340 with size: 0.000183 MiB 00:05:18.828 element at address: 0x2000004ff400 with size: 0.000183 MiB 00:05:18.828 element at address: 0x2000004ff4c0 with size: 0.000183 MiB 00:05:18.828 element at address: 0x2000004ff580 with size: 0.000183 MiB 00:05:18.828 element at address: 0x2000004ff640 with size: 0.000183 MiB 00:05:18.828 element at address: 0x2000004ff700 with size: 0.000183 MiB 00:05:18.828 element at address: 0x2000004ff7c0 with size: 0.000183 MiB 00:05:18.828 element at address: 0x2000004ff880 with size: 0.000183 MiB 00:05:18.828 element at address: 0x2000004ff940 with size: 0.000183 MiB 00:05:18.828 element at address: 0x2000004ffa00 with size: 0.000183 MiB 00:05:18.828 element at address: 0x2000004ffac0 with size: 0.000183 MiB 00:05:18.828 element at address: 0x2000004ffcc0 with size: 0.000183 MiB 00:05:18.828 element at address: 0x2000004ffd80 with size: 0.000183 MiB 00:05:18.828 element at address: 0x2000004ffe40 with size: 0.000183 MiB 00:05:18.828 element at address: 0x20000085a0c0 with size: 0.000183 MiB 00:05:18.828 element at address: 0x20000085a2c0 with size: 0.000183 MiB 00:05:18.828 element at address: 0x20000085e580 with size: 0.000183 MiB 00:05:18.828 element at address: 0x20000087e840 with size: 0.000183 MiB 00:05:18.828 element at address: 0x20000087e900 with size: 0.000183 MiB 00:05:18.828 element at address: 0x20000087e9c0 with size: 0.000183 MiB 00:05:18.828 element at address: 0x20000087ea80 with size: 0.000183 MiB 00:05:18.828 element at address: 0x20000087eb40 with size: 0.000183 MiB 00:05:18.828 element at address: 0x20000087ec00 with size: 0.000183 MiB 00:05:18.828 element at address: 0x20000087ecc0 with size: 0.000183 MiB 00:05:18.828 element at address: 0x20000087ed80 with size: 0.000183 MiB 00:05:18.828 element at address: 0x20000087ee40 with size: 0.000183 MiB 00:05:18.828 element at address: 0x20000087ef00 with size: 0.000183 MiB 00:05:18.828 element at address: 0x20000087efc0 with size: 0.000183 MiB 00:05:18.828 element at address: 0x20000087f080 with size: 0.000183 MiB 00:05:18.828 element at address: 0x20000087f140 with size: 0.000183 MiB 00:05:18.828 element at address: 0x20000087f200 with size: 0.000183 MiB 00:05:18.828 element at address: 0x20000087f2c0 with size: 0.000183 MiB 00:05:18.828 element at address: 0x20000087f380 with size: 0.000183 MiB 00:05:18.828 element at address: 0x20000087f440 with size: 0.000183 MiB 00:05:18.828 element at address: 0x20000087f500 with size: 0.000183 MiB 00:05:18.828 element at address: 0x20000087f5c0 with size: 0.000183 MiB 00:05:18.828 element at address: 0x20000087f680 with size: 0.000183 MiB 00:05:18.828 element at address: 0x2000008ff940 with size: 0.000183 MiB 00:05:18.828 element at address: 0x2000008ffb40 with size: 0.000183 MiB 00:05:18.828 element at address: 0x200000c7c7c0 with size: 0.000183 MiB 00:05:18.828 element at address: 0x200000c7c880 with size: 0.000183 MiB 00:05:18.828 element at address: 0x200000c7c940 with size: 0.000183 MiB 00:05:18.828 element at address: 0x200000c7ca00 with size: 0.000183 MiB 00:05:18.828 element at address: 0x200000c7cac0 with size: 0.000183 MiB 00:05:18.828 element at address: 0x200000c7cb80 with size: 0.000183 MiB 00:05:18.828 element at address: 0x200000c7cc40 with size: 0.000183 MiB 00:05:18.828 element at address: 0x200000c7cd00 with size: 0.000183 MiB 00:05:18.828 element at address: 0x200000c7cdc0 with size: 0.000183 MiB 00:05:18.828 element at address: 0x200000c7ce80 with size: 0.000183 MiB 00:05:18.828 element at address: 0x200000c7cf40 with size: 0.000183 MiB 00:05:18.828 element at address: 0x200000c7d000 with size: 0.000183 MiB 00:05:18.828 element at address: 0x200000c7d0c0 with size: 0.000183 MiB 00:05:18.828 element at address: 0x200000c7d180 with size: 0.000183 MiB 00:05:18.828 element at address: 0x200000c7d240 with size: 0.000183 MiB 00:05:18.828 element at address: 0x200000c7d300 with size: 0.000183 MiB 00:05:18.828 element at address: 0x200000c7d3c0 with size: 0.000183 MiB 00:05:18.828 element at address: 0x200000c7d480 with size: 0.000183 MiB 00:05:18.828 element at address: 0x200000c7d540 with size: 0.000183 MiB 00:05:18.828 element at address: 0x200000c7d600 with size: 0.000183 MiB 00:05:18.828 element at address: 0x200000c7d6c0 with size: 0.000183 MiB 00:05:18.828 element at address: 0x200000c7d780 with size: 0.000183 MiB 00:05:18.828 element at address: 0x200000c7d840 with size: 0.000183 MiB 00:05:18.828 element at address: 0x200000c7d900 with size: 0.000183 MiB 00:05:18.828 element at address: 0x200000c7d9c0 with size: 0.000183 MiB 00:05:18.828 element at address: 0x200000c7da80 with size: 0.000183 MiB 00:05:18.828 element at address: 0x200000c7db40 with size: 0.000183 MiB 00:05:18.828 element at address: 0x200000c7dc00 with size: 0.000183 MiB 00:05:18.828 element at address: 0x200000c7dcc0 with size: 0.000183 MiB 00:05:18.828 element at address: 0x200000c7dd80 with size: 0.000183 MiB 00:05:18.828 element at address: 0x200000c7de40 with size: 0.000183 MiB 00:05:18.828 element at address: 0x200000c7df00 with size: 0.000183 MiB 00:05:18.828 element at address: 0x200000c7dfc0 with size: 0.000183 MiB 00:05:18.828 element at address: 0x200000c7e080 with size: 0.000183 MiB 00:05:18.828 element at address: 0x200000c7e140 with size: 0.000183 MiB 00:05:18.828 element at address: 0x200000c7e200 with size: 0.000183 MiB 00:05:18.828 element at address: 0x200000c7e2c0 with size: 0.000183 MiB 00:05:18.828 element at address: 0x200000c7e380 with size: 0.000183 MiB 00:05:18.828 element at address: 0x200000c7e440 with size: 0.000183 MiB 00:05:18.828 element at address: 0x200000c7e500 with size: 0.000183 MiB 00:05:18.828 element at address: 0x200000c7e5c0 with size: 0.000183 MiB 00:05:18.828 element at address: 0x200000c7e680 with size: 0.000183 MiB 00:05:18.828 element at address: 0x200000c7e740 with size: 0.000183 MiB 00:05:18.828 element at address: 0x200000c7e800 with size: 0.000183 MiB 00:05:18.828 element at address: 0x200000c7e8c0 with size: 0.000183 MiB 00:05:18.828 element at address: 0x200000c7e980 with size: 0.000183 MiB 00:05:18.828 element at address: 0x200000c7ea40 with size: 0.000183 MiB 00:05:18.828 element at address: 0x200000c7eb00 with size: 0.000183 MiB 00:05:18.829 element at address: 0x200000c7ebc0 with size: 0.000183 MiB 00:05:18.829 element at address: 0x200000c7ec80 with size: 0.000183 MiB 00:05:18.829 element at address: 0x200000c7ed40 with size: 0.000183 MiB 00:05:18.829 element at address: 0x200000cff000 with size: 0.000183 MiB 00:05:18.829 element at address: 0x200000cff0c0 with size: 0.000183 MiB 00:05:18.829 element at address: 0x200003e7af40 with size: 0.000183 MiB 00:05:18.829 element at address: 0x200003e7b000 with size: 0.000183 MiB 00:05:18.829 element at address: 0x200003e7b0c0 with size: 0.000183 MiB 00:05:18.829 element at address: 0x200003e7b180 with size: 0.000183 MiB 00:05:18.829 element at address: 0x200003e7b240 with size: 0.000183 MiB 00:05:18.829 element at address: 0x200003e7b300 with size: 0.000183 MiB 00:05:18.829 element at address: 0x200003e7b3c0 with size: 0.000183 MiB 00:05:18.829 element at address: 0x200003e7b480 with size: 0.000183 MiB 00:05:18.829 element at address: 0x200003e7b540 with size: 0.000183 MiB 00:05:18.829 element at address: 0x200003e7b600 with size: 0.000183 MiB 00:05:18.829 element at address: 0x200003e7b6c0 with size: 0.000183 MiB 00:05:18.829 element at address: 0x200003efb980 with size: 0.000183 MiB 00:05:18.829 element at address: 0x2000064fdd80 with size: 0.000183 MiB 00:05:18.829 element at address: 0x20000a67d280 with size: 0.000183 MiB 00:05:18.829 element at address: 0x20000a67d340 with size: 0.000183 MiB 00:05:18.829 element at address: 0x20000a67d400 with size: 0.000183 MiB 00:05:18.829 element at address: 0x20000a67d4c0 with size: 0.000183 MiB 00:05:18.829 element at address: 0x20000a67d580 with size: 0.000183 MiB 00:05:18.829 element at address: 0x20000a67d640 with size: 0.000183 MiB 00:05:18.829 element at address: 0x20000a67d700 with size: 0.000183 MiB 00:05:18.829 element at address: 0x20000a67d7c0 with size: 0.000183 MiB 00:05:18.829 element at address: 0x20000a67d880 with size: 0.000183 MiB 00:05:18.829 element at address: 0x20000a67d940 with size: 0.000183 MiB 00:05:18.829 element at address: 0x20000a67da00 with size: 0.000183 MiB 00:05:18.829 element at address: 0x20000a67dac0 with size: 0.000183 MiB 00:05:18.829 element at address: 0x20000a6fdd80 with size: 0.000183 MiB 00:05:18.829 element at address: 0x200012cf1bc0 with size: 0.000183 MiB 00:05:18.829 element at address: 0x2000196efc40 with size: 0.000183 MiB 00:05:18.829 element at address: 0x2000196efd00 with size: 0.000183 MiB 00:05:18.829 element at address: 0x2000198bc740 with size: 0.000183 MiB 00:05:18.829 element at address: 0x20001ae919c0 with size: 0.000183 MiB 00:05:18.829 element at address: 0x20001ae91a80 with size: 0.000183 MiB 00:05:18.829 element at address: 0x20001ae91b40 with size: 0.000183 MiB 00:05:18.829 element at address: 0x20001ae91c00 with size: 0.000183 MiB 00:05:18.829 element at address: 0x20001ae91cc0 with size: 0.000183 MiB 00:05:18.829 element at address: 0x20001ae91d80 with size: 0.000183 MiB 00:05:18.829 element at address: 0x20001ae91e40 with size: 0.000183 MiB 00:05:18.829 element at address: 0x20001ae91f00 with size: 0.000183 MiB 00:05:18.829 element at address: 0x20001ae91fc0 with size: 0.000183 MiB 00:05:18.829 element at address: 0x20001ae92080 with size: 0.000183 MiB 00:05:18.829 element at address: 0x20001ae92140 with size: 0.000183 MiB 00:05:18.829 element at address: 0x20001ae92200 with size: 0.000183 MiB 00:05:18.829 element at address: 0x20001ae922c0 with size: 0.000183 MiB 00:05:18.829 element at address: 0x20001ae92380 with size: 0.000183 MiB 00:05:18.829 element at address: 0x20001ae92440 with size: 0.000183 MiB 00:05:18.829 element at address: 0x20001ae92500 with size: 0.000183 MiB 00:05:18.829 element at address: 0x20001ae925c0 with size: 0.000183 MiB 00:05:18.829 element at address: 0x20001ae92680 with size: 0.000183 MiB 00:05:18.829 element at address: 0x20001ae92740 with size: 0.000183 MiB 00:05:18.829 element at address: 0x20001ae92800 with size: 0.000183 MiB 00:05:18.829 element at address: 0x20001ae928c0 with size: 0.000183 MiB 00:05:18.829 element at address: 0x20001ae92980 with size: 0.000183 MiB 00:05:18.829 element at address: 0x20001ae92a40 with size: 0.000183 MiB 00:05:18.829 element at address: 0x20001ae92b00 with size: 0.000183 MiB 00:05:18.829 element at address: 0x20001ae92bc0 with size: 0.000183 MiB 00:05:18.829 element at address: 0x20001ae92c80 with size: 0.000183 MiB 00:05:18.829 element at address: 0x20001ae92d40 with size: 0.000183 MiB 00:05:18.829 element at address: 0x20001ae92e00 with size: 0.000183 MiB 00:05:18.829 element at address: 0x20001ae92ec0 with size: 0.000183 MiB 00:05:18.829 element at address: 0x20001ae92f80 with size: 0.000183 MiB 00:05:18.829 element at address: 0x20001ae93040 with size: 0.000183 MiB 00:05:18.829 element at address: 0x20001ae93100 with size: 0.000183 MiB 00:05:18.829 element at address: 0x20001ae931c0 with size: 0.000183 MiB 00:05:18.829 element at address: 0x20001ae93280 with size: 0.000183 MiB 00:05:18.829 element at address: 0x20001ae93340 with size: 0.000183 MiB 00:05:18.829 element at address: 0x20001ae93400 with size: 0.000183 MiB 00:05:18.829 element at address: 0x20001ae934c0 with size: 0.000183 MiB 00:05:18.829 element at address: 0x20001ae93580 with size: 0.000183 MiB 00:05:18.829 element at address: 0x20001ae93640 with size: 0.000183 MiB 00:05:18.829 element at address: 0x20001ae93700 with size: 0.000183 MiB 00:05:18.829 element at address: 0x20001ae937c0 with size: 0.000183 MiB 00:05:18.829 element at address: 0x20001ae93880 with size: 0.000183 MiB 00:05:18.829 element at address: 0x20001ae93940 with size: 0.000183 MiB 00:05:18.829 element at address: 0x20001ae93a00 with size: 0.000183 MiB 00:05:18.829 element at address: 0x20001ae93ac0 with size: 0.000183 MiB 00:05:18.829 element at address: 0x20001ae93b80 with size: 0.000183 MiB 00:05:18.829 element at address: 0x20001ae93c40 with size: 0.000183 MiB 00:05:18.829 element at address: 0x20001ae93d00 with size: 0.000183 MiB 00:05:18.829 element at address: 0x20001ae93dc0 with size: 0.000183 MiB 00:05:18.829 element at address: 0x20001ae93e80 with size: 0.000183 MiB 00:05:18.829 element at address: 0x20001ae93f40 with size: 0.000183 MiB 00:05:18.829 element at address: 0x20001ae94000 with size: 0.000183 MiB 00:05:18.829 element at address: 0x20001ae940c0 with size: 0.000183 MiB 00:05:18.829 element at address: 0x20001ae94180 with size: 0.000183 MiB 00:05:18.829 element at address: 0x20001ae94240 with size: 0.000183 MiB 00:05:18.829 element at address: 0x20001ae94300 with size: 0.000183 MiB 00:05:18.829 element at address: 0x20001ae943c0 with size: 0.000183 MiB 00:05:18.829 element at address: 0x20001ae94480 with size: 0.000183 MiB 00:05:18.829 element at address: 0x20001ae94540 with size: 0.000183 MiB 00:05:18.829 element at address: 0x20001ae94600 with size: 0.000183 MiB 00:05:18.829 element at address: 0x20001ae946c0 with size: 0.000183 MiB 00:05:18.829 element at address: 0x20001ae94780 with size: 0.000183 MiB 00:05:18.829 element at address: 0x20001ae94840 with size: 0.000183 MiB 00:05:18.829 element at address: 0x20001ae94900 with size: 0.000183 MiB 00:05:18.829 element at address: 0x20001ae949c0 with size: 0.000183 MiB 00:05:18.829 element at address: 0x20001ae94a80 with size: 0.000183 MiB 00:05:18.829 element at address: 0x20001ae94b40 with size: 0.000183 MiB 00:05:18.829 element at address: 0x20001ae94c00 with size: 0.000183 MiB 00:05:18.829 element at address: 0x20001ae94cc0 with size: 0.000183 MiB 00:05:18.829 element at address: 0x20001ae94d80 with size: 0.000183 MiB 00:05:18.829 element at address: 0x20001ae94e40 with size: 0.000183 MiB 00:05:18.829 element at address: 0x20001ae94f00 with size: 0.000183 MiB 00:05:18.829 element at address: 0x20001ae94fc0 with size: 0.000183 MiB 00:05:18.829 element at address: 0x20001ae95080 with size: 0.000183 MiB 00:05:18.830 element at address: 0x20001ae95140 with size: 0.000183 MiB 00:05:18.830 element at address: 0x20001ae95200 with size: 0.000183 MiB 00:05:18.830 element at address: 0x20001ae952c0 with size: 0.000183 MiB 00:05:18.830 element at address: 0x20001ae95380 with size: 0.000183 MiB 00:05:18.830 element at address: 0x20001ae95440 with size: 0.000183 MiB 00:05:18.830 element at address: 0x200028265740 with size: 0.000183 MiB 00:05:18.830 element at address: 0x200028265800 with size: 0.000183 MiB 00:05:18.830 element at address: 0x20002826c400 with size: 0.000183 MiB 00:05:18.830 element at address: 0x20002826c600 with size: 0.000183 MiB 00:05:18.830 element at address: 0x20002826c6c0 with size: 0.000183 MiB 00:05:18.830 element at address: 0x20002826c780 with size: 0.000183 MiB 00:05:18.830 element at address: 0x20002826c840 with size: 0.000183 MiB 00:05:18.830 element at address: 0x20002826c900 with size: 0.000183 MiB 00:05:18.830 element at address: 0x20002826c9c0 with size: 0.000183 MiB 00:05:18.830 element at address: 0x20002826ca80 with size: 0.000183 MiB 00:05:18.830 element at address: 0x20002826cb40 with size: 0.000183 MiB 00:05:18.830 element at address: 0x20002826cc00 with size: 0.000183 MiB 00:05:18.830 element at address: 0x20002826ccc0 with size: 0.000183 MiB 00:05:18.830 element at address: 0x20002826cd80 with size: 0.000183 MiB 00:05:18.830 element at address: 0x20002826ce40 with size: 0.000183 MiB 00:05:18.830 element at address: 0x20002826cf00 with size: 0.000183 MiB 00:05:18.830 element at address: 0x20002826cfc0 with size: 0.000183 MiB 00:05:18.830 element at address: 0x20002826d080 with size: 0.000183 MiB 00:05:18.830 element at address: 0x20002826d140 with size: 0.000183 MiB 00:05:18.830 element at address: 0x20002826d200 with size: 0.000183 MiB 00:05:18.830 element at address: 0x20002826d2c0 with size: 0.000183 MiB 00:05:18.830 element at address: 0x20002826d380 with size: 0.000183 MiB 00:05:18.830 element at address: 0x20002826d440 with size: 0.000183 MiB 00:05:18.830 element at address: 0x20002826d500 with size: 0.000183 MiB 00:05:18.830 element at address: 0x20002826d5c0 with size: 0.000183 MiB 00:05:18.830 element at address: 0x20002826d680 with size: 0.000183 MiB 00:05:18.830 element at address: 0x20002826d740 with size: 0.000183 MiB 00:05:18.830 element at address: 0x20002826d800 with size: 0.000183 MiB 00:05:18.830 element at address: 0x20002826d8c0 with size: 0.000183 MiB 00:05:18.830 element at address: 0x20002826d980 with size: 0.000183 MiB 00:05:18.830 element at address: 0x20002826da40 with size: 0.000183 MiB 00:05:18.830 element at address: 0x20002826db00 with size: 0.000183 MiB 00:05:18.830 element at address: 0x20002826dbc0 with size: 0.000183 MiB 00:05:18.830 element at address: 0x20002826dc80 with size: 0.000183 MiB 00:05:18.830 element at address: 0x20002826dd40 with size: 0.000183 MiB 00:05:18.830 element at address: 0x20002826de00 with size: 0.000183 MiB 00:05:18.830 element at address: 0x20002826dec0 with size: 0.000183 MiB 00:05:18.830 element at address: 0x20002826df80 with size: 0.000183 MiB 00:05:18.830 element at address: 0x20002826e040 with size: 0.000183 MiB 00:05:18.830 element at address: 0x20002826e100 with size: 0.000183 MiB 00:05:18.830 element at address: 0x20002826e1c0 with size: 0.000183 MiB 00:05:18.830 element at address: 0x20002826e280 with size: 0.000183 MiB 00:05:18.830 element at address: 0x20002826e340 with size: 0.000183 MiB 00:05:18.830 element at address: 0x20002826e400 with size: 0.000183 MiB 00:05:18.830 element at address: 0x20002826e4c0 with size: 0.000183 MiB 00:05:18.830 element at address: 0x20002826e580 with size: 0.000183 MiB 00:05:18.830 element at address: 0x20002826e640 with size: 0.000183 MiB 00:05:18.830 element at address: 0x20002826e700 with size: 0.000183 MiB 00:05:18.830 element at address: 0x20002826e7c0 with size: 0.000183 MiB 00:05:18.830 element at address: 0x20002826e880 with size: 0.000183 MiB 00:05:18.830 element at address: 0x20002826e940 with size: 0.000183 MiB 00:05:18.830 element at address: 0x20002826ea00 with size: 0.000183 MiB 00:05:18.830 element at address: 0x20002826eac0 with size: 0.000183 MiB 00:05:18.830 element at address: 0x20002826eb80 with size: 0.000183 MiB 00:05:18.830 element at address: 0x20002826ec40 with size: 0.000183 MiB 00:05:18.830 element at address: 0x20002826ed00 with size: 0.000183 MiB 00:05:18.830 element at address: 0x20002826edc0 with size: 0.000183 MiB 00:05:18.830 element at address: 0x20002826ee80 with size: 0.000183 MiB 00:05:18.830 element at address: 0x20002826ef40 with size: 0.000183 MiB 00:05:18.830 element at address: 0x20002826f000 with size: 0.000183 MiB 00:05:18.830 element at address: 0x20002826f0c0 with size: 0.000183 MiB 00:05:18.830 element at address: 0x20002826f180 with size: 0.000183 MiB 00:05:18.830 element at address: 0x20002826f240 with size: 0.000183 MiB 00:05:18.830 element at address: 0x20002826f300 with size: 0.000183 MiB 00:05:18.830 element at address: 0x20002826f3c0 with size: 0.000183 MiB 00:05:18.830 element at address: 0x20002826f480 with size: 0.000183 MiB 00:05:18.830 element at address: 0x20002826f540 with size: 0.000183 MiB 00:05:18.830 element at address: 0x20002826f600 with size: 0.000183 MiB 00:05:18.830 element at address: 0x20002826f6c0 with size: 0.000183 MiB 00:05:18.830 element at address: 0x20002826f780 with size: 0.000183 MiB 00:05:18.830 element at address: 0x20002826f840 with size: 0.000183 MiB 00:05:18.830 element at address: 0x20002826f900 with size: 0.000183 MiB 00:05:18.830 element at address: 0x20002826f9c0 with size: 0.000183 MiB 00:05:18.830 element at address: 0x20002826fa80 with size: 0.000183 MiB 00:05:18.830 element at address: 0x20002826fb40 with size: 0.000183 MiB 00:05:18.830 element at address: 0x20002826fc00 with size: 0.000183 MiB 00:05:18.830 element at address: 0x20002826fcc0 with size: 0.000183 MiB 00:05:18.830 element at address: 0x20002826fd80 with size: 0.000183 MiB 00:05:18.830 element at address: 0x20002826fe40 with size: 0.000183 MiB 00:05:18.830 element at address: 0x20002826ff00 with size: 0.000183 MiB 00:05:18.830 list of memzone associated elements. size: 607.928894 MiB 00:05:18.830 element at address: 0x20001ae95500 with size: 211.416748 MiB 00:05:18.830 associated memzone info: size: 211.416626 MiB name: MP_PDU_immediate_data_Pool_0 00:05:18.830 element at address: 0x20002826ffc0 with size: 157.562561 MiB 00:05:18.830 associated memzone info: size: 157.562439 MiB name: MP_PDU_data_out_Pool_0 00:05:18.830 element at address: 0x200012df1e80 with size: 100.055054 MiB 00:05:18.830 associated memzone info: size: 100.054932 MiB name: MP_bdev_io_69742_0 00:05:18.830 element at address: 0x200000dff380 with size: 48.003052 MiB 00:05:18.830 associated memzone info: size: 48.002930 MiB name: MP_msgpool_69742_0 00:05:18.830 element at address: 0x200003ffdb80 with size: 36.008911 MiB 00:05:18.830 associated memzone info: size: 36.008789 MiB name: MP_fsdev_io_69742_0 00:05:18.830 element at address: 0x2000199be940 with size: 20.255554 MiB 00:05:18.830 associated memzone info: size: 20.255432 MiB name: MP_PDU_Pool_0 00:05:18.830 element at address: 0x2000321feb40 with size: 18.005066 MiB 00:05:18.830 associated memzone info: size: 18.004944 MiB name: MP_SCSI_TASK_Pool_0 00:05:18.830 element at address: 0x2000004fff00 with size: 3.000244 MiB 00:05:18.830 associated memzone info: size: 3.000122 MiB name: MP_evtpool_69742_0 00:05:18.830 element at address: 0x2000009ffe00 with size: 2.000488 MiB 00:05:18.830 associated memzone info: size: 2.000366 MiB name: RG_MP_msgpool_69742 00:05:18.830 element at address: 0x2000002d7d00 with size: 1.008118 MiB 00:05:18.830 associated memzone info: size: 1.007996 MiB name: MP_evtpool_69742 00:05:18.830 element at address: 0x20000a6fde40 with size: 1.008118 MiB 00:05:18.831 associated memzone info: size: 1.007996 MiB name: MP_PDU_Pool 00:05:18.831 element at address: 0x2000198bc800 with size: 1.008118 MiB 00:05:18.831 associated memzone info: size: 1.007996 MiB name: MP_PDU_immediate_data_Pool 00:05:18.831 element at address: 0x2000064fde40 with size: 1.008118 MiB 00:05:18.831 associated memzone info: size: 1.007996 MiB name: MP_PDU_data_out_Pool 00:05:18.831 element at address: 0x200003efba40 with size: 1.008118 MiB 00:05:18.831 associated memzone info: size: 1.007996 MiB name: MP_SCSI_TASK_Pool 00:05:18.831 element at address: 0x200000cff180 with size: 1.000488 MiB 00:05:18.831 associated memzone info: size: 1.000366 MiB name: RG_ring_0_69742 00:05:18.831 element at address: 0x2000008ffc00 with size: 1.000488 MiB 00:05:18.831 associated memzone info: size: 1.000366 MiB name: RG_ring_1_69742 00:05:18.831 element at address: 0x200012cf1c80 with size: 1.000488 MiB 00:05:18.831 associated memzone info: size: 1.000366 MiB name: RG_ring_4_69742 00:05:18.831 element at address: 0x2000320fe940 with size: 1.000488 MiB 00:05:18.831 associated memzone info: size: 1.000366 MiB name: RG_ring_5_69742 00:05:18.831 element at address: 0x20000087f740 with size: 0.500488 MiB 00:05:18.831 associated memzone info: size: 0.500366 MiB name: RG_MP_fsdev_io_69742 00:05:18.831 element at address: 0x200000c7ee00 with size: 0.500488 MiB 00:05:18.831 associated memzone info: size: 0.500366 MiB name: RG_MP_bdev_io_69742 00:05:18.831 element at address: 0x20000a67db80 with size: 0.500488 MiB 00:05:18.831 associated memzone info: size: 0.500366 MiB name: RG_MP_PDU_Pool 00:05:18.831 element at address: 0x200003e7b780 with size: 0.500488 MiB 00:05:18.831 associated memzone info: size: 0.500366 MiB name: RG_MP_SCSI_TASK_Pool 00:05:18.831 element at address: 0x20001987c540 with size: 0.250488 MiB 00:05:18.831 associated memzone info: size: 0.250366 MiB name: RG_MP_PDU_immediate_data_Pool 00:05:18.831 element at address: 0x2000002b7a40 with size: 0.125488 MiB 00:05:18.831 associated memzone info: size: 0.125366 MiB name: RG_MP_evtpool_69742 00:05:18.831 element at address: 0x20000085e640 with size: 0.125488 MiB 00:05:18.831 associated memzone info: size: 0.125366 MiB name: RG_ring_2_69742 00:05:18.831 element at address: 0x2000064f5b80 with size: 0.031738 MiB 00:05:18.831 associated memzone info: size: 0.031616 MiB name: RG_MP_PDU_data_out_Pool 00:05:18.831 element at address: 0x2000282658c0 with size: 0.023743 MiB 00:05:18.831 associated memzone info: size: 0.023621 MiB name: MP_Session_Pool_0 00:05:18.831 element at address: 0x20000085a380 with size: 0.016113 MiB 00:05:18.831 associated memzone info: size: 0.015991 MiB name: RG_ring_3_69742 00:05:18.831 element at address: 0x20002826ba00 with size: 0.002441 MiB 00:05:18.831 associated memzone info: size: 0.002319 MiB name: RG_MP_Session_Pool 00:05:18.831 element at address: 0x2000004ffb80 with size: 0.000305 MiB 00:05:18.831 associated memzone info: size: 0.000183 MiB name: MP_msgpool_69742 00:05:18.831 element at address: 0x2000008ffa00 with size: 0.000305 MiB 00:05:18.831 associated memzone info: size: 0.000183 MiB name: MP_fsdev_io_69742 00:05:18.831 element at address: 0x20000085a180 with size: 0.000305 MiB 00:05:18.831 associated memzone info: size: 0.000183 MiB name: MP_bdev_io_69742 00:05:18.831 element at address: 0x20002826c4c0 with size: 0.000305 MiB 00:05:18.831 associated memzone info: size: 0.000183 MiB name: MP_Session_Pool 00:05:18.831 23:39:06 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@25 -- # trap - SIGINT SIGTERM EXIT 00:05:18.831 23:39:06 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@26 -- # killprocess 69742 00:05:18.831 23:39:06 dpdk_mem_utility -- common/autotest_common.sh@954 -- # '[' -z 69742 ']' 00:05:18.831 23:39:06 dpdk_mem_utility -- common/autotest_common.sh@958 -- # kill -0 69742 00:05:18.831 23:39:06 dpdk_mem_utility -- common/autotest_common.sh@959 -- # uname 00:05:18.831 23:39:06 dpdk_mem_utility -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:05:18.831 23:39:06 dpdk_mem_utility -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 69742 00:05:18.831 23:39:06 dpdk_mem_utility -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:05:18.831 23:39:06 dpdk_mem_utility -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:05:18.831 23:39:06 dpdk_mem_utility -- common/autotest_common.sh@972 -- # echo 'killing process with pid 69742' 00:05:18.831 killing process with pid 69742 00:05:18.831 23:39:06 dpdk_mem_utility -- common/autotest_common.sh@973 -- # kill 69742 00:05:18.831 23:39:06 dpdk_mem_utility -- common/autotest_common.sh@978 -- # wait 69742 00:05:19.091 00:05:19.091 real 0m1.682s 00:05:19.091 user 0m1.631s 00:05:19.091 sys 0m0.517s 00:05:19.091 23:39:07 dpdk_mem_utility -- common/autotest_common.sh@1130 -- # xtrace_disable 00:05:19.091 ************************************ 00:05:19.091 END TEST dpdk_mem_utility 00:05:19.091 ************************************ 00:05:19.091 23:39:07 dpdk_mem_utility -- common/autotest_common.sh@10 -- # set +x 00:05:19.091 23:39:07 -- spdk/autotest.sh@168 -- # run_test event /home/vagrant/spdk_repo/spdk/test/event/event.sh 00:05:19.091 23:39:07 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:05:19.091 23:39:07 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:05:19.091 23:39:07 -- common/autotest_common.sh@10 -- # set +x 00:05:19.091 ************************************ 00:05:19.091 START TEST event 00:05:19.091 ************************************ 00:05:19.091 23:39:07 event -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/event/event.sh 00:05:19.352 * Looking for test storage... 00:05:19.352 * Found test storage at /home/vagrant/spdk_repo/spdk/test/event 00:05:19.352 23:39:07 event -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:05:19.352 23:39:07 event -- common/autotest_common.sh@1693 -- # lcov --version 00:05:19.352 23:39:07 event -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:05:19.352 23:39:07 event -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:05:19.352 23:39:07 event -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:05:19.352 23:39:07 event -- scripts/common.sh@333 -- # local ver1 ver1_l 00:05:19.352 23:39:07 event -- scripts/common.sh@334 -- # local ver2 ver2_l 00:05:19.352 23:39:07 event -- scripts/common.sh@336 -- # IFS=.-: 00:05:19.352 23:39:07 event -- scripts/common.sh@336 -- # read -ra ver1 00:05:19.352 23:39:07 event -- scripts/common.sh@337 -- # IFS=.-: 00:05:19.352 23:39:07 event -- scripts/common.sh@337 -- # read -ra ver2 00:05:19.352 23:39:07 event -- scripts/common.sh@338 -- # local 'op=<' 00:05:19.352 23:39:07 event -- scripts/common.sh@340 -- # ver1_l=2 00:05:19.352 23:39:07 event -- scripts/common.sh@341 -- # ver2_l=1 00:05:19.352 23:39:07 event -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:05:19.352 23:39:07 event -- scripts/common.sh@344 -- # case "$op" in 00:05:19.352 23:39:07 event -- scripts/common.sh@345 -- # : 1 00:05:19.352 23:39:07 event -- scripts/common.sh@364 -- # (( v = 0 )) 00:05:19.352 23:39:07 event -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:05:19.352 23:39:07 event -- scripts/common.sh@365 -- # decimal 1 00:05:19.352 23:39:07 event -- scripts/common.sh@353 -- # local d=1 00:05:19.352 23:39:07 event -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:05:19.352 23:39:07 event -- scripts/common.sh@355 -- # echo 1 00:05:19.352 23:39:07 event -- scripts/common.sh@365 -- # ver1[v]=1 00:05:19.352 23:39:07 event -- scripts/common.sh@366 -- # decimal 2 00:05:19.352 23:39:07 event -- scripts/common.sh@353 -- # local d=2 00:05:19.352 23:39:07 event -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:05:19.352 23:39:07 event -- scripts/common.sh@355 -- # echo 2 00:05:19.352 23:39:07 event -- scripts/common.sh@366 -- # ver2[v]=2 00:05:19.352 23:39:07 event -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:05:19.352 23:39:07 event -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:05:19.352 23:39:07 event -- scripts/common.sh@368 -- # return 0 00:05:19.352 23:39:07 event -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:05:19.352 23:39:07 event -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:05:19.352 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:19.352 --rc genhtml_branch_coverage=1 00:05:19.352 --rc genhtml_function_coverage=1 00:05:19.352 --rc genhtml_legend=1 00:05:19.352 --rc geninfo_all_blocks=1 00:05:19.352 --rc geninfo_unexecuted_blocks=1 00:05:19.352 00:05:19.352 ' 00:05:19.352 23:39:07 event -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:05:19.352 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:19.352 --rc genhtml_branch_coverage=1 00:05:19.352 --rc genhtml_function_coverage=1 00:05:19.352 --rc genhtml_legend=1 00:05:19.352 --rc geninfo_all_blocks=1 00:05:19.352 --rc geninfo_unexecuted_blocks=1 00:05:19.352 00:05:19.352 ' 00:05:19.352 23:39:07 event -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:05:19.352 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:19.352 --rc genhtml_branch_coverage=1 00:05:19.352 --rc genhtml_function_coverage=1 00:05:19.352 --rc genhtml_legend=1 00:05:19.352 --rc geninfo_all_blocks=1 00:05:19.352 --rc geninfo_unexecuted_blocks=1 00:05:19.352 00:05:19.352 ' 00:05:19.352 23:39:07 event -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:05:19.352 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:19.352 --rc genhtml_branch_coverage=1 00:05:19.352 --rc genhtml_function_coverage=1 00:05:19.352 --rc genhtml_legend=1 00:05:19.352 --rc geninfo_all_blocks=1 00:05:19.352 --rc geninfo_unexecuted_blocks=1 00:05:19.352 00:05:19.352 ' 00:05:19.352 23:39:07 event -- event/event.sh@9 -- # source /home/vagrant/spdk_repo/spdk/test/bdev/nbd_common.sh 00:05:19.352 23:39:07 event -- bdev/nbd_common.sh@6 -- # set -e 00:05:19.352 23:39:07 event -- event/event.sh@45 -- # run_test event_perf /home/vagrant/spdk_repo/spdk/test/event/event_perf/event_perf -m 0xF -t 1 00:05:19.352 23:39:07 event -- common/autotest_common.sh@1105 -- # '[' 6 -le 1 ']' 00:05:19.352 23:39:07 event -- common/autotest_common.sh@1111 -- # xtrace_disable 00:05:19.352 23:39:07 event -- common/autotest_common.sh@10 -- # set +x 00:05:19.352 ************************************ 00:05:19.352 START TEST event_perf 00:05:19.352 ************************************ 00:05:19.352 23:39:07 event.event_perf -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/event/event_perf/event_perf -m 0xF -t 1 00:05:19.352 Running I/O for 1 seconds...[2024-11-26 23:39:07.467823] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:05:19.352 [2024-11-26 23:39:07.468424] [ DPDK EAL parameters: event_perf --no-shconf -c 0xF --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid69828 ] 00:05:19.612 [2024-11-26 23:39:07.625298] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:05:19.612 [2024-11-26 23:39:07.657964] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:05:19.612 [2024-11-26 23:39:07.658208] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:05:19.612 Running I/O for 1 seconds...[2024-11-26 23:39:07.658152] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:05:19.612 [2024-11-26 23:39:07.658378] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:05:20.994 00:05:20.994 lcore 0: 209077 00:05:20.995 lcore 1: 209078 00:05:20.995 lcore 2: 209078 00:05:20.995 lcore 3: 209078 00:05:20.995 done. 00:05:20.995 00:05:20.995 real 0m1.289s 00:05:20.995 user 0m4.071s 00:05:20.995 sys 0m0.095s 00:05:20.995 23:39:08 event.event_perf -- common/autotest_common.sh@1130 -- # xtrace_disable 00:05:20.995 23:39:08 event.event_perf -- common/autotest_common.sh@10 -- # set +x 00:05:20.995 ************************************ 00:05:20.995 END TEST event_perf 00:05:20.995 ************************************ 00:05:20.995 23:39:08 event -- event/event.sh@46 -- # run_test event_reactor /home/vagrant/spdk_repo/spdk/test/event/reactor/reactor -t 1 00:05:20.995 23:39:08 event -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:05:20.995 23:39:08 event -- common/autotest_common.sh@1111 -- # xtrace_disable 00:05:20.995 23:39:08 event -- common/autotest_common.sh@10 -- # set +x 00:05:20.995 ************************************ 00:05:20.995 START TEST event_reactor 00:05:20.995 ************************************ 00:05:20.995 23:39:08 event.event_reactor -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/event/reactor/reactor -t 1 00:05:20.995 [2024-11-26 23:39:08.830969] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:05:20.995 [2024-11-26 23:39:08.831165] [ DPDK EAL parameters: reactor --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid69862 ] 00:05:20.995 [2024-11-26 23:39:08.985761] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:05:20.995 [2024-11-26 23:39:09.019124] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:05:21.940 test_start 00:05:21.940 oneshot 00:05:21.940 tick 100 00:05:21.940 tick 100 00:05:21.940 tick 250 00:05:21.940 tick 100 00:05:21.940 tick 100 00:05:21.940 tick 100 00:05:21.940 tick 250 00:05:21.940 tick 500 00:05:21.940 tick 100 00:05:21.940 tick 100 00:05:21.940 tick 250 00:05:21.940 tick 100 00:05:21.940 tick 100 00:05:21.940 test_end 00:05:22.200 00:05:22.200 real 0m1.290s 00:05:22.200 user 0m1.104s 00:05:22.200 sys 0m0.079s 00:05:22.200 23:39:10 event.event_reactor -- common/autotest_common.sh@1130 -- # xtrace_disable 00:05:22.200 23:39:10 event.event_reactor -- common/autotest_common.sh@10 -- # set +x 00:05:22.200 ************************************ 00:05:22.200 END TEST event_reactor 00:05:22.200 ************************************ 00:05:22.200 23:39:10 event -- event/event.sh@47 -- # run_test event_reactor_perf /home/vagrant/spdk_repo/spdk/test/event/reactor_perf/reactor_perf -t 1 00:05:22.200 23:39:10 event -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:05:22.200 23:39:10 event -- common/autotest_common.sh@1111 -- # xtrace_disable 00:05:22.200 23:39:10 event -- common/autotest_common.sh@10 -- # set +x 00:05:22.200 ************************************ 00:05:22.200 START TEST event_reactor_perf 00:05:22.200 ************************************ 00:05:22.200 23:39:10 event.event_reactor_perf -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/event/reactor_perf/reactor_perf -t 1 00:05:22.200 [2024-11-26 23:39:10.185700] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:05:22.200 [2024-11-26 23:39:10.185881] [ DPDK EAL parameters: reactor_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid69903 ] 00:05:22.460 [2024-11-26 23:39:10.338122] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:05:22.460 [2024-11-26 23:39:10.362916] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:05:23.405 test_start 00:05:23.405 test_end 00:05:23.405 Performance: 399528 events per second 00:05:23.405 00:05:23.405 real 0m1.274s 00:05:23.405 user 0m1.096s 00:05:23.405 sys 0m0.072s 00:05:23.405 23:39:11 event.event_reactor_perf -- common/autotest_common.sh@1130 -- # xtrace_disable 00:05:23.405 23:39:11 event.event_reactor_perf -- common/autotest_common.sh@10 -- # set +x 00:05:23.406 ************************************ 00:05:23.406 END TEST event_reactor_perf 00:05:23.406 ************************************ 00:05:23.406 23:39:11 event -- event/event.sh@49 -- # uname -s 00:05:23.406 23:39:11 event -- event/event.sh@49 -- # '[' Linux = Linux ']' 00:05:23.406 23:39:11 event -- event/event.sh@50 -- # run_test event_scheduler /home/vagrant/spdk_repo/spdk/test/event/scheduler/scheduler.sh 00:05:23.406 23:39:11 event -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:05:23.406 23:39:11 event -- common/autotest_common.sh@1111 -- # xtrace_disable 00:05:23.406 23:39:11 event -- common/autotest_common.sh@10 -- # set +x 00:05:23.406 ************************************ 00:05:23.406 START TEST event_scheduler 00:05:23.406 ************************************ 00:05:23.406 23:39:11 event.event_scheduler -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/event/scheduler/scheduler.sh 00:05:23.666 * Looking for test storage... 00:05:23.666 * Found test storage at /home/vagrant/spdk_repo/spdk/test/event/scheduler 00:05:23.666 23:39:11 event.event_scheduler -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:05:23.666 23:39:11 event.event_scheduler -- common/autotest_common.sh@1693 -- # lcov --version 00:05:23.666 23:39:11 event.event_scheduler -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:05:23.666 23:39:11 event.event_scheduler -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:05:23.666 23:39:11 event.event_scheduler -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:05:23.666 23:39:11 event.event_scheduler -- scripts/common.sh@333 -- # local ver1 ver1_l 00:05:23.666 23:39:11 event.event_scheduler -- scripts/common.sh@334 -- # local ver2 ver2_l 00:05:23.666 23:39:11 event.event_scheduler -- scripts/common.sh@336 -- # IFS=.-: 00:05:23.666 23:39:11 event.event_scheduler -- scripts/common.sh@336 -- # read -ra ver1 00:05:23.666 23:39:11 event.event_scheduler -- scripts/common.sh@337 -- # IFS=.-: 00:05:23.666 23:39:11 event.event_scheduler -- scripts/common.sh@337 -- # read -ra ver2 00:05:23.666 23:39:11 event.event_scheduler -- scripts/common.sh@338 -- # local 'op=<' 00:05:23.666 23:39:11 event.event_scheduler -- scripts/common.sh@340 -- # ver1_l=2 00:05:23.666 23:39:11 event.event_scheduler -- scripts/common.sh@341 -- # ver2_l=1 00:05:23.666 23:39:11 event.event_scheduler -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:05:23.666 23:39:11 event.event_scheduler -- scripts/common.sh@344 -- # case "$op" in 00:05:23.666 23:39:11 event.event_scheduler -- scripts/common.sh@345 -- # : 1 00:05:23.666 23:39:11 event.event_scheduler -- scripts/common.sh@364 -- # (( v = 0 )) 00:05:23.666 23:39:11 event.event_scheduler -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:05:23.666 23:39:11 event.event_scheduler -- scripts/common.sh@365 -- # decimal 1 00:05:23.666 23:39:11 event.event_scheduler -- scripts/common.sh@353 -- # local d=1 00:05:23.666 23:39:11 event.event_scheduler -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:05:23.666 23:39:11 event.event_scheduler -- scripts/common.sh@355 -- # echo 1 00:05:23.666 23:39:11 event.event_scheduler -- scripts/common.sh@365 -- # ver1[v]=1 00:05:23.666 23:39:11 event.event_scheduler -- scripts/common.sh@366 -- # decimal 2 00:05:23.666 23:39:11 event.event_scheduler -- scripts/common.sh@353 -- # local d=2 00:05:23.666 23:39:11 event.event_scheduler -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:05:23.666 23:39:11 event.event_scheduler -- scripts/common.sh@355 -- # echo 2 00:05:23.666 23:39:11 event.event_scheduler -- scripts/common.sh@366 -- # ver2[v]=2 00:05:23.666 23:39:11 event.event_scheduler -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:05:23.666 23:39:11 event.event_scheduler -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:05:23.666 23:39:11 event.event_scheduler -- scripts/common.sh@368 -- # return 0 00:05:23.666 23:39:11 event.event_scheduler -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:05:23.666 23:39:11 event.event_scheduler -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:05:23.666 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:23.666 --rc genhtml_branch_coverage=1 00:05:23.666 --rc genhtml_function_coverage=1 00:05:23.666 --rc genhtml_legend=1 00:05:23.666 --rc geninfo_all_blocks=1 00:05:23.666 --rc geninfo_unexecuted_blocks=1 00:05:23.666 00:05:23.666 ' 00:05:23.666 23:39:11 event.event_scheduler -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:05:23.666 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:23.666 --rc genhtml_branch_coverage=1 00:05:23.666 --rc genhtml_function_coverage=1 00:05:23.666 --rc genhtml_legend=1 00:05:23.666 --rc geninfo_all_blocks=1 00:05:23.666 --rc geninfo_unexecuted_blocks=1 00:05:23.666 00:05:23.666 ' 00:05:23.666 23:39:11 event.event_scheduler -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:05:23.666 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:23.666 --rc genhtml_branch_coverage=1 00:05:23.666 --rc genhtml_function_coverage=1 00:05:23.666 --rc genhtml_legend=1 00:05:23.666 --rc geninfo_all_blocks=1 00:05:23.666 --rc geninfo_unexecuted_blocks=1 00:05:23.666 00:05:23.666 ' 00:05:23.666 23:39:11 event.event_scheduler -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:05:23.666 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:23.666 --rc genhtml_branch_coverage=1 00:05:23.666 --rc genhtml_function_coverage=1 00:05:23.666 --rc genhtml_legend=1 00:05:23.666 --rc geninfo_all_blocks=1 00:05:23.666 --rc geninfo_unexecuted_blocks=1 00:05:23.666 00:05:23.666 ' 00:05:23.666 23:39:11 event.event_scheduler -- scheduler/scheduler.sh@29 -- # rpc=rpc_cmd 00:05:23.666 23:39:11 event.event_scheduler -- scheduler/scheduler.sh@35 -- # scheduler_pid=69969 00:05:23.666 23:39:11 event.event_scheduler -- scheduler/scheduler.sh@34 -- # /home/vagrant/spdk_repo/spdk/test/event/scheduler/scheduler -m 0xF -p 0x2 --wait-for-rpc -f 00:05:23.666 23:39:11 event.event_scheduler -- scheduler/scheduler.sh@36 -- # trap 'killprocess $scheduler_pid; exit 1' SIGINT SIGTERM EXIT 00:05:23.666 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:05:23.666 23:39:11 event.event_scheduler -- scheduler/scheduler.sh@37 -- # waitforlisten 69969 00:05:23.666 23:39:11 event.event_scheduler -- common/autotest_common.sh@835 -- # '[' -z 69969 ']' 00:05:23.666 23:39:11 event.event_scheduler -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:05:23.666 23:39:11 event.event_scheduler -- common/autotest_common.sh@840 -- # local max_retries=100 00:05:23.666 23:39:11 event.event_scheduler -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:05:23.666 23:39:11 event.event_scheduler -- common/autotest_common.sh@844 -- # xtrace_disable 00:05:23.666 23:39:11 event.event_scheduler -- common/autotest_common.sh@10 -- # set +x 00:05:23.926 [2024-11-26 23:39:11.798280] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:05:23.926 [2024-11-26 23:39:11.798499] [ DPDK EAL parameters: scheduler --no-shconf -c 0xF --main-lcore=2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid69969 ] 00:05:23.926 [2024-11-26 23:39:11.931846] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:05:23.926 [2024-11-26 23:39:11.959826] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:05:23.926 [2024-11-26 23:39:11.960068] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:05:23.926 [2024-11-26 23:39:11.960109] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:05:23.926 [2024-11-26 23:39:11.960246] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:05:24.866 23:39:12 event.event_scheduler -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:05:24.866 23:39:12 event.event_scheduler -- common/autotest_common.sh@868 -- # return 0 00:05:24.866 23:39:12 event.event_scheduler -- scheduler/scheduler.sh@39 -- # rpc_cmd framework_set_scheduler dynamic 00:05:24.866 23:39:12 event.event_scheduler -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:24.866 23:39:12 event.event_scheduler -- common/autotest_common.sh@10 -- # set +x 00:05:24.866 POWER: failed to open /sys/devices/system/cpu/cpu%u/cpufreq/scaling_governor 00:05:24.866 POWER: Cannot set governor of lcore 0 to userspace 00:05:24.866 POWER: failed to open /sys/devices/system/cpu/cpu%u/cpufreq/scaling_governor 00:05:24.866 POWER: Cannot set governor of lcore 0 to performance 00:05:24.866 POWER: failed to open /sys/devices/system/cpu/cpu%u/cpufreq/scaling_governor 00:05:24.866 POWER: Cannot set governor of lcore 0 to userspace 00:05:24.866 GUEST_CHANNEL: Unable to to connect to '/dev/virtio-ports/virtio.serial.port.poweragent.0' with error No such file or directory 00:05:24.866 POWER: Unable to set Power Management Environment for lcore 0 00:05:24.866 [2024-11-26 23:39:12.632543] dpdk_governor.c: 135:_init_core: *ERROR*: Failed to initialize on core0 00:05:24.866 [2024-11-26 23:39:12.632563] dpdk_governor.c: 196:_init: *ERROR*: Failed to initialize on core0 00:05:24.866 [2024-11-26 23:39:12.632586] scheduler_dynamic.c: 280:init: *NOTICE*: Unable to initialize dpdk governor 00:05:24.866 [2024-11-26 23:39:12.632611] scheduler_dynamic.c: 427:set_opts: *NOTICE*: Setting scheduler load limit to 20 00:05:24.866 [2024-11-26 23:39:12.632624] scheduler_dynamic.c: 429:set_opts: *NOTICE*: Setting scheduler core limit to 80 00:05:24.866 [2024-11-26 23:39:12.632634] scheduler_dynamic.c: 431:set_opts: *NOTICE*: Setting scheduler core busy to 95 00:05:24.866 23:39:12 event.event_scheduler -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:24.866 23:39:12 event.event_scheduler -- scheduler/scheduler.sh@40 -- # rpc_cmd framework_start_init 00:05:24.866 23:39:12 event.event_scheduler -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:24.866 23:39:12 event.event_scheduler -- common/autotest_common.sh@10 -- # set +x 00:05:24.866 [2024-11-26 23:39:12.709083] scheduler.c: 382:test_start: *NOTICE*: Scheduler test application started. 00:05:24.866 23:39:12 event.event_scheduler -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:24.866 23:39:12 event.event_scheduler -- scheduler/scheduler.sh@43 -- # run_test scheduler_create_thread scheduler_create_thread 00:05:24.866 23:39:12 event.event_scheduler -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:05:24.866 23:39:12 event.event_scheduler -- common/autotest_common.sh@1111 -- # xtrace_disable 00:05:24.866 23:39:12 event.event_scheduler -- common/autotest_common.sh@10 -- # set +x 00:05:24.866 ************************************ 00:05:24.866 START TEST scheduler_create_thread 00:05:24.866 ************************************ 00:05:24.866 23:39:12 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@1129 -- # scheduler_create_thread 00:05:24.866 23:39:12 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@12 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n active_pinned -m 0x1 -a 100 00:05:24.866 23:39:12 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:24.866 23:39:12 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:05:24.866 2 00:05:24.866 23:39:12 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:24.866 23:39:12 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@13 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n active_pinned -m 0x2 -a 100 00:05:24.866 23:39:12 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:24.866 23:39:12 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:05:24.866 3 00:05:24.866 23:39:12 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:24.866 23:39:12 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@14 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n active_pinned -m 0x4 -a 100 00:05:24.866 23:39:12 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:24.866 23:39:12 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:05:24.866 4 00:05:24.866 23:39:12 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:24.866 23:39:12 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@15 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n active_pinned -m 0x8 -a 100 00:05:24.866 23:39:12 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:24.866 23:39:12 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:05:24.866 5 00:05:24.866 23:39:12 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:24.866 23:39:12 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@16 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n idle_pinned -m 0x1 -a 0 00:05:24.866 23:39:12 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:24.866 23:39:12 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:05:24.866 6 00:05:24.866 23:39:12 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:24.866 23:39:12 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@17 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n idle_pinned -m 0x2 -a 0 00:05:24.866 23:39:12 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:24.866 23:39:12 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:05:24.866 7 00:05:24.866 23:39:12 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:24.866 23:39:12 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@18 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n idle_pinned -m 0x4 -a 0 00:05:24.866 23:39:12 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:24.866 23:39:12 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:05:24.866 8 00:05:24.866 23:39:12 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:24.866 23:39:12 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@19 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n idle_pinned -m 0x8 -a 0 00:05:24.866 23:39:12 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:24.866 23:39:12 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:05:24.866 9 00:05:24.866 23:39:12 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:24.866 23:39:12 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@21 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n one_third_active -a 30 00:05:24.866 23:39:12 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:24.866 23:39:12 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:05:25.435 10 00:05:25.435 23:39:13 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:25.435 23:39:13 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@22 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n half_active -a 0 00:05:25.435 23:39:13 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:25.435 23:39:13 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:05:26.816 23:39:14 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:26.816 23:39:14 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@22 -- # thread_id=11 00:05:26.816 23:39:14 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@23 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_set_active 11 50 00:05:26.816 23:39:14 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:26.816 23:39:14 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:05:27.386 23:39:15 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:27.386 23:39:15 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@25 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n deleted -a 100 00:05:27.386 23:39:15 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:27.386 23:39:15 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:05:28.327 23:39:16 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:28.327 23:39:16 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@25 -- # thread_id=12 00:05:28.327 23:39:16 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@26 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_delete 12 00:05:28.327 23:39:16 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:28.327 23:39:16 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:05:28.897 ************************************ 00:05:28.897 END TEST scheduler_create_thread 00:05:28.897 ************************************ 00:05:28.897 23:39:16 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:28.897 00:05:28.897 real 0m4.210s 00:05:28.897 user 0m0.029s 00:05:28.897 sys 0m0.007s 00:05:28.897 23:39:16 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@1130 -- # xtrace_disable 00:05:28.897 23:39:16 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:05:28.897 23:39:16 event.event_scheduler -- scheduler/scheduler.sh@45 -- # trap - SIGINT SIGTERM EXIT 00:05:28.897 23:39:16 event.event_scheduler -- scheduler/scheduler.sh@46 -- # killprocess 69969 00:05:28.897 23:39:16 event.event_scheduler -- common/autotest_common.sh@954 -- # '[' -z 69969 ']' 00:05:28.897 23:39:16 event.event_scheduler -- common/autotest_common.sh@958 -- # kill -0 69969 00:05:28.897 23:39:16 event.event_scheduler -- common/autotest_common.sh@959 -- # uname 00:05:28.897 23:39:16 event.event_scheduler -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:05:28.897 23:39:16 event.event_scheduler -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 69969 00:05:29.157 killing process with pid 69969 00:05:29.157 23:39:17 event.event_scheduler -- common/autotest_common.sh@960 -- # process_name=reactor_2 00:05:29.157 23:39:17 event.event_scheduler -- common/autotest_common.sh@964 -- # '[' reactor_2 = sudo ']' 00:05:29.157 23:39:17 event.event_scheduler -- common/autotest_common.sh@972 -- # echo 'killing process with pid 69969' 00:05:29.157 23:39:17 event.event_scheduler -- common/autotest_common.sh@973 -- # kill 69969 00:05:29.157 23:39:17 event.event_scheduler -- common/autotest_common.sh@978 -- # wait 69969 00:05:29.157 [2024-11-26 23:39:17.213382] scheduler.c: 360:test_shutdown: *NOTICE*: Scheduler test application stopped. 00:05:29.416 00:05:29.416 real 0m5.983s 00:05:29.416 user 0m13.045s 00:05:29.416 sys 0m0.463s 00:05:29.416 23:39:17 event.event_scheduler -- common/autotest_common.sh@1130 -- # xtrace_disable 00:05:29.416 ************************************ 00:05:29.416 END TEST event_scheduler 00:05:29.416 ************************************ 00:05:29.416 23:39:17 event.event_scheduler -- common/autotest_common.sh@10 -- # set +x 00:05:29.416 23:39:17 event -- event/event.sh@51 -- # modprobe -n nbd 00:05:29.416 23:39:17 event -- event/event.sh@52 -- # run_test app_repeat app_repeat_test 00:05:29.416 23:39:17 event -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:05:29.416 23:39:17 event -- common/autotest_common.sh@1111 -- # xtrace_disable 00:05:29.416 23:39:17 event -- common/autotest_common.sh@10 -- # set +x 00:05:29.416 ************************************ 00:05:29.416 START TEST app_repeat 00:05:29.416 ************************************ 00:05:29.416 23:39:17 event.app_repeat -- common/autotest_common.sh@1129 -- # app_repeat_test 00:05:29.416 23:39:17 event.app_repeat -- event/event.sh@12 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:05:29.416 23:39:17 event.app_repeat -- event/event.sh@13 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:05:29.416 23:39:17 event.app_repeat -- event/event.sh@13 -- # local nbd_list 00:05:29.416 23:39:17 event.app_repeat -- event/event.sh@14 -- # bdev_list=('Malloc0' 'Malloc1') 00:05:29.416 23:39:17 event.app_repeat -- event/event.sh@14 -- # local bdev_list 00:05:29.416 23:39:17 event.app_repeat -- event/event.sh@15 -- # local repeat_times=4 00:05:29.416 23:39:17 event.app_repeat -- event/event.sh@17 -- # modprobe nbd 00:05:29.675 23:39:17 event.app_repeat -- event/event.sh@19 -- # repeat_pid=70086 00:05:29.675 23:39:17 event.app_repeat -- event/event.sh@20 -- # trap 'killprocess $repeat_pid; exit 1' SIGINT SIGTERM EXIT 00:05:29.675 23:39:17 event.app_repeat -- event/event.sh@18 -- # /home/vagrant/spdk_repo/spdk/test/event/app_repeat/app_repeat -r /var/tmp/spdk-nbd.sock -m 0x3 -t 4 00:05:29.675 Process app_repeat pid: 70086 00:05:29.675 spdk_app_start Round 0 00:05:29.675 23:39:17 event.app_repeat -- event/event.sh@21 -- # echo 'Process app_repeat pid: 70086' 00:05:29.675 23:39:17 event.app_repeat -- event/event.sh@23 -- # for i in {0..2} 00:05:29.675 23:39:17 event.app_repeat -- event/event.sh@24 -- # echo 'spdk_app_start Round 0' 00:05:29.675 23:39:17 event.app_repeat -- event/event.sh@25 -- # waitforlisten 70086 /var/tmp/spdk-nbd.sock 00:05:29.675 23:39:17 event.app_repeat -- common/autotest_common.sh@835 -- # '[' -z 70086 ']' 00:05:29.675 23:39:17 event.app_repeat -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk-nbd.sock 00:05:29.675 23:39:17 event.app_repeat -- common/autotest_common.sh@840 -- # local max_retries=100 00:05:29.675 23:39:17 event.app_repeat -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock...' 00:05:29.675 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock... 00:05:29.675 23:39:17 event.app_repeat -- common/autotest_common.sh@844 -- # xtrace_disable 00:05:29.675 23:39:17 event.app_repeat -- common/autotest_common.sh@10 -- # set +x 00:05:29.675 [2024-11-26 23:39:17.603417] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:05:29.675 [2024-11-26 23:39:17.603628] [ DPDK EAL parameters: app_repeat --no-shconf -c 0x3 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid70086 ] 00:05:29.675 [2024-11-26 23:39:17.757027] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 2 00:05:29.675 [2024-11-26 23:39:17.783062] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:05:29.675 [2024-11-26 23:39:17.783159] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:05:30.615 23:39:18 event.app_repeat -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:05:30.615 23:39:18 event.app_repeat -- common/autotest_common.sh@868 -- # return 0 00:05:30.615 23:39:18 event.app_repeat -- event/event.sh@27 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:05:30.615 Malloc0 00:05:30.615 23:39:18 event.app_repeat -- event/event.sh@28 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:05:30.875 Malloc1 00:05:30.875 23:39:18 event.app_repeat -- event/event.sh@30 -- # nbd_rpc_data_verify /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:05:30.875 23:39:18 event.app_repeat -- bdev/nbd_common.sh@90 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:05:30.875 23:39:18 event.app_repeat -- bdev/nbd_common.sh@91 -- # bdev_list=('Malloc0' 'Malloc1') 00:05:30.875 23:39:18 event.app_repeat -- bdev/nbd_common.sh@91 -- # local bdev_list 00:05:30.875 23:39:18 event.app_repeat -- bdev/nbd_common.sh@92 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:05:30.875 23:39:18 event.app_repeat -- bdev/nbd_common.sh@92 -- # local nbd_list 00:05:30.875 23:39:18 event.app_repeat -- bdev/nbd_common.sh@94 -- # nbd_start_disks /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:05:30.875 23:39:18 event.app_repeat -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:05:30.875 23:39:18 event.app_repeat -- bdev/nbd_common.sh@10 -- # bdev_list=('Malloc0' 'Malloc1') 00:05:30.875 23:39:18 event.app_repeat -- bdev/nbd_common.sh@10 -- # local bdev_list 00:05:30.875 23:39:18 event.app_repeat -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:05:30.875 23:39:18 event.app_repeat -- bdev/nbd_common.sh@11 -- # local nbd_list 00:05:30.875 23:39:18 event.app_repeat -- bdev/nbd_common.sh@12 -- # local i 00:05:30.875 23:39:18 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:05:30.875 23:39:18 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:05:30.875 23:39:18 event.app_repeat -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc0 /dev/nbd0 00:05:31.135 /dev/nbd0 00:05:31.135 23:39:19 event.app_repeat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:05:31.135 23:39:19 event.app_repeat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:05:31.135 23:39:19 event.app_repeat -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:05:31.135 23:39:19 event.app_repeat -- common/autotest_common.sh@873 -- # local i 00:05:31.136 23:39:19 event.app_repeat -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:05:31.136 23:39:19 event.app_repeat -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:05:31.136 23:39:19 event.app_repeat -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:05:31.136 23:39:19 event.app_repeat -- common/autotest_common.sh@877 -- # break 00:05:31.136 23:39:19 event.app_repeat -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:05:31.136 23:39:19 event.app_repeat -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:05:31.136 23:39:19 event.app_repeat -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:05:31.136 1+0 records in 00:05:31.136 1+0 records out 00:05:31.136 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000569672 s, 7.2 MB/s 00:05:31.136 23:39:19 event.app_repeat -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:05:31.136 23:39:19 event.app_repeat -- common/autotest_common.sh@890 -- # size=4096 00:05:31.136 23:39:19 event.app_repeat -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:05:31.136 23:39:19 event.app_repeat -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:05:31.136 23:39:19 event.app_repeat -- common/autotest_common.sh@893 -- # return 0 00:05:31.136 23:39:19 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:05:31.136 23:39:19 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:05:31.136 23:39:19 event.app_repeat -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc1 /dev/nbd1 00:05:31.395 /dev/nbd1 00:05:31.395 23:39:19 event.app_repeat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:05:31.395 23:39:19 event.app_repeat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:05:31.395 23:39:19 event.app_repeat -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:05:31.395 23:39:19 event.app_repeat -- common/autotest_common.sh@873 -- # local i 00:05:31.395 23:39:19 event.app_repeat -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:05:31.395 23:39:19 event.app_repeat -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:05:31.395 23:39:19 event.app_repeat -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:05:31.395 23:39:19 event.app_repeat -- common/autotest_common.sh@877 -- # break 00:05:31.395 23:39:19 event.app_repeat -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:05:31.395 23:39:19 event.app_repeat -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:05:31.395 23:39:19 event.app_repeat -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:05:31.395 1+0 records in 00:05:31.395 1+0 records out 00:05:31.395 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000205785 s, 19.9 MB/s 00:05:31.395 23:39:19 event.app_repeat -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:05:31.395 23:39:19 event.app_repeat -- common/autotest_common.sh@890 -- # size=4096 00:05:31.395 23:39:19 event.app_repeat -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:05:31.395 23:39:19 event.app_repeat -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:05:31.395 23:39:19 event.app_repeat -- common/autotest_common.sh@893 -- # return 0 00:05:31.395 23:39:19 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:05:31.395 23:39:19 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:05:31.395 23:39:19 event.app_repeat -- bdev/nbd_common.sh@95 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:05:31.395 23:39:19 event.app_repeat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:05:31.395 23:39:19 event.app_repeat -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:05:31.655 23:39:19 event.app_repeat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[ 00:05:31.655 { 00:05:31.655 "nbd_device": "/dev/nbd0", 00:05:31.655 "bdev_name": "Malloc0" 00:05:31.655 }, 00:05:31.655 { 00:05:31.655 "nbd_device": "/dev/nbd1", 00:05:31.655 "bdev_name": "Malloc1" 00:05:31.655 } 00:05:31.655 ]' 00:05:31.655 23:39:19 event.app_repeat -- bdev/nbd_common.sh@64 -- # echo '[ 00:05:31.655 { 00:05:31.655 "nbd_device": "/dev/nbd0", 00:05:31.655 "bdev_name": "Malloc0" 00:05:31.655 }, 00:05:31.655 { 00:05:31.655 "nbd_device": "/dev/nbd1", 00:05:31.655 "bdev_name": "Malloc1" 00:05:31.655 } 00:05:31.655 ]' 00:05:31.655 23:39:19 event.app_repeat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:05:31.655 23:39:19 event.app_repeat -- bdev/nbd_common.sh@64 -- # nbd_disks_name='/dev/nbd0 00:05:31.655 /dev/nbd1' 00:05:31.655 23:39:19 event.app_repeat -- bdev/nbd_common.sh@65 -- # echo '/dev/nbd0 00:05:31.655 /dev/nbd1' 00:05:31.655 23:39:19 event.app_repeat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:05:31.655 23:39:19 event.app_repeat -- bdev/nbd_common.sh@65 -- # count=2 00:05:31.655 23:39:19 event.app_repeat -- bdev/nbd_common.sh@66 -- # echo 2 00:05:31.655 23:39:19 event.app_repeat -- bdev/nbd_common.sh@95 -- # count=2 00:05:31.655 23:39:19 event.app_repeat -- bdev/nbd_common.sh@96 -- # '[' 2 -ne 2 ']' 00:05:31.655 23:39:19 event.app_repeat -- bdev/nbd_common.sh@100 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' write 00:05:31.655 23:39:19 event.app_repeat -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:05:31.655 23:39:19 event.app_repeat -- bdev/nbd_common.sh@70 -- # local nbd_list 00:05:31.655 23:39:19 event.app_repeat -- bdev/nbd_common.sh@71 -- # local operation=write 00:05:31.655 23:39:19 event.app_repeat -- bdev/nbd_common.sh@72 -- # local tmp_file=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest 00:05:31.655 23:39:19 event.app_repeat -- bdev/nbd_common.sh@74 -- # '[' write = write ']' 00:05:31.655 23:39:19 event.app_repeat -- bdev/nbd_common.sh@76 -- # dd if=/dev/urandom of=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest bs=4096 count=256 00:05:31.655 256+0 records in 00:05:31.655 256+0 records out 00:05:31.655 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0138408 s, 75.8 MB/s 00:05:31.655 23:39:19 event.app_repeat -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:05:31.655 23:39:19 event.app_repeat -- bdev/nbd_common.sh@78 -- # dd if=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest of=/dev/nbd0 bs=4096 count=256 oflag=direct 00:05:31.655 256+0 records in 00:05:31.655 256+0 records out 00:05:31.655 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0245878 s, 42.6 MB/s 00:05:31.655 23:39:19 event.app_repeat -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:05:31.655 23:39:19 event.app_repeat -- bdev/nbd_common.sh@78 -- # dd if=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest of=/dev/nbd1 bs=4096 count=256 oflag=direct 00:05:31.655 256+0 records in 00:05:31.655 256+0 records out 00:05:31.655 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0233961 s, 44.8 MB/s 00:05:31.655 23:39:19 event.app_repeat -- bdev/nbd_common.sh@101 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' verify 00:05:31.655 23:39:19 event.app_repeat -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:05:31.655 23:39:19 event.app_repeat -- bdev/nbd_common.sh@70 -- # local nbd_list 00:05:31.655 23:39:19 event.app_repeat -- bdev/nbd_common.sh@71 -- # local operation=verify 00:05:31.655 23:39:19 event.app_repeat -- bdev/nbd_common.sh@72 -- # local tmp_file=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest 00:05:31.655 23:39:19 event.app_repeat -- bdev/nbd_common.sh@74 -- # '[' verify = write ']' 00:05:31.655 23:39:19 event.app_repeat -- bdev/nbd_common.sh@80 -- # '[' verify = verify ']' 00:05:31.655 23:39:19 event.app_repeat -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:05:31.655 23:39:19 event.app_repeat -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /home/vagrant/spdk_repo/spdk/test/event/nbdrandtest /dev/nbd0 00:05:31.655 23:39:19 event.app_repeat -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:05:31.655 23:39:19 event.app_repeat -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /home/vagrant/spdk_repo/spdk/test/event/nbdrandtest /dev/nbd1 00:05:31.655 23:39:19 event.app_repeat -- bdev/nbd_common.sh@85 -- # rm /home/vagrant/spdk_repo/spdk/test/event/nbdrandtest 00:05:31.655 23:39:19 event.app_repeat -- bdev/nbd_common.sh@103 -- # nbd_stop_disks /var/tmp/spdk-nbd.sock '/dev/nbd0 /dev/nbd1' 00:05:31.655 23:39:19 event.app_repeat -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:05:31.655 23:39:19 event.app_repeat -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:05:31.655 23:39:19 event.app_repeat -- bdev/nbd_common.sh@50 -- # local nbd_list 00:05:31.655 23:39:19 event.app_repeat -- bdev/nbd_common.sh@51 -- # local i 00:05:31.655 23:39:19 event.app_repeat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:05:31.655 23:39:19 event.app_repeat -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd0 00:05:31.915 23:39:19 event.app_repeat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:05:31.915 23:39:19 event.app_repeat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:05:31.915 23:39:19 event.app_repeat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:05:31.915 23:39:19 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:05:31.915 23:39:19 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:05:31.915 23:39:19 event.app_repeat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:05:31.915 23:39:19 event.app_repeat -- bdev/nbd_common.sh@41 -- # break 00:05:31.915 23:39:19 event.app_repeat -- bdev/nbd_common.sh@45 -- # return 0 00:05:31.915 23:39:19 event.app_repeat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:05:31.915 23:39:19 event.app_repeat -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd1 00:05:32.174 23:39:20 event.app_repeat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:05:32.174 23:39:20 event.app_repeat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:05:32.174 23:39:20 event.app_repeat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:05:32.174 23:39:20 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:05:32.174 23:39:20 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:05:32.174 23:39:20 event.app_repeat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:05:32.174 23:39:20 event.app_repeat -- bdev/nbd_common.sh@41 -- # break 00:05:32.174 23:39:20 event.app_repeat -- bdev/nbd_common.sh@45 -- # return 0 00:05:32.174 23:39:20 event.app_repeat -- bdev/nbd_common.sh@104 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:05:32.174 23:39:20 event.app_repeat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:05:32.174 23:39:20 event.app_repeat -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:05:32.433 23:39:20 event.app_repeat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:05:32.433 23:39:20 event.app_repeat -- bdev/nbd_common.sh@64 -- # echo '[]' 00:05:32.433 23:39:20 event.app_repeat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:05:32.433 23:39:20 event.app_repeat -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:05:32.433 23:39:20 event.app_repeat -- bdev/nbd_common.sh@65 -- # echo '' 00:05:32.433 23:39:20 event.app_repeat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:05:32.433 23:39:20 event.app_repeat -- bdev/nbd_common.sh@65 -- # true 00:05:32.433 23:39:20 event.app_repeat -- bdev/nbd_common.sh@65 -- # count=0 00:05:32.433 23:39:20 event.app_repeat -- bdev/nbd_common.sh@66 -- # echo 0 00:05:32.433 23:39:20 event.app_repeat -- bdev/nbd_common.sh@104 -- # count=0 00:05:32.433 23:39:20 event.app_repeat -- bdev/nbd_common.sh@105 -- # '[' 0 -ne 0 ']' 00:05:32.433 23:39:20 event.app_repeat -- bdev/nbd_common.sh@109 -- # return 0 00:05:32.433 23:39:20 event.app_repeat -- event/event.sh@34 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock spdk_kill_instance SIGTERM 00:05:32.693 23:39:20 event.app_repeat -- event/event.sh@35 -- # sleep 3 00:05:32.693 [2024-11-26 23:39:20.781094] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 2 00:05:32.693 [2024-11-26 23:39:20.805306] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:05:32.693 [2024-11-26 23:39:20.805313] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:05:32.953 [2024-11-26 23:39:20.846655] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_register' already registered. 00:05:32.953 [2024-11-26 23:39:20.846754] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_unregister' already registered. 00:05:36.251 spdk_app_start Round 1 00:05:36.251 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock... 00:05:36.251 23:39:23 event.app_repeat -- event/event.sh@23 -- # for i in {0..2} 00:05:36.251 23:39:23 event.app_repeat -- event/event.sh@24 -- # echo 'spdk_app_start Round 1' 00:05:36.251 23:39:23 event.app_repeat -- event/event.sh@25 -- # waitforlisten 70086 /var/tmp/spdk-nbd.sock 00:05:36.251 23:39:23 event.app_repeat -- common/autotest_common.sh@835 -- # '[' -z 70086 ']' 00:05:36.251 23:39:23 event.app_repeat -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk-nbd.sock 00:05:36.251 23:39:23 event.app_repeat -- common/autotest_common.sh@840 -- # local max_retries=100 00:05:36.251 23:39:23 event.app_repeat -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock...' 00:05:36.251 23:39:23 event.app_repeat -- common/autotest_common.sh@844 -- # xtrace_disable 00:05:36.251 23:39:23 event.app_repeat -- common/autotest_common.sh@10 -- # set +x 00:05:36.251 23:39:23 event.app_repeat -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:05:36.251 23:39:23 event.app_repeat -- common/autotest_common.sh@868 -- # return 0 00:05:36.251 23:39:23 event.app_repeat -- event/event.sh@27 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:05:36.251 Malloc0 00:05:36.251 23:39:24 event.app_repeat -- event/event.sh@28 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:05:36.251 Malloc1 00:05:36.251 23:39:24 event.app_repeat -- event/event.sh@30 -- # nbd_rpc_data_verify /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:05:36.251 23:39:24 event.app_repeat -- bdev/nbd_common.sh@90 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:05:36.251 23:39:24 event.app_repeat -- bdev/nbd_common.sh@91 -- # bdev_list=('Malloc0' 'Malloc1') 00:05:36.251 23:39:24 event.app_repeat -- bdev/nbd_common.sh@91 -- # local bdev_list 00:05:36.251 23:39:24 event.app_repeat -- bdev/nbd_common.sh@92 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:05:36.251 23:39:24 event.app_repeat -- bdev/nbd_common.sh@92 -- # local nbd_list 00:05:36.251 23:39:24 event.app_repeat -- bdev/nbd_common.sh@94 -- # nbd_start_disks /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:05:36.251 23:39:24 event.app_repeat -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:05:36.251 23:39:24 event.app_repeat -- bdev/nbd_common.sh@10 -- # bdev_list=('Malloc0' 'Malloc1') 00:05:36.251 23:39:24 event.app_repeat -- bdev/nbd_common.sh@10 -- # local bdev_list 00:05:36.251 23:39:24 event.app_repeat -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:05:36.251 23:39:24 event.app_repeat -- bdev/nbd_common.sh@11 -- # local nbd_list 00:05:36.251 23:39:24 event.app_repeat -- bdev/nbd_common.sh@12 -- # local i 00:05:36.251 23:39:24 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:05:36.251 23:39:24 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:05:36.251 23:39:24 event.app_repeat -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc0 /dev/nbd0 00:05:36.511 /dev/nbd0 00:05:36.511 23:39:24 event.app_repeat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:05:36.511 23:39:24 event.app_repeat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:05:36.511 23:39:24 event.app_repeat -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:05:36.511 23:39:24 event.app_repeat -- common/autotest_common.sh@873 -- # local i 00:05:36.511 23:39:24 event.app_repeat -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:05:36.511 23:39:24 event.app_repeat -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:05:36.511 23:39:24 event.app_repeat -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:05:36.511 23:39:24 event.app_repeat -- common/autotest_common.sh@877 -- # break 00:05:36.511 23:39:24 event.app_repeat -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:05:36.511 23:39:24 event.app_repeat -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:05:36.511 23:39:24 event.app_repeat -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:05:36.511 1+0 records in 00:05:36.511 1+0 records out 00:05:36.511 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000214164 s, 19.1 MB/s 00:05:36.511 23:39:24 event.app_repeat -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:05:36.511 23:39:24 event.app_repeat -- common/autotest_common.sh@890 -- # size=4096 00:05:36.511 23:39:24 event.app_repeat -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:05:36.511 23:39:24 event.app_repeat -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:05:36.511 23:39:24 event.app_repeat -- common/autotest_common.sh@893 -- # return 0 00:05:36.511 23:39:24 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:05:36.511 23:39:24 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:05:36.511 23:39:24 event.app_repeat -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc1 /dev/nbd1 00:05:36.770 /dev/nbd1 00:05:36.770 23:39:24 event.app_repeat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:05:36.770 23:39:24 event.app_repeat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:05:36.770 23:39:24 event.app_repeat -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:05:36.770 23:39:24 event.app_repeat -- common/autotest_common.sh@873 -- # local i 00:05:36.770 23:39:24 event.app_repeat -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:05:36.770 23:39:24 event.app_repeat -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:05:36.770 23:39:24 event.app_repeat -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:05:36.770 23:39:24 event.app_repeat -- common/autotest_common.sh@877 -- # break 00:05:36.770 23:39:24 event.app_repeat -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:05:36.770 23:39:24 event.app_repeat -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:05:36.770 23:39:24 event.app_repeat -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:05:36.770 1+0 records in 00:05:36.770 1+0 records out 00:05:36.770 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000418015 s, 9.8 MB/s 00:05:36.770 23:39:24 event.app_repeat -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:05:36.770 23:39:24 event.app_repeat -- common/autotest_common.sh@890 -- # size=4096 00:05:36.770 23:39:24 event.app_repeat -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:05:36.770 23:39:24 event.app_repeat -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:05:36.770 23:39:24 event.app_repeat -- common/autotest_common.sh@893 -- # return 0 00:05:36.770 23:39:24 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:05:36.770 23:39:24 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:05:36.770 23:39:24 event.app_repeat -- bdev/nbd_common.sh@95 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:05:36.770 23:39:24 event.app_repeat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:05:36.770 23:39:24 event.app_repeat -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:05:37.031 23:39:24 event.app_repeat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[ 00:05:37.031 { 00:05:37.031 "nbd_device": "/dev/nbd0", 00:05:37.031 "bdev_name": "Malloc0" 00:05:37.031 }, 00:05:37.031 { 00:05:37.031 "nbd_device": "/dev/nbd1", 00:05:37.031 "bdev_name": "Malloc1" 00:05:37.031 } 00:05:37.031 ]' 00:05:37.031 23:39:24 event.app_repeat -- bdev/nbd_common.sh@64 -- # echo '[ 00:05:37.031 { 00:05:37.031 "nbd_device": "/dev/nbd0", 00:05:37.031 "bdev_name": "Malloc0" 00:05:37.031 }, 00:05:37.031 { 00:05:37.031 "nbd_device": "/dev/nbd1", 00:05:37.031 "bdev_name": "Malloc1" 00:05:37.031 } 00:05:37.031 ]' 00:05:37.031 23:39:24 event.app_repeat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:05:37.031 23:39:25 event.app_repeat -- bdev/nbd_common.sh@64 -- # nbd_disks_name='/dev/nbd0 00:05:37.031 /dev/nbd1' 00:05:37.031 23:39:25 event.app_repeat -- bdev/nbd_common.sh@65 -- # echo '/dev/nbd0 00:05:37.031 /dev/nbd1' 00:05:37.031 23:39:25 event.app_repeat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:05:37.031 23:39:25 event.app_repeat -- bdev/nbd_common.sh@65 -- # count=2 00:05:37.031 23:39:25 event.app_repeat -- bdev/nbd_common.sh@66 -- # echo 2 00:05:37.031 23:39:25 event.app_repeat -- bdev/nbd_common.sh@95 -- # count=2 00:05:37.031 23:39:25 event.app_repeat -- bdev/nbd_common.sh@96 -- # '[' 2 -ne 2 ']' 00:05:37.031 23:39:25 event.app_repeat -- bdev/nbd_common.sh@100 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' write 00:05:37.031 23:39:25 event.app_repeat -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:05:37.031 23:39:25 event.app_repeat -- bdev/nbd_common.sh@70 -- # local nbd_list 00:05:37.031 23:39:25 event.app_repeat -- bdev/nbd_common.sh@71 -- # local operation=write 00:05:37.031 23:39:25 event.app_repeat -- bdev/nbd_common.sh@72 -- # local tmp_file=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest 00:05:37.031 23:39:25 event.app_repeat -- bdev/nbd_common.sh@74 -- # '[' write = write ']' 00:05:37.031 23:39:25 event.app_repeat -- bdev/nbd_common.sh@76 -- # dd if=/dev/urandom of=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest bs=4096 count=256 00:05:37.031 256+0 records in 00:05:37.031 256+0 records out 00:05:37.031 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.012865 s, 81.5 MB/s 00:05:37.031 23:39:25 event.app_repeat -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:05:37.031 23:39:25 event.app_repeat -- bdev/nbd_common.sh@78 -- # dd if=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest of=/dev/nbd0 bs=4096 count=256 oflag=direct 00:05:37.031 256+0 records in 00:05:37.031 256+0 records out 00:05:37.031 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0196582 s, 53.3 MB/s 00:05:37.031 23:39:25 event.app_repeat -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:05:37.031 23:39:25 event.app_repeat -- bdev/nbd_common.sh@78 -- # dd if=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest of=/dev/nbd1 bs=4096 count=256 oflag=direct 00:05:37.031 256+0 records in 00:05:37.031 256+0 records out 00:05:37.031 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0218741 s, 47.9 MB/s 00:05:37.031 23:39:25 event.app_repeat -- bdev/nbd_common.sh@101 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' verify 00:05:37.031 23:39:25 event.app_repeat -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:05:37.031 23:39:25 event.app_repeat -- bdev/nbd_common.sh@70 -- # local nbd_list 00:05:37.031 23:39:25 event.app_repeat -- bdev/nbd_common.sh@71 -- # local operation=verify 00:05:37.031 23:39:25 event.app_repeat -- bdev/nbd_common.sh@72 -- # local tmp_file=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest 00:05:37.031 23:39:25 event.app_repeat -- bdev/nbd_common.sh@74 -- # '[' verify = write ']' 00:05:37.031 23:39:25 event.app_repeat -- bdev/nbd_common.sh@80 -- # '[' verify = verify ']' 00:05:37.031 23:39:25 event.app_repeat -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:05:37.031 23:39:25 event.app_repeat -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /home/vagrant/spdk_repo/spdk/test/event/nbdrandtest /dev/nbd0 00:05:37.031 23:39:25 event.app_repeat -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:05:37.031 23:39:25 event.app_repeat -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /home/vagrant/spdk_repo/spdk/test/event/nbdrandtest /dev/nbd1 00:05:37.031 23:39:25 event.app_repeat -- bdev/nbd_common.sh@85 -- # rm /home/vagrant/spdk_repo/spdk/test/event/nbdrandtest 00:05:37.031 23:39:25 event.app_repeat -- bdev/nbd_common.sh@103 -- # nbd_stop_disks /var/tmp/spdk-nbd.sock '/dev/nbd0 /dev/nbd1' 00:05:37.031 23:39:25 event.app_repeat -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:05:37.031 23:39:25 event.app_repeat -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:05:37.032 23:39:25 event.app_repeat -- bdev/nbd_common.sh@50 -- # local nbd_list 00:05:37.032 23:39:25 event.app_repeat -- bdev/nbd_common.sh@51 -- # local i 00:05:37.032 23:39:25 event.app_repeat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:05:37.032 23:39:25 event.app_repeat -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd0 00:05:37.292 23:39:25 event.app_repeat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:05:37.292 23:39:25 event.app_repeat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:05:37.292 23:39:25 event.app_repeat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:05:37.292 23:39:25 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:05:37.292 23:39:25 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:05:37.292 23:39:25 event.app_repeat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:05:37.292 23:39:25 event.app_repeat -- bdev/nbd_common.sh@41 -- # break 00:05:37.292 23:39:25 event.app_repeat -- bdev/nbd_common.sh@45 -- # return 0 00:05:37.292 23:39:25 event.app_repeat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:05:37.292 23:39:25 event.app_repeat -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd1 00:05:37.553 23:39:25 event.app_repeat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:05:37.553 23:39:25 event.app_repeat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:05:37.553 23:39:25 event.app_repeat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:05:37.553 23:39:25 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:05:37.553 23:39:25 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:05:37.553 23:39:25 event.app_repeat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:05:37.553 23:39:25 event.app_repeat -- bdev/nbd_common.sh@41 -- # break 00:05:37.553 23:39:25 event.app_repeat -- bdev/nbd_common.sh@45 -- # return 0 00:05:37.553 23:39:25 event.app_repeat -- bdev/nbd_common.sh@104 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:05:37.553 23:39:25 event.app_repeat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:05:37.553 23:39:25 event.app_repeat -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:05:37.813 23:39:25 event.app_repeat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:05:37.813 23:39:25 event.app_repeat -- bdev/nbd_common.sh@64 -- # echo '[]' 00:05:37.813 23:39:25 event.app_repeat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:05:37.813 23:39:25 event.app_repeat -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:05:37.813 23:39:25 event.app_repeat -- bdev/nbd_common.sh@65 -- # echo '' 00:05:37.813 23:39:25 event.app_repeat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:05:37.813 23:39:25 event.app_repeat -- bdev/nbd_common.sh@65 -- # true 00:05:37.813 23:39:25 event.app_repeat -- bdev/nbd_common.sh@65 -- # count=0 00:05:37.813 23:39:25 event.app_repeat -- bdev/nbd_common.sh@66 -- # echo 0 00:05:37.813 23:39:25 event.app_repeat -- bdev/nbd_common.sh@104 -- # count=0 00:05:37.813 23:39:25 event.app_repeat -- bdev/nbd_common.sh@105 -- # '[' 0 -ne 0 ']' 00:05:37.813 23:39:25 event.app_repeat -- bdev/nbd_common.sh@109 -- # return 0 00:05:37.813 23:39:25 event.app_repeat -- event/event.sh@34 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock spdk_kill_instance SIGTERM 00:05:38.073 23:39:26 event.app_repeat -- event/event.sh@35 -- # sleep 3 00:05:38.073 [2024-11-26 23:39:26.185672] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 2 00:05:38.348 [2024-11-26 23:39:26.208862] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:05:38.349 [2024-11-26 23:39:26.208885] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:05:38.349 [2024-11-26 23:39:26.249951] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_register' already registered. 00:05:38.349 [2024-11-26 23:39:26.250026] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_unregister' already registered. 00:05:40.946 23:39:29 event.app_repeat -- event/event.sh@23 -- # for i in {0..2} 00:05:40.946 spdk_app_start Round 2 00:05:40.946 23:39:29 event.app_repeat -- event/event.sh@24 -- # echo 'spdk_app_start Round 2' 00:05:40.946 23:39:29 event.app_repeat -- event/event.sh@25 -- # waitforlisten 70086 /var/tmp/spdk-nbd.sock 00:05:40.946 23:39:29 event.app_repeat -- common/autotest_common.sh@835 -- # '[' -z 70086 ']' 00:05:40.946 23:39:29 event.app_repeat -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk-nbd.sock 00:05:40.946 23:39:29 event.app_repeat -- common/autotest_common.sh@840 -- # local max_retries=100 00:05:40.946 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock... 00:05:40.946 23:39:29 event.app_repeat -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock...' 00:05:40.946 23:39:29 event.app_repeat -- common/autotest_common.sh@844 -- # xtrace_disable 00:05:40.946 23:39:29 event.app_repeat -- common/autotest_common.sh@10 -- # set +x 00:05:41.205 23:39:29 event.app_repeat -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:05:41.205 23:39:29 event.app_repeat -- common/autotest_common.sh@868 -- # return 0 00:05:41.205 23:39:29 event.app_repeat -- event/event.sh@27 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:05:41.465 Malloc0 00:05:41.465 23:39:29 event.app_repeat -- event/event.sh@28 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:05:41.723 Malloc1 00:05:41.723 23:39:29 event.app_repeat -- event/event.sh@30 -- # nbd_rpc_data_verify /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:05:41.723 23:39:29 event.app_repeat -- bdev/nbd_common.sh@90 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:05:41.723 23:39:29 event.app_repeat -- bdev/nbd_common.sh@91 -- # bdev_list=('Malloc0' 'Malloc1') 00:05:41.723 23:39:29 event.app_repeat -- bdev/nbd_common.sh@91 -- # local bdev_list 00:05:41.723 23:39:29 event.app_repeat -- bdev/nbd_common.sh@92 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:05:41.723 23:39:29 event.app_repeat -- bdev/nbd_common.sh@92 -- # local nbd_list 00:05:41.723 23:39:29 event.app_repeat -- bdev/nbd_common.sh@94 -- # nbd_start_disks /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:05:41.723 23:39:29 event.app_repeat -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:05:41.723 23:39:29 event.app_repeat -- bdev/nbd_common.sh@10 -- # bdev_list=('Malloc0' 'Malloc1') 00:05:41.723 23:39:29 event.app_repeat -- bdev/nbd_common.sh@10 -- # local bdev_list 00:05:41.723 23:39:29 event.app_repeat -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:05:41.723 23:39:29 event.app_repeat -- bdev/nbd_common.sh@11 -- # local nbd_list 00:05:41.723 23:39:29 event.app_repeat -- bdev/nbd_common.sh@12 -- # local i 00:05:41.723 23:39:29 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:05:41.723 23:39:29 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:05:41.723 23:39:29 event.app_repeat -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc0 /dev/nbd0 00:05:41.983 /dev/nbd0 00:05:41.983 23:39:29 event.app_repeat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:05:41.983 23:39:29 event.app_repeat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:05:41.983 23:39:29 event.app_repeat -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:05:41.983 23:39:29 event.app_repeat -- common/autotest_common.sh@873 -- # local i 00:05:41.983 23:39:29 event.app_repeat -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:05:41.983 23:39:29 event.app_repeat -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:05:41.983 23:39:29 event.app_repeat -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:05:41.983 23:39:29 event.app_repeat -- common/autotest_common.sh@877 -- # break 00:05:41.983 23:39:29 event.app_repeat -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:05:41.983 23:39:29 event.app_repeat -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:05:41.983 23:39:29 event.app_repeat -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:05:41.983 1+0 records in 00:05:41.983 1+0 records out 00:05:41.983 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.0002147 s, 19.1 MB/s 00:05:41.983 23:39:29 event.app_repeat -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:05:41.983 23:39:29 event.app_repeat -- common/autotest_common.sh@890 -- # size=4096 00:05:41.983 23:39:29 event.app_repeat -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:05:41.983 23:39:29 event.app_repeat -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:05:41.983 23:39:29 event.app_repeat -- common/autotest_common.sh@893 -- # return 0 00:05:41.983 23:39:29 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:05:41.983 23:39:29 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:05:41.983 23:39:29 event.app_repeat -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc1 /dev/nbd1 00:05:42.243 /dev/nbd1 00:05:42.243 23:39:30 event.app_repeat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:05:42.243 23:39:30 event.app_repeat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:05:42.243 23:39:30 event.app_repeat -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:05:42.243 23:39:30 event.app_repeat -- common/autotest_common.sh@873 -- # local i 00:05:42.243 23:39:30 event.app_repeat -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:05:42.243 23:39:30 event.app_repeat -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:05:42.243 23:39:30 event.app_repeat -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:05:42.243 23:39:30 event.app_repeat -- common/autotest_common.sh@877 -- # break 00:05:42.243 23:39:30 event.app_repeat -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:05:42.243 23:39:30 event.app_repeat -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:05:42.243 23:39:30 event.app_repeat -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:05:42.243 1+0 records in 00:05:42.243 1+0 records out 00:05:42.243 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000228637 s, 17.9 MB/s 00:05:42.243 23:39:30 event.app_repeat -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:05:42.243 23:39:30 event.app_repeat -- common/autotest_common.sh@890 -- # size=4096 00:05:42.243 23:39:30 event.app_repeat -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:05:42.243 23:39:30 event.app_repeat -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:05:42.243 23:39:30 event.app_repeat -- common/autotest_common.sh@893 -- # return 0 00:05:42.243 23:39:30 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:05:42.243 23:39:30 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:05:42.243 23:39:30 event.app_repeat -- bdev/nbd_common.sh@95 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:05:42.243 23:39:30 event.app_repeat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:05:42.243 23:39:30 event.app_repeat -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:05:42.503 23:39:30 event.app_repeat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[ 00:05:42.503 { 00:05:42.503 "nbd_device": "/dev/nbd0", 00:05:42.503 "bdev_name": "Malloc0" 00:05:42.503 }, 00:05:42.503 { 00:05:42.503 "nbd_device": "/dev/nbd1", 00:05:42.503 "bdev_name": "Malloc1" 00:05:42.503 } 00:05:42.503 ]' 00:05:42.503 23:39:30 event.app_repeat -- bdev/nbd_common.sh@64 -- # echo '[ 00:05:42.503 { 00:05:42.503 "nbd_device": "/dev/nbd0", 00:05:42.503 "bdev_name": "Malloc0" 00:05:42.503 }, 00:05:42.503 { 00:05:42.503 "nbd_device": "/dev/nbd1", 00:05:42.503 "bdev_name": "Malloc1" 00:05:42.503 } 00:05:42.503 ]' 00:05:42.503 23:39:30 event.app_repeat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:05:42.503 23:39:30 event.app_repeat -- bdev/nbd_common.sh@64 -- # nbd_disks_name='/dev/nbd0 00:05:42.503 /dev/nbd1' 00:05:42.503 23:39:30 event.app_repeat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:05:42.503 23:39:30 event.app_repeat -- bdev/nbd_common.sh@65 -- # echo '/dev/nbd0 00:05:42.503 /dev/nbd1' 00:05:42.503 23:39:30 event.app_repeat -- bdev/nbd_common.sh@65 -- # count=2 00:05:42.503 23:39:30 event.app_repeat -- bdev/nbd_common.sh@66 -- # echo 2 00:05:42.503 23:39:30 event.app_repeat -- bdev/nbd_common.sh@95 -- # count=2 00:05:42.503 23:39:30 event.app_repeat -- bdev/nbd_common.sh@96 -- # '[' 2 -ne 2 ']' 00:05:42.503 23:39:30 event.app_repeat -- bdev/nbd_common.sh@100 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' write 00:05:42.503 23:39:30 event.app_repeat -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:05:42.503 23:39:30 event.app_repeat -- bdev/nbd_common.sh@70 -- # local nbd_list 00:05:42.503 23:39:30 event.app_repeat -- bdev/nbd_common.sh@71 -- # local operation=write 00:05:42.503 23:39:30 event.app_repeat -- bdev/nbd_common.sh@72 -- # local tmp_file=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest 00:05:42.503 23:39:30 event.app_repeat -- bdev/nbd_common.sh@74 -- # '[' write = write ']' 00:05:42.503 23:39:30 event.app_repeat -- bdev/nbd_common.sh@76 -- # dd if=/dev/urandom of=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest bs=4096 count=256 00:05:42.503 256+0 records in 00:05:42.503 256+0 records out 00:05:42.503 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0136063 s, 77.1 MB/s 00:05:42.503 23:39:30 event.app_repeat -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:05:42.503 23:39:30 event.app_repeat -- bdev/nbd_common.sh@78 -- # dd if=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest of=/dev/nbd0 bs=4096 count=256 oflag=direct 00:05:42.503 256+0 records in 00:05:42.503 256+0 records out 00:05:42.503 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0252002 s, 41.6 MB/s 00:05:42.503 23:39:30 event.app_repeat -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:05:42.503 23:39:30 event.app_repeat -- bdev/nbd_common.sh@78 -- # dd if=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest of=/dev/nbd1 bs=4096 count=256 oflag=direct 00:05:42.503 256+0 records in 00:05:42.503 256+0 records out 00:05:42.503 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0189062 s, 55.5 MB/s 00:05:42.503 23:39:30 event.app_repeat -- bdev/nbd_common.sh@101 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' verify 00:05:42.503 23:39:30 event.app_repeat -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:05:42.503 23:39:30 event.app_repeat -- bdev/nbd_common.sh@70 -- # local nbd_list 00:05:42.503 23:39:30 event.app_repeat -- bdev/nbd_common.sh@71 -- # local operation=verify 00:05:42.503 23:39:30 event.app_repeat -- bdev/nbd_common.sh@72 -- # local tmp_file=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest 00:05:42.503 23:39:30 event.app_repeat -- bdev/nbd_common.sh@74 -- # '[' verify = write ']' 00:05:42.503 23:39:30 event.app_repeat -- bdev/nbd_common.sh@80 -- # '[' verify = verify ']' 00:05:42.503 23:39:30 event.app_repeat -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:05:42.503 23:39:30 event.app_repeat -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /home/vagrant/spdk_repo/spdk/test/event/nbdrandtest /dev/nbd0 00:05:42.503 23:39:30 event.app_repeat -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:05:42.503 23:39:30 event.app_repeat -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /home/vagrant/spdk_repo/spdk/test/event/nbdrandtest /dev/nbd1 00:05:42.503 23:39:30 event.app_repeat -- bdev/nbd_common.sh@85 -- # rm /home/vagrant/spdk_repo/spdk/test/event/nbdrandtest 00:05:42.503 23:39:30 event.app_repeat -- bdev/nbd_common.sh@103 -- # nbd_stop_disks /var/tmp/spdk-nbd.sock '/dev/nbd0 /dev/nbd1' 00:05:42.503 23:39:30 event.app_repeat -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:05:42.503 23:39:30 event.app_repeat -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:05:42.503 23:39:30 event.app_repeat -- bdev/nbd_common.sh@50 -- # local nbd_list 00:05:42.503 23:39:30 event.app_repeat -- bdev/nbd_common.sh@51 -- # local i 00:05:42.503 23:39:30 event.app_repeat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:05:42.503 23:39:30 event.app_repeat -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd0 00:05:42.764 23:39:30 event.app_repeat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:05:42.764 23:39:30 event.app_repeat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:05:42.764 23:39:30 event.app_repeat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:05:42.764 23:39:30 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:05:42.764 23:39:30 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:05:42.764 23:39:30 event.app_repeat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:05:42.764 23:39:30 event.app_repeat -- bdev/nbd_common.sh@41 -- # break 00:05:42.764 23:39:30 event.app_repeat -- bdev/nbd_common.sh@45 -- # return 0 00:05:42.764 23:39:30 event.app_repeat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:05:42.764 23:39:30 event.app_repeat -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd1 00:05:43.023 23:39:30 event.app_repeat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:05:43.023 23:39:30 event.app_repeat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:05:43.023 23:39:30 event.app_repeat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:05:43.023 23:39:30 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:05:43.023 23:39:30 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:05:43.023 23:39:30 event.app_repeat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:05:43.023 23:39:30 event.app_repeat -- bdev/nbd_common.sh@41 -- # break 00:05:43.023 23:39:30 event.app_repeat -- bdev/nbd_common.sh@45 -- # return 0 00:05:43.023 23:39:30 event.app_repeat -- bdev/nbd_common.sh@104 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:05:43.023 23:39:30 event.app_repeat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:05:43.023 23:39:30 event.app_repeat -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:05:43.282 23:39:31 event.app_repeat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:05:43.282 23:39:31 event.app_repeat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:05:43.282 23:39:31 event.app_repeat -- bdev/nbd_common.sh@64 -- # echo '[]' 00:05:43.282 23:39:31 event.app_repeat -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:05:43.282 23:39:31 event.app_repeat -- bdev/nbd_common.sh@65 -- # echo '' 00:05:43.282 23:39:31 event.app_repeat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:05:43.282 23:39:31 event.app_repeat -- bdev/nbd_common.sh@65 -- # true 00:05:43.282 23:39:31 event.app_repeat -- bdev/nbd_common.sh@65 -- # count=0 00:05:43.282 23:39:31 event.app_repeat -- bdev/nbd_common.sh@66 -- # echo 0 00:05:43.282 23:39:31 event.app_repeat -- bdev/nbd_common.sh@104 -- # count=0 00:05:43.282 23:39:31 event.app_repeat -- bdev/nbd_common.sh@105 -- # '[' 0 -ne 0 ']' 00:05:43.282 23:39:31 event.app_repeat -- bdev/nbd_common.sh@109 -- # return 0 00:05:43.282 23:39:31 event.app_repeat -- event/event.sh@34 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock spdk_kill_instance SIGTERM 00:05:43.541 23:39:31 event.app_repeat -- event/event.sh@35 -- # sleep 3 00:05:43.541 [2024-11-26 23:39:31.563237] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 2 00:05:43.541 [2024-11-26 23:39:31.586419] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:05:43.541 [2024-11-26 23:39:31.586425] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:05:43.541 [2024-11-26 23:39:31.627457] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_register' already registered. 00:05:43.541 [2024-11-26 23:39:31.627517] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_unregister' already registered. 00:05:46.837 23:39:34 event.app_repeat -- event/event.sh@38 -- # waitforlisten 70086 /var/tmp/spdk-nbd.sock 00:05:46.837 23:39:34 event.app_repeat -- common/autotest_common.sh@835 -- # '[' -z 70086 ']' 00:05:46.837 23:39:34 event.app_repeat -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk-nbd.sock 00:05:46.837 23:39:34 event.app_repeat -- common/autotest_common.sh@840 -- # local max_retries=100 00:05:46.837 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock... 00:05:46.837 23:39:34 event.app_repeat -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock...' 00:05:46.837 23:39:34 event.app_repeat -- common/autotest_common.sh@844 -- # xtrace_disable 00:05:46.837 23:39:34 event.app_repeat -- common/autotest_common.sh@10 -- # set +x 00:05:46.837 23:39:34 event.app_repeat -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:05:46.837 23:39:34 event.app_repeat -- common/autotest_common.sh@868 -- # return 0 00:05:46.837 23:39:34 event.app_repeat -- event/event.sh@39 -- # killprocess 70086 00:05:46.837 23:39:34 event.app_repeat -- common/autotest_common.sh@954 -- # '[' -z 70086 ']' 00:05:46.837 23:39:34 event.app_repeat -- common/autotest_common.sh@958 -- # kill -0 70086 00:05:46.837 23:39:34 event.app_repeat -- common/autotest_common.sh@959 -- # uname 00:05:46.837 23:39:34 event.app_repeat -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:05:46.837 23:39:34 event.app_repeat -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 70086 00:05:46.837 23:39:34 event.app_repeat -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:05:46.837 23:39:34 event.app_repeat -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:05:46.837 killing process with pid 70086 00:05:46.837 23:39:34 event.app_repeat -- common/autotest_common.sh@972 -- # echo 'killing process with pid 70086' 00:05:46.837 23:39:34 event.app_repeat -- common/autotest_common.sh@973 -- # kill 70086 00:05:46.837 23:39:34 event.app_repeat -- common/autotest_common.sh@978 -- # wait 70086 00:05:46.837 spdk_app_start is called in Round 0. 00:05:46.837 Shutdown signal received, stop current app iteration 00:05:46.837 Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 reinitialization... 00:05:46.837 spdk_app_start is called in Round 1. 00:05:46.837 Shutdown signal received, stop current app iteration 00:05:46.837 Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 reinitialization... 00:05:46.837 spdk_app_start is called in Round 2. 00:05:46.837 Shutdown signal received, stop current app iteration 00:05:46.837 Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 reinitialization... 00:05:46.837 spdk_app_start is called in Round 3. 00:05:46.837 Shutdown signal received, stop current app iteration 00:05:46.837 23:39:34 event.app_repeat -- event/event.sh@40 -- # trap - SIGINT SIGTERM EXIT 00:05:46.837 23:39:34 event.app_repeat -- event/event.sh@42 -- # return 0 00:05:46.837 00:05:46.837 real 0m17.299s 00:05:46.837 user 0m38.205s 00:05:46.837 sys 0m2.660s 00:05:46.837 23:39:34 event.app_repeat -- common/autotest_common.sh@1130 -- # xtrace_disable 00:05:46.837 23:39:34 event.app_repeat -- common/autotest_common.sh@10 -- # set +x 00:05:46.837 ************************************ 00:05:46.837 END TEST app_repeat 00:05:46.837 ************************************ 00:05:46.837 23:39:34 event -- event/event.sh@54 -- # (( SPDK_TEST_CRYPTO == 0 )) 00:05:46.837 23:39:34 event -- event/event.sh@55 -- # run_test cpu_locks /home/vagrant/spdk_repo/spdk/test/event/cpu_locks.sh 00:05:46.837 23:39:34 event -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:05:46.837 23:39:34 event -- common/autotest_common.sh@1111 -- # xtrace_disable 00:05:46.837 23:39:34 event -- common/autotest_common.sh@10 -- # set +x 00:05:46.837 ************************************ 00:05:46.837 START TEST cpu_locks 00:05:46.837 ************************************ 00:05:46.837 23:39:34 event.cpu_locks -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/event/cpu_locks.sh 00:05:47.098 * Looking for test storage... 00:05:47.098 * Found test storage at /home/vagrant/spdk_repo/spdk/test/event 00:05:47.098 23:39:35 event.cpu_locks -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:05:47.098 23:39:35 event.cpu_locks -- common/autotest_common.sh@1693 -- # lcov --version 00:05:47.098 23:39:35 event.cpu_locks -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:05:47.098 23:39:35 event.cpu_locks -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:05:47.098 23:39:35 event.cpu_locks -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:05:47.098 23:39:35 event.cpu_locks -- scripts/common.sh@333 -- # local ver1 ver1_l 00:05:47.098 23:39:35 event.cpu_locks -- scripts/common.sh@334 -- # local ver2 ver2_l 00:05:47.098 23:39:35 event.cpu_locks -- scripts/common.sh@336 -- # IFS=.-: 00:05:47.098 23:39:35 event.cpu_locks -- scripts/common.sh@336 -- # read -ra ver1 00:05:47.098 23:39:35 event.cpu_locks -- scripts/common.sh@337 -- # IFS=.-: 00:05:47.098 23:39:35 event.cpu_locks -- scripts/common.sh@337 -- # read -ra ver2 00:05:47.098 23:39:35 event.cpu_locks -- scripts/common.sh@338 -- # local 'op=<' 00:05:47.098 23:39:35 event.cpu_locks -- scripts/common.sh@340 -- # ver1_l=2 00:05:47.098 23:39:35 event.cpu_locks -- scripts/common.sh@341 -- # ver2_l=1 00:05:47.098 23:39:35 event.cpu_locks -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:05:47.098 23:39:35 event.cpu_locks -- scripts/common.sh@344 -- # case "$op" in 00:05:47.098 23:39:35 event.cpu_locks -- scripts/common.sh@345 -- # : 1 00:05:47.098 23:39:35 event.cpu_locks -- scripts/common.sh@364 -- # (( v = 0 )) 00:05:47.098 23:39:35 event.cpu_locks -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:05:47.098 23:39:35 event.cpu_locks -- scripts/common.sh@365 -- # decimal 1 00:05:47.098 23:39:35 event.cpu_locks -- scripts/common.sh@353 -- # local d=1 00:05:47.098 23:39:35 event.cpu_locks -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:05:47.098 23:39:35 event.cpu_locks -- scripts/common.sh@355 -- # echo 1 00:05:47.098 23:39:35 event.cpu_locks -- scripts/common.sh@365 -- # ver1[v]=1 00:05:47.098 23:39:35 event.cpu_locks -- scripts/common.sh@366 -- # decimal 2 00:05:47.098 23:39:35 event.cpu_locks -- scripts/common.sh@353 -- # local d=2 00:05:47.098 23:39:35 event.cpu_locks -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:05:47.098 23:39:35 event.cpu_locks -- scripts/common.sh@355 -- # echo 2 00:05:47.098 23:39:35 event.cpu_locks -- scripts/common.sh@366 -- # ver2[v]=2 00:05:47.098 23:39:35 event.cpu_locks -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:05:47.098 23:39:35 event.cpu_locks -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:05:47.098 23:39:35 event.cpu_locks -- scripts/common.sh@368 -- # return 0 00:05:47.098 23:39:35 event.cpu_locks -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:05:47.098 23:39:35 event.cpu_locks -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:05:47.098 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:47.098 --rc genhtml_branch_coverage=1 00:05:47.098 --rc genhtml_function_coverage=1 00:05:47.098 --rc genhtml_legend=1 00:05:47.098 --rc geninfo_all_blocks=1 00:05:47.098 --rc geninfo_unexecuted_blocks=1 00:05:47.098 00:05:47.098 ' 00:05:47.098 23:39:35 event.cpu_locks -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:05:47.098 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:47.098 --rc genhtml_branch_coverage=1 00:05:47.098 --rc genhtml_function_coverage=1 00:05:47.098 --rc genhtml_legend=1 00:05:47.098 --rc geninfo_all_blocks=1 00:05:47.098 --rc geninfo_unexecuted_blocks=1 00:05:47.098 00:05:47.098 ' 00:05:47.098 23:39:35 event.cpu_locks -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:05:47.098 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:47.098 --rc genhtml_branch_coverage=1 00:05:47.098 --rc genhtml_function_coverage=1 00:05:47.098 --rc genhtml_legend=1 00:05:47.098 --rc geninfo_all_blocks=1 00:05:47.098 --rc geninfo_unexecuted_blocks=1 00:05:47.098 00:05:47.098 ' 00:05:47.098 23:39:35 event.cpu_locks -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:05:47.098 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:47.098 --rc genhtml_branch_coverage=1 00:05:47.098 --rc genhtml_function_coverage=1 00:05:47.098 --rc genhtml_legend=1 00:05:47.098 --rc geninfo_all_blocks=1 00:05:47.098 --rc geninfo_unexecuted_blocks=1 00:05:47.098 00:05:47.098 ' 00:05:47.098 23:39:35 event.cpu_locks -- event/cpu_locks.sh@11 -- # rpc_sock1=/var/tmp/spdk.sock 00:05:47.098 23:39:35 event.cpu_locks -- event/cpu_locks.sh@12 -- # rpc_sock2=/var/tmp/spdk2.sock 00:05:47.098 23:39:35 event.cpu_locks -- event/cpu_locks.sh@164 -- # trap cleanup EXIT SIGTERM SIGINT 00:05:47.098 23:39:35 event.cpu_locks -- event/cpu_locks.sh@166 -- # run_test default_locks default_locks 00:05:47.098 23:39:35 event.cpu_locks -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:05:47.098 23:39:35 event.cpu_locks -- common/autotest_common.sh@1111 -- # xtrace_disable 00:05:47.098 23:39:35 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:05:47.098 ************************************ 00:05:47.098 START TEST default_locks 00:05:47.098 ************************************ 00:05:47.098 23:39:35 event.cpu_locks.default_locks -- common/autotest_common.sh@1129 -- # default_locks 00:05:47.098 23:39:35 event.cpu_locks.default_locks -- event/cpu_locks.sh@46 -- # spdk_tgt_pid=70506 00:05:47.098 23:39:35 event.cpu_locks.default_locks -- event/cpu_locks.sh@45 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 00:05:47.098 23:39:35 event.cpu_locks.default_locks -- event/cpu_locks.sh@47 -- # waitforlisten 70506 00:05:47.098 23:39:35 event.cpu_locks.default_locks -- common/autotest_common.sh@835 -- # '[' -z 70506 ']' 00:05:47.098 23:39:35 event.cpu_locks.default_locks -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:05:47.098 23:39:35 event.cpu_locks.default_locks -- common/autotest_common.sh@840 -- # local max_retries=100 00:05:47.098 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:05:47.098 23:39:35 event.cpu_locks.default_locks -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:05:47.098 23:39:35 event.cpu_locks.default_locks -- common/autotest_common.sh@844 -- # xtrace_disable 00:05:47.098 23:39:35 event.cpu_locks.default_locks -- common/autotest_common.sh@10 -- # set +x 00:05:47.358 [2024-11-26 23:39:35.238719] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:05:47.358 [2024-11-26 23:39:35.238864] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid70506 ] 00:05:47.358 [2024-11-26 23:39:35.392265] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:05:47.358 [2024-11-26 23:39:35.417722] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:05:47.927 23:39:36 event.cpu_locks.default_locks -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:05:47.927 23:39:36 event.cpu_locks.default_locks -- common/autotest_common.sh@868 -- # return 0 00:05:47.927 23:39:36 event.cpu_locks.default_locks -- event/cpu_locks.sh@49 -- # locks_exist 70506 00:05:47.927 23:39:36 event.cpu_locks.default_locks -- event/cpu_locks.sh@22 -- # lslocks -p 70506 00:05:47.927 23:39:36 event.cpu_locks.default_locks -- event/cpu_locks.sh@22 -- # grep -q spdk_cpu_lock 00:05:48.496 23:39:36 event.cpu_locks.default_locks -- event/cpu_locks.sh@50 -- # killprocess 70506 00:05:48.496 23:39:36 event.cpu_locks.default_locks -- common/autotest_common.sh@954 -- # '[' -z 70506 ']' 00:05:48.496 23:39:36 event.cpu_locks.default_locks -- common/autotest_common.sh@958 -- # kill -0 70506 00:05:48.496 23:39:36 event.cpu_locks.default_locks -- common/autotest_common.sh@959 -- # uname 00:05:48.496 23:39:36 event.cpu_locks.default_locks -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:05:48.496 23:39:36 event.cpu_locks.default_locks -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 70506 00:05:48.496 23:39:36 event.cpu_locks.default_locks -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:05:48.496 23:39:36 event.cpu_locks.default_locks -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:05:48.496 killing process with pid 70506 00:05:48.496 23:39:36 event.cpu_locks.default_locks -- common/autotest_common.sh@972 -- # echo 'killing process with pid 70506' 00:05:48.496 23:39:36 event.cpu_locks.default_locks -- common/autotest_common.sh@973 -- # kill 70506 00:05:48.496 23:39:36 event.cpu_locks.default_locks -- common/autotest_common.sh@978 -- # wait 70506 00:05:48.757 23:39:36 event.cpu_locks.default_locks -- event/cpu_locks.sh@52 -- # NOT waitforlisten 70506 00:05:48.757 23:39:36 event.cpu_locks.default_locks -- common/autotest_common.sh@652 -- # local es=0 00:05:48.757 23:39:36 event.cpu_locks.default_locks -- common/autotest_common.sh@654 -- # valid_exec_arg waitforlisten 70506 00:05:48.757 23:39:36 event.cpu_locks.default_locks -- common/autotest_common.sh@640 -- # local arg=waitforlisten 00:05:48.757 23:39:36 event.cpu_locks.default_locks -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:05:48.757 23:39:36 event.cpu_locks.default_locks -- common/autotest_common.sh@644 -- # type -t waitforlisten 00:05:48.757 23:39:36 event.cpu_locks.default_locks -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:05:48.757 23:39:36 event.cpu_locks.default_locks -- common/autotest_common.sh@655 -- # waitforlisten 70506 00:05:48.757 23:39:36 event.cpu_locks.default_locks -- common/autotest_common.sh@835 -- # '[' -z 70506 ']' 00:05:48.757 23:39:36 event.cpu_locks.default_locks -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:05:48.757 23:39:36 event.cpu_locks.default_locks -- common/autotest_common.sh@840 -- # local max_retries=100 00:05:48.757 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:05:48.757 23:39:36 event.cpu_locks.default_locks -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:05:48.757 23:39:36 event.cpu_locks.default_locks -- common/autotest_common.sh@844 -- # xtrace_disable 00:05:48.757 23:39:36 event.cpu_locks.default_locks -- common/autotest_common.sh@10 -- # set +x 00:05:48.757 /home/vagrant/spdk_repo/spdk/test/common/autotest_common.sh: line 850: kill: (70506) - No such process 00:05:48.757 ERROR: process (pid: 70506) is no longer running 00:05:48.757 23:39:36 event.cpu_locks.default_locks -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:05:48.757 23:39:36 event.cpu_locks.default_locks -- common/autotest_common.sh@868 -- # return 1 00:05:48.757 23:39:36 event.cpu_locks.default_locks -- common/autotest_common.sh@655 -- # es=1 00:05:48.757 23:39:36 event.cpu_locks.default_locks -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:05:48.757 23:39:36 event.cpu_locks.default_locks -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:05:48.757 23:39:36 event.cpu_locks.default_locks -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:05:48.757 23:39:36 event.cpu_locks.default_locks -- event/cpu_locks.sh@54 -- # no_locks 00:05:48.757 23:39:36 event.cpu_locks.default_locks -- event/cpu_locks.sh@26 -- # lock_files=() 00:05:48.757 23:39:36 event.cpu_locks.default_locks -- event/cpu_locks.sh@26 -- # local lock_files 00:05:48.757 23:39:36 event.cpu_locks.default_locks -- event/cpu_locks.sh@27 -- # (( 0 != 0 )) 00:05:48.757 00:05:48.757 real 0m1.718s 00:05:48.757 user 0m1.719s 00:05:48.757 sys 0m0.563s 00:05:48.757 23:39:36 event.cpu_locks.default_locks -- common/autotest_common.sh@1130 -- # xtrace_disable 00:05:48.757 23:39:36 event.cpu_locks.default_locks -- common/autotest_common.sh@10 -- # set +x 00:05:48.757 ************************************ 00:05:48.757 END TEST default_locks 00:05:48.757 ************************************ 00:05:49.017 23:39:36 event.cpu_locks -- event/cpu_locks.sh@167 -- # run_test default_locks_via_rpc default_locks_via_rpc 00:05:49.017 23:39:36 event.cpu_locks -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:05:49.017 23:39:36 event.cpu_locks -- common/autotest_common.sh@1111 -- # xtrace_disable 00:05:49.017 23:39:36 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:05:49.017 ************************************ 00:05:49.017 START TEST default_locks_via_rpc 00:05:49.017 ************************************ 00:05:49.017 23:39:36 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@1129 -- # default_locks_via_rpc 00:05:49.017 23:39:36 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@62 -- # spdk_tgt_pid=70559 00:05:49.017 23:39:36 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@61 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 00:05:49.017 23:39:36 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@63 -- # waitforlisten 70559 00:05:49.017 23:39:36 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@835 -- # '[' -z 70559 ']' 00:05:49.017 23:39:36 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:05:49.017 23:39:36 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@840 -- # local max_retries=100 00:05:49.017 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:05:49.017 23:39:36 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:05:49.017 23:39:36 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@844 -- # xtrace_disable 00:05:49.017 23:39:36 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:05:49.017 [2024-11-26 23:39:37.028717] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:05:49.017 [2024-11-26 23:39:37.028868] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid70559 ] 00:05:49.277 [2024-11-26 23:39:37.181745] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:05:49.277 [2024-11-26 23:39:37.206354] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:05:49.846 23:39:37 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:05:49.846 23:39:37 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@868 -- # return 0 00:05:49.846 23:39:37 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@65 -- # rpc_cmd framework_disable_cpumask_locks 00:05:49.846 23:39:37 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:49.846 23:39:37 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:05:49.846 23:39:37 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:49.846 23:39:37 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@67 -- # no_locks 00:05:49.846 23:39:37 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@26 -- # lock_files=() 00:05:49.846 23:39:37 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@26 -- # local lock_files 00:05:49.846 23:39:37 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@27 -- # (( 0 != 0 )) 00:05:49.846 23:39:37 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@69 -- # rpc_cmd framework_enable_cpumask_locks 00:05:49.846 23:39:37 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:49.846 23:39:37 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:05:49.846 23:39:37 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:49.846 23:39:37 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@71 -- # locks_exist 70559 00:05:49.846 23:39:37 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@22 -- # lslocks -p 70559 00:05:49.846 23:39:37 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@22 -- # grep -q spdk_cpu_lock 00:05:50.105 23:39:38 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@73 -- # killprocess 70559 00:05:50.105 23:39:38 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@954 -- # '[' -z 70559 ']' 00:05:50.105 23:39:38 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@958 -- # kill -0 70559 00:05:50.105 23:39:38 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@959 -- # uname 00:05:50.105 23:39:38 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:05:50.105 23:39:38 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 70559 00:05:50.105 23:39:38 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:05:50.105 23:39:38 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:05:50.105 killing process with pid 70559 00:05:50.105 23:39:38 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@972 -- # echo 'killing process with pid 70559' 00:05:50.105 23:39:38 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@973 -- # kill 70559 00:05:50.105 23:39:38 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@978 -- # wait 70559 00:05:50.365 00:05:50.365 real 0m1.503s 00:05:50.365 user 0m1.485s 00:05:50.365 sys 0m0.474s 00:05:50.365 23:39:38 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@1130 -- # xtrace_disable 00:05:50.365 23:39:38 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:05:50.365 ************************************ 00:05:50.365 END TEST default_locks_via_rpc 00:05:50.365 ************************************ 00:05:50.365 23:39:38 event.cpu_locks -- event/cpu_locks.sh@168 -- # run_test non_locking_app_on_locked_coremask non_locking_app_on_locked_coremask 00:05:50.365 23:39:38 event.cpu_locks -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:05:50.365 23:39:38 event.cpu_locks -- common/autotest_common.sh@1111 -- # xtrace_disable 00:05:50.365 23:39:38 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:05:50.624 ************************************ 00:05:50.624 START TEST non_locking_app_on_locked_coremask 00:05:50.624 ************************************ 00:05:50.624 23:39:38 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@1129 -- # non_locking_app_on_locked_coremask 00:05:50.624 23:39:38 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@80 -- # spdk_tgt_pid=70600 00:05:50.624 23:39:38 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@79 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 00:05:50.624 23:39:38 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@81 -- # waitforlisten 70600 /var/tmp/spdk.sock 00:05:50.624 23:39:38 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@835 -- # '[' -z 70600 ']' 00:05:50.624 23:39:38 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:05:50.624 23:39:38 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@840 -- # local max_retries=100 00:05:50.624 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:05:50.624 23:39:38 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:05:50.624 23:39:38 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@844 -- # xtrace_disable 00:05:50.624 23:39:38 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@10 -- # set +x 00:05:50.624 [2024-11-26 23:39:38.598928] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:05:50.624 [2024-11-26 23:39:38.599042] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid70600 ] 00:05:50.884 [2024-11-26 23:39:38.754775] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:05:50.884 [2024-11-26 23:39:38.778824] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:05:51.463 23:39:39 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:05:51.463 23:39:39 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@868 -- # return 0 00:05:51.463 23:39:39 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@83 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 --disable-cpumask-locks -r /var/tmp/spdk2.sock 00:05:51.463 23:39:39 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@84 -- # spdk_tgt_pid2=70616 00:05:51.463 23:39:39 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@85 -- # waitforlisten 70616 /var/tmp/spdk2.sock 00:05:51.463 23:39:39 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@835 -- # '[' -z 70616 ']' 00:05:51.463 23:39:39 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk2.sock 00:05:51.463 23:39:39 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@840 -- # local max_retries=100 00:05:51.463 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock... 00:05:51.463 23:39:39 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock...' 00:05:51.463 23:39:39 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@844 -- # xtrace_disable 00:05:51.463 23:39:39 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@10 -- # set +x 00:05:51.463 [2024-11-26 23:39:39.460951] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:05:51.463 [2024-11-26 23:39:39.461082] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid70616 ] 00:05:51.722 [2024-11-26 23:39:39.609973] app.c: 916:spdk_app_start: *NOTICE*: CPU core locks deactivated. 00:05:51.722 [2024-11-26 23:39:39.610025] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:05:51.722 [2024-11-26 23:39:39.665819] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:05:52.291 23:39:40 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:05:52.291 23:39:40 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@868 -- # return 0 00:05:52.291 23:39:40 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@87 -- # locks_exist 70600 00:05:52.291 23:39:40 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@22 -- # lslocks -p 70600 00:05:52.291 23:39:40 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@22 -- # grep -q spdk_cpu_lock 00:05:52.860 23:39:40 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@89 -- # killprocess 70600 00:05:52.860 23:39:40 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@954 -- # '[' -z 70600 ']' 00:05:52.860 23:39:40 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@958 -- # kill -0 70600 00:05:52.860 23:39:40 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@959 -- # uname 00:05:52.860 23:39:40 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:05:52.860 23:39:40 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 70600 00:05:52.860 killing process with pid 70600 00:05:52.860 23:39:40 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:05:52.860 23:39:40 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:05:52.860 23:39:40 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@972 -- # echo 'killing process with pid 70600' 00:05:52.860 23:39:40 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@973 -- # kill 70600 00:05:52.860 23:39:40 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@978 -- # wait 70600 00:05:53.430 23:39:41 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@90 -- # killprocess 70616 00:05:53.430 23:39:41 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@954 -- # '[' -z 70616 ']' 00:05:53.430 23:39:41 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@958 -- # kill -0 70616 00:05:53.430 23:39:41 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@959 -- # uname 00:05:53.430 23:39:41 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:05:53.430 23:39:41 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 70616 00:05:53.690 23:39:41 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:05:53.690 23:39:41 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:05:53.690 23:39:41 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@972 -- # echo 'killing process with pid 70616' 00:05:53.690 killing process with pid 70616 00:05:53.690 23:39:41 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@973 -- # kill 70616 00:05:53.690 23:39:41 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@978 -- # wait 70616 00:05:53.950 00:05:53.950 real 0m3.420s 00:05:53.950 user 0m3.585s 00:05:53.950 sys 0m1.016s 00:05:53.950 23:39:41 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@1130 -- # xtrace_disable 00:05:53.950 23:39:41 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@10 -- # set +x 00:05:53.950 ************************************ 00:05:53.950 END TEST non_locking_app_on_locked_coremask 00:05:53.950 ************************************ 00:05:53.950 23:39:41 event.cpu_locks -- event/cpu_locks.sh@169 -- # run_test locking_app_on_unlocked_coremask locking_app_on_unlocked_coremask 00:05:53.950 23:39:41 event.cpu_locks -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:05:53.950 23:39:41 event.cpu_locks -- common/autotest_common.sh@1111 -- # xtrace_disable 00:05:53.950 23:39:41 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:05:53.950 ************************************ 00:05:53.950 START TEST locking_app_on_unlocked_coremask 00:05:53.950 ************************************ 00:05:53.950 23:39:41 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@1129 -- # locking_app_on_unlocked_coremask 00:05:53.950 23:39:41 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@98 -- # spdk_tgt_pid=70685 00:05:53.950 23:39:41 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@97 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 --disable-cpumask-locks 00:05:53.950 23:39:41 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@99 -- # waitforlisten 70685 /var/tmp/spdk.sock 00:05:53.950 23:39:41 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@835 -- # '[' -z 70685 ']' 00:05:53.950 23:39:41 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:05:53.950 23:39:41 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@840 -- # local max_retries=100 00:05:53.950 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:05:53.950 23:39:41 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:05:53.950 23:39:41 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@844 -- # xtrace_disable 00:05:53.950 23:39:41 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@10 -- # set +x 00:05:54.210 [2024-11-26 23:39:42.084995] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:05:54.210 [2024-11-26 23:39:42.085135] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid70685 ] 00:05:54.210 [2024-11-26 23:39:42.237258] app.c: 916:spdk_app_start: *NOTICE*: CPU core locks deactivated. 00:05:54.210 [2024-11-26 23:39:42.237324] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:05:54.210 [2024-11-26 23:39:42.261943] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:05:55.145 23:39:42 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:05:55.145 23:39:42 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@868 -- # return 0 00:05:55.145 23:39:42 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@102 -- # spdk_tgt_pid2=70695 00:05:55.145 23:39:42 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@103 -- # waitforlisten 70695 /var/tmp/spdk2.sock 00:05:55.145 23:39:42 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@835 -- # '[' -z 70695 ']' 00:05:55.145 23:39:42 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk2.sock 00:05:55.145 23:39:42 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@840 -- # local max_retries=100 00:05:55.145 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock... 00:05:55.145 23:39:42 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock...' 00:05:55.145 23:39:42 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@844 -- # xtrace_disable 00:05:55.145 23:39:42 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@101 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 -r /var/tmp/spdk2.sock 00:05:55.145 23:39:42 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@10 -- # set +x 00:05:55.145 [2024-11-26 23:39:42.996579] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:05:55.145 [2024-11-26 23:39:42.996705] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid70695 ] 00:05:55.145 [2024-11-26 23:39:43.143268] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:05:55.145 [2024-11-26 23:39:43.192999] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:05:55.712 23:39:43 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:05:55.712 23:39:43 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@868 -- # return 0 00:05:55.712 23:39:43 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@105 -- # locks_exist 70695 00:05:55.712 23:39:43 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@22 -- # lslocks -p 70695 00:05:55.712 23:39:43 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@22 -- # grep -q spdk_cpu_lock 00:05:56.281 23:39:44 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@107 -- # killprocess 70685 00:05:56.281 23:39:44 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@954 -- # '[' -z 70685 ']' 00:05:56.281 23:39:44 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@958 -- # kill -0 70685 00:05:56.281 23:39:44 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@959 -- # uname 00:05:56.281 23:39:44 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:05:56.281 23:39:44 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 70685 00:05:56.281 23:39:44 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:05:56.281 23:39:44 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:05:56.281 killing process with pid 70685 00:05:56.281 23:39:44 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@972 -- # echo 'killing process with pid 70685' 00:05:56.281 23:39:44 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@973 -- # kill 70685 00:05:56.281 23:39:44 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@978 -- # wait 70685 00:05:56.885 23:39:44 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@108 -- # killprocess 70695 00:05:56.885 23:39:44 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@954 -- # '[' -z 70695 ']' 00:05:56.885 23:39:44 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@958 -- # kill -0 70695 00:05:56.885 23:39:44 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@959 -- # uname 00:05:56.885 23:39:44 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:05:56.885 23:39:44 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 70695 00:05:56.885 23:39:44 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:05:56.885 23:39:44 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:05:56.885 killing process with pid 70695 00:05:56.886 23:39:44 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@972 -- # echo 'killing process with pid 70695' 00:05:56.886 23:39:44 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@973 -- # kill 70695 00:05:56.886 23:39:44 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@978 -- # wait 70695 00:05:57.145 00:05:57.145 real 0m3.273s 00:05:57.145 user 0m3.460s 00:05:57.145 sys 0m0.963s 00:05:57.145 23:39:45 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@1130 -- # xtrace_disable 00:05:57.145 23:39:45 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@10 -- # set +x 00:05:57.145 ************************************ 00:05:57.145 END TEST locking_app_on_unlocked_coremask 00:05:57.145 ************************************ 00:05:57.403 23:39:45 event.cpu_locks -- event/cpu_locks.sh@170 -- # run_test locking_app_on_locked_coremask locking_app_on_locked_coremask 00:05:57.403 23:39:45 event.cpu_locks -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:05:57.403 23:39:45 event.cpu_locks -- common/autotest_common.sh@1111 -- # xtrace_disable 00:05:57.403 23:39:45 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:05:57.403 ************************************ 00:05:57.403 START TEST locking_app_on_locked_coremask 00:05:57.403 ************************************ 00:05:57.403 23:39:45 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@1129 -- # locking_app_on_locked_coremask 00:05:57.403 23:39:45 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@115 -- # spdk_tgt_pid=70762 00:05:57.403 23:39:45 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@114 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 00:05:57.403 23:39:45 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@116 -- # waitforlisten 70762 /var/tmp/spdk.sock 00:05:57.403 23:39:45 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@835 -- # '[' -z 70762 ']' 00:05:57.403 23:39:45 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:05:57.403 23:39:45 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@840 -- # local max_retries=100 00:05:57.403 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:05:57.403 23:39:45 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:05:57.403 23:39:45 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@844 -- # xtrace_disable 00:05:57.403 23:39:45 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@10 -- # set +x 00:05:57.403 [2024-11-26 23:39:45.427575] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:05:57.403 [2024-11-26 23:39:45.427711] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid70762 ] 00:05:57.661 [2024-11-26 23:39:45.583838] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:05:57.661 [2024-11-26 23:39:45.608380] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:05:58.229 23:39:46 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:05:58.229 23:39:46 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@868 -- # return 0 00:05:58.229 23:39:46 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@118 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 -r /var/tmp/spdk2.sock 00:05:58.229 23:39:46 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@119 -- # spdk_tgt_pid2=70778 00:05:58.229 23:39:46 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@120 -- # NOT waitforlisten 70778 /var/tmp/spdk2.sock 00:05:58.229 23:39:46 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@652 -- # local es=0 00:05:58.229 23:39:46 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@654 -- # valid_exec_arg waitforlisten 70778 /var/tmp/spdk2.sock 00:05:58.229 23:39:46 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@640 -- # local arg=waitforlisten 00:05:58.229 23:39:46 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:05:58.229 23:39:46 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@644 -- # type -t waitforlisten 00:05:58.229 23:39:46 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:05:58.229 23:39:46 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@655 -- # waitforlisten 70778 /var/tmp/spdk2.sock 00:05:58.229 23:39:46 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@835 -- # '[' -z 70778 ']' 00:05:58.229 23:39:46 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk2.sock 00:05:58.229 23:39:46 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@840 -- # local max_retries=100 00:05:58.229 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock... 00:05:58.229 23:39:46 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock...' 00:05:58.229 23:39:46 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@844 -- # xtrace_disable 00:05:58.229 23:39:46 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@10 -- # set +x 00:05:58.229 [2024-11-26 23:39:46.301983] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:05:58.229 [2024-11-26 23:39:46.302130] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid70778 ] 00:05:58.487 [2024-11-26 23:39:46.451492] app.c: 781:claim_cpu_cores: *ERROR*: Cannot create lock on core 0, probably process 70762 has claimed it. 00:05:58.487 [2024-11-26 23:39:46.451554] app.c: 912:spdk_app_start: *ERROR*: Unable to acquire lock on assigned core mask - exiting. 00:05:59.056 ERROR: process (pid: 70778) is no longer running 00:05:59.056 /home/vagrant/spdk_repo/spdk/test/common/autotest_common.sh: line 850: kill: (70778) - No such process 00:05:59.056 23:39:46 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:05:59.056 23:39:46 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@868 -- # return 1 00:05:59.056 23:39:46 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@655 -- # es=1 00:05:59.056 23:39:46 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:05:59.056 23:39:46 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:05:59.056 23:39:46 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:05:59.056 23:39:46 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@122 -- # locks_exist 70762 00:05:59.056 23:39:46 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@22 -- # lslocks -p 70762 00:05:59.056 23:39:46 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@22 -- # grep -q spdk_cpu_lock 00:05:59.315 23:39:47 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@124 -- # killprocess 70762 00:05:59.315 23:39:47 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@954 -- # '[' -z 70762 ']' 00:05:59.315 23:39:47 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@958 -- # kill -0 70762 00:05:59.315 23:39:47 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@959 -- # uname 00:05:59.315 23:39:47 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:05:59.315 23:39:47 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 70762 00:05:59.315 23:39:47 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:05:59.315 23:39:47 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:05:59.315 killing process with pid 70762 00:05:59.315 23:39:47 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@972 -- # echo 'killing process with pid 70762' 00:05:59.315 23:39:47 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@973 -- # kill 70762 00:05:59.315 23:39:47 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@978 -- # wait 70762 00:05:59.885 00:05:59.885 real 0m2.373s 00:05:59.885 user 0m2.549s 00:05:59.885 sys 0m0.683s 00:05:59.885 23:39:47 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@1130 -- # xtrace_disable 00:05:59.885 23:39:47 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@10 -- # set +x 00:05:59.885 ************************************ 00:05:59.885 END TEST locking_app_on_locked_coremask 00:05:59.885 ************************************ 00:05:59.885 23:39:47 event.cpu_locks -- event/cpu_locks.sh@171 -- # run_test locking_overlapped_coremask locking_overlapped_coremask 00:05:59.885 23:39:47 event.cpu_locks -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:05:59.885 23:39:47 event.cpu_locks -- common/autotest_common.sh@1111 -- # xtrace_disable 00:05:59.885 23:39:47 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:05:59.885 ************************************ 00:05:59.885 START TEST locking_overlapped_coremask 00:05:59.885 ************************************ 00:05:59.885 23:39:47 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@1129 -- # locking_overlapped_coremask 00:05:59.885 23:39:47 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@132 -- # spdk_tgt_pid=70820 00:05:59.885 23:39:47 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@131 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x7 00:05:59.885 23:39:47 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@133 -- # waitforlisten 70820 /var/tmp/spdk.sock 00:05:59.885 23:39:47 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@835 -- # '[' -z 70820 ']' 00:05:59.885 23:39:47 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:05:59.885 23:39:47 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@840 -- # local max_retries=100 00:05:59.885 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:05:59.885 23:39:47 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:05:59.885 23:39:47 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@844 -- # xtrace_disable 00:05:59.885 23:39:47 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@10 -- # set +x 00:05:59.885 [2024-11-26 23:39:47.872709] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:05:59.885 [2024-11-26 23:39:47.872849] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x7 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid70820 ] 00:06:00.144 [2024-11-26 23:39:48.029064] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 3 00:06:00.144 [2024-11-26 23:39:48.056266] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:06:00.144 [2024-11-26 23:39:48.056372] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:00.144 [2024-11-26 23:39:48.056491] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:06:00.712 23:39:48 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:00.712 23:39:48 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@868 -- # return 0 00:06:00.712 23:39:48 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@135 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1c -r /var/tmp/spdk2.sock 00:06:00.712 23:39:48 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@136 -- # spdk_tgt_pid2=70838 00:06:00.712 23:39:48 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@137 -- # NOT waitforlisten 70838 /var/tmp/spdk2.sock 00:06:00.712 23:39:48 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@652 -- # local es=0 00:06:00.712 23:39:48 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@654 -- # valid_exec_arg waitforlisten 70838 /var/tmp/spdk2.sock 00:06:00.712 23:39:48 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@640 -- # local arg=waitforlisten 00:06:00.712 23:39:48 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:06:00.712 23:39:48 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@644 -- # type -t waitforlisten 00:06:00.712 23:39:48 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:06:00.712 23:39:48 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@655 -- # waitforlisten 70838 /var/tmp/spdk2.sock 00:06:00.712 23:39:48 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@835 -- # '[' -z 70838 ']' 00:06:00.712 23:39:48 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk2.sock 00:06:00.712 23:39:48 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:00.712 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock... 00:06:00.712 23:39:48 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock...' 00:06:00.712 23:39:48 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:00.712 23:39:48 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@10 -- # set +x 00:06:00.712 [2024-11-26 23:39:48.748601] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:06:00.712 [2024-11-26 23:39:48.748729] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1c --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid70838 ] 00:06:00.971 [2024-11-26 23:39:48.900295] app.c: 781:claim_cpu_cores: *ERROR*: Cannot create lock on core 2, probably process 70820 has claimed it. 00:06:00.972 [2024-11-26 23:39:48.900374] app.c: 912:spdk_app_start: *ERROR*: Unable to acquire lock on assigned core mask - exiting. 00:06:01.541 ERROR: process (pid: 70838) is no longer running 00:06:01.541 /home/vagrant/spdk_repo/spdk/test/common/autotest_common.sh: line 850: kill: (70838) - No such process 00:06:01.541 23:39:49 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:01.541 23:39:49 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@868 -- # return 1 00:06:01.541 23:39:49 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@655 -- # es=1 00:06:01.541 23:39:49 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:06:01.541 23:39:49 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:06:01.541 23:39:49 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:06:01.541 23:39:49 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@139 -- # check_remaining_locks 00:06:01.541 23:39:49 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@36 -- # locks=(/var/tmp/spdk_cpu_lock_*) 00:06:01.541 23:39:49 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@37 -- # locks_expected=(/var/tmp/spdk_cpu_lock_{000..002}) 00:06:01.541 23:39:49 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@38 -- # [[ /var/tmp/spdk_cpu_lock_000 /var/tmp/spdk_cpu_lock_001 /var/tmp/spdk_cpu_lock_002 == \/\v\a\r\/\t\m\p\/\s\p\d\k\_\c\p\u\_\l\o\c\k\_\0\0\0\ \/\v\a\r\/\t\m\p\/\s\p\d\k\_\c\p\u\_\l\o\c\k\_\0\0\1\ \/\v\a\r\/\t\m\p\/\s\p\d\k\_\c\p\u\_\l\o\c\k\_\0\0\2 ]] 00:06:01.541 23:39:49 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@141 -- # killprocess 70820 00:06:01.541 23:39:49 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@954 -- # '[' -z 70820 ']' 00:06:01.541 23:39:49 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@958 -- # kill -0 70820 00:06:01.541 23:39:49 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@959 -- # uname 00:06:01.541 23:39:49 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:06:01.541 23:39:49 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 70820 00:06:01.541 23:39:49 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:06:01.541 23:39:49 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:06:01.541 killing process with pid 70820 00:06:01.541 23:39:49 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@972 -- # echo 'killing process with pid 70820' 00:06:01.541 23:39:49 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@973 -- # kill 70820 00:06:01.541 23:39:49 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@978 -- # wait 70820 00:06:01.803 00:06:01.803 real 0m2.030s 00:06:01.803 user 0m5.484s 00:06:01.803 sys 0m0.490s 00:06:01.803 23:39:49 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:01.803 23:39:49 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@10 -- # set +x 00:06:01.803 ************************************ 00:06:01.803 END TEST locking_overlapped_coremask 00:06:01.803 ************************************ 00:06:01.803 23:39:49 event.cpu_locks -- event/cpu_locks.sh@172 -- # run_test locking_overlapped_coremask_via_rpc locking_overlapped_coremask_via_rpc 00:06:01.803 23:39:49 event.cpu_locks -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:06:01.803 23:39:49 event.cpu_locks -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:01.803 23:39:49 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:06:01.803 ************************************ 00:06:01.803 START TEST locking_overlapped_coremask_via_rpc 00:06:01.803 ************************************ 00:06:01.803 23:39:49 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@1129 -- # locking_overlapped_coremask_via_rpc 00:06:01.803 23:39:49 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@148 -- # spdk_tgt_pid=70880 00:06:01.803 23:39:49 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@149 -- # waitforlisten 70880 /var/tmp/spdk.sock 00:06:01.803 23:39:49 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@147 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x7 --disable-cpumask-locks 00:06:01.803 23:39:49 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@835 -- # '[' -z 70880 ']' 00:06:01.803 23:39:49 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:01.803 23:39:49 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:01.803 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:01.803 23:39:49 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:01.803 23:39:49 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:01.803 23:39:49 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:02.063 [2024-11-26 23:39:49.974825] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:06:02.063 [2024-11-26 23:39:49.974952] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x7 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid70880 ] 00:06:02.063 [2024-11-26 23:39:50.127634] app.c: 916:spdk_app_start: *NOTICE*: CPU core locks deactivated. 00:06:02.063 [2024-11-26 23:39:50.127683] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 3 00:06:02.063 [2024-11-26 23:39:50.154199] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:02.063 [2024-11-26 23:39:50.154144] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:06:02.063 [2024-11-26 23:39:50.154288] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:06:03.001 23:39:50 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:03.001 23:39:50 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@868 -- # return 0 00:06:03.001 23:39:50 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@151 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1c -r /var/tmp/spdk2.sock --disable-cpumask-locks 00:06:03.001 23:39:50 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@152 -- # spdk_tgt_pid2=70898 00:06:03.001 23:39:50 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@153 -- # waitforlisten 70898 /var/tmp/spdk2.sock 00:06:03.001 23:39:50 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@835 -- # '[' -z 70898 ']' 00:06:03.001 23:39:50 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk2.sock 00:06:03.001 23:39:50 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:03.001 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock... 00:06:03.001 23:39:50 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock...' 00:06:03.001 23:39:50 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:03.001 23:39:50 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:03.001 [2024-11-26 23:39:50.848465] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:06:03.001 [2024-11-26 23:39:50.848595] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1c --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid70898 ] 00:06:03.001 [2024-11-26 23:39:51.000839] app.c: 916:spdk_app_start: *NOTICE*: CPU core locks deactivated. 00:06:03.001 [2024-11-26 23:39:51.000891] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 3 00:06:03.001 [2024-11-26 23:39:51.061754] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:06:03.001 [2024-11-26 23:39:51.061850] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 4 00:06:03.001 [2024-11-26 23:39:51.061781] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:06:03.940 23:39:51 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:03.940 23:39:51 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@868 -- # return 0 00:06:03.940 23:39:51 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@155 -- # rpc_cmd framework_enable_cpumask_locks 00:06:03.940 23:39:51 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:03.940 23:39:51 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:03.940 23:39:51 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:03.940 23:39:51 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@156 -- # NOT rpc_cmd -s /var/tmp/spdk2.sock framework_enable_cpumask_locks 00:06:03.940 23:39:51 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@652 -- # local es=0 00:06:03.940 23:39:51 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd -s /var/tmp/spdk2.sock framework_enable_cpumask_locks 00:06:03.940 23:39:51 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:06:03.940 23:39:51 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:06:03.940 23:39:51 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:06:03.940 23:39:51 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:06:03.940 23:39:51 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@655 -- # rpc_cmd -s /var/tmp/spdk2.sock framework_enable_cpumask_locks 00:06:03.940 23:39:51 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:03.940 23:39:51 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:03.940 [2024-11-26 23:39:51.728514] app.c: 781:claim_cpu_cores: *ERROR*: Cannot create lock on core 2, probably process 70880 has claimed it. 00:06:03.940 request: 00:06:03.940 { 00:06:03.940 "method": "framework_enable_cpumask_locks", 00:06:03.940 "req_id": 1 00:06:03.940 } 00:06:03.940 Got JSON-RPC error response 00:06:03.940 response: 00:06:03.940 { 00:06:03.940 "code": -32603, 00:06:03.940 "message": "Failed to claim CPU core: 2" 00:06:03.940 } 00:06:03.940 23:39:51 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:06:03.940 23:39:51 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@655 -- # es=1 00:06:03.940 23:39:51 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:06:03.940 23:39:51 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:06:03.940 23:39:51 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:06:03.940 23:39:51 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@158 -- # waitforlisten 70880 /var/tmp/spdk.sock 00:06:03.941 23:39:51 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@835 -- # '[' -z 70880 ']' 00:06:03.941 23:39:51 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:03.941 23:39:51 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:03.941 23:39:51 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:03.941 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:03.941 23:39:51 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:03.941 23:39:51 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:03.941 23:39:51 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:03.941 23:39:51 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@868 -- # return 0 00:06:03.941 23:39:51 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@159 -- # waitforlisten 70898 /var/tmp/spdk2.sock 00:06:03.941 23:39:51 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@835 -- # '[' -z 70898 ']' 00:06:03.941 23:39:51 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk2.sock 00:06:03.941 23:39:51 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:03.941 23:39:51 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock...' 00:06:03.941 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock... 00:06:03.941 23:39:51 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:03.941 23:39:51 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:04.200 23:39:52 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:04.200 23:39:52 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@868 -- # return 0 00:06:04.200 23:39:52 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@161 -- # check_remaining_locks 00:06:04.200 23:39:52 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@36 -- # locks=(/var/tmp/spdk_cpu_lock_*) 00:06:04.201 23:39:52 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@37 -- # locks_expected=(/var/tmp/spdk_cpu_lock_{000..002}) 00:06:04.201 23:39:52 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@38 -- # [[ /var/tmp/spdk_cpu_lock_000 /var/tmp/spdk_cpu_lock_001 /var/tmp/spdk_cpu_lock_002 == \/\v\a\r\/\t\m\p\/\s\p\d\k\_\c\p\u\_\l\o\c\k\_\0\0\0\ \/\v\a\r\/\t\m\p\/\s\p\d\k\_\c\p\u\_\l\o\c\k\_\0\0\1\ \/\v\a\r\/\t\m\p\/\s\p\d\k\_\c\p\u\_\l\o\c\k\_\0\0\2 ]] 00:06:04.201 00:06:04.201 real 0m2.316s 00:06:04.201 user 0m1.088s 00:06:04.201 sys 0m0.157s 00:06:04.201 23:39:52 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:04.201 23:39:52 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:04.201 ************************************ 00:06:04.201 END TEST locking_overlapped_coremask_via_rpc 00:06:04.201 ************************************ 00:06:04.201 23:39:52 event.cpu_locks -- event/cpu_locks.sh@174 -- # cleanup 00:06:04.201 23:39:52 event.cpu_locks -- event/cpu_locks.sh@15 -- # [[ -z 70880 ]] 00:06:04.201 23:39:52 event.cpu_locks -- event/cpu_locks.sh@15 -- # killprocess 70880 00:06:04.201 23:39:52 event.cpu_locks -- common/autotest_common.sh@954 -- # '[' -z 70880 ']' 00:06:04.201 23:39:52 event.cpu_locks -- common/autotest_common.sh@958 -- # kill -0 70880 00:06:04.201 23:39:52 event.cpu_locks -- common/autotest_common.sh@959 -- # uname 00:06:04.201 23:39:52 event.cpu_locks -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:06:04.201 23:39:52 event.cpu_locks -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 70880 00:06:04.201 23:39:52 event.cpu_locks -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:06:04.201 23:39:52 event.cpu_locks -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:06:04.201 killing process with pid 70880 00:06:04.201 23:39:52 event.cpu_locks -- common/autotest_common.sh@972 -- # echo 'killing process with pid 70880' 00:06:04.201 23:39:52 event.cpu_locks -- common/autotest_common.sh@973 -- # kill 70880 00:06:04.201 23:39:52 event.cpu_locks -- common/autotest_common.sh@978 -- # wait 70880 00:06:04.780 23:39:52 event.cpu_locks -- event/cpu_locks.sh@16 -- # [[ -z 70898 ]] 00:06:04.780 23:39:52 event.cpu_locks -- event/cpu_locks.sh@16 -- # killprocess 70898 00:06:04.780 23:39:52 event.cpu_locks -- common/autotest_common.sh@954 -- # '[' -z 70898 ']' 00:06:04.780 23:39:52 event.cpu_locks -- common/autotest_common.sh@958 -- # kill -0 70898 00:06:04.780 23:39:52 event.cpu_locks -- common/autotest_common.sh@959 -- # uname 00:06:04.780 23:39:52 event.cpu_locks -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:06:04.780 23:39:52 event.cpu_locks -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 70898 00:06:04.780 23:39:52 event.cpu_locks -- common/autotest_common.sh@960 -- # process_name=reactor_2 00:06:04.780 23:39:52 event.cpu_locks -- common/autotest_common.sh@964 -- # '[' reactor_2 = sudo ']' 00:06:04.780 killing process with pid 70898 00:06:04.780 23:39:52 event.cpu_locks -- common/autotest_common.sh@972 -- # echo 'killing process with pid 70898' 00:06:04.780 23:39:52 event.cpu_locks -- common/autotest_common.sh@973 -- # kill 70898 00:06:04.780 23:39:52 event.cpu_locks -- common/autotest_common.sh@978 -- # wait 70898 00:06:05.042 23:39:53 event.cpu_locks -- event/cpu_locks.sh@18 -- # rm -f 00:06:05.042 23:39:53 event.cpu_locks -- event/cpu_locks.sh@1 -- # cleanup 00:06:05.042 23:39:53 event.cpu_locks -- event/cpu_locks.sh@15 -- # [[ -z 70880 ]] 00:06:05.042 23:39:53 event.cpu_locks -- event/cpu_locks.sh@15 -- # killprocess 70880 00:06:05.042 23:39:53 event.cpu_locks -- common/autotest_common.sh@954 -- # '[' -z 70880 ']' 00:06:05.042 23:39:53 event.cpu_locks -- common/autotest_common.sh@958 -- # kill -0 70880 00:06:05.042 /home/vagrant/spdk_repo/spdk/test/common/autotest_common.sh: line 958: kill: (70880) - No such process 00:06:05.042 Process with pid 70880 is not found 00:06:05.042 23:39:53 event.cpu_locks -- common/autotest_common.sh@981 -- # echo 'Process with pid 70880 is not found' 00:06:05.042 23:39:53 event.cpu_locks -- event/cpu_locks.sh@16 -- # [[ -z 70898 ]] 00:06:05.042 23:39:53 event.cpu_locks -- event/cpu_locks.sh@16 -- # killprocess 70898 00:06:05.042 23:39:53 event.cpu_locks -- common/autotest_common.sh@954 -- # '[' -z 70898 ']' 00:06:05.042 23:39:53 event.cpu_locks -- common/autotest_common.sh@958 -- # kill -0 70898 00:06:05.042 /home/vagrant/spdk_repo/spdk/test/common/autotest_common.sh: line 958: kill: (70898) - No such process 00:06:05.042 Process with pid 70898 is not found 00:06:05.042 23:39:53 event.cpu_locks -- common/autotest_common.sh@981 -- # echo 'Process with pid 70898 is not found' 00:06:05.042 23:39:53 event.cpu_locks -- event/cpu_locks.sh@18 -- # rm -f 00:06:05.042 00:06:05.042 real 0m18.171s 00:06:05.042 user 0m30.931s 00:06:05.042 sys 0m5.430s 00:06:05.042 23:39:53 event.cpu_locks -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:05.042 23:39:53 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:06:05.042 ************************************ 00:06:05.042 END TEST cpu_locks 00:06:05.042 ************************************ 00:06:05.042 00:06:05.042 real 0m45.958s 00:06:05.042 user 1m28.699s 00:06:05.042 sys 0m9.214s 00:06:05.042 23:39:53 event -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:05.042 23:39:53 event -- common/autotest_common.sh@10 -- # set +x 00:06:05.042 ************************************ 00:06:05.042 END TEST event 00:06:05.042 ************************************ 00:06:05.302 23:39:53 -- spdk/autotest.sh@169 -- # run_test thread /home/vagrant/spdk_repo/spdk/test/thread/thread.sh 00:06:05.302 23:39:53 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:06:05.302 23:39:53 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:05.302 23:39:53 -- common/autotest_common.sh@10 -- # set +x 00:06:05.302 ************************************ 00:06:05.302 START TEST thread 00:06:05.302 ************************************ 00:06:05.302 23:39:53 thread -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/thread/thread.sh 00:06:05.302 * Looking for test storage... 00:06:05.302 * Found test storage at /home/vagrant/spdk_repo/spdk/test/thread 00:06:05.302 23:39:53 thread -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:06:05.302 23:39:53 thread -- common/autotest_common.sh@1693 -- # lcov --version 00:06:05.302 23:39:53 thread -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:06:05.302 23:39:53 thread -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:06:05.302 23:39:53 thread -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:06:05.302 23:39:53 thread -- scripts/common.sh@333 -- # local ver1 ver1_l 00:06:05.302 23:39:53 thread -- scripts/common.sh@334 -- # local ver2 ver2_l 00:06:05.302 23:39:53 thread -- scripts/common.sh@336 -- # IFS=.-: 00:06:05.302 23:39:53 thread -- scripts/common.sh@336 -- # read -ra ver1 00:06:05.302 23:39:53 thread -- scripts/common.sh@337 -- # IFS=.-: 00:06:05.302 23:39:53 thread -- scripts/common.sh@337 -- # read -ra ver2 00:06:05.302 23:39:53 thread -- scripts/common.sh@338 -- # local 'op=<' 00:06:05.302 23:39:53 thread -- scripts/common.sh@340 -- # ver1_l=2 00:06:05.302 23:39:53 thread -- scripts/common.sh@341 -- # ver2_l=1 00:06:05.302 23:39:53 thread -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:06:05.302 23:39:53 thread -- scripts/common.sh@344 -- # case "$op" in 00:06:05.302 23:39:53 thread -- scripts/common.sh@345 -- # : 1 00:06:05.302 23:39:53 thread -- scripts/common.sh@364 -- # (( v = 0 )) 00:06:05.302 23:39:53 thread -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:06:05.302 23:39:53 thread -- scripts/common.sh@365 -- # decimal 1 00:06:05.302 23:39:53 thread -- scripts/common.sh@353 -- # local d=1 00:06:05.302 23:39:53 thread -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:06:05.302 23:39:53 thread -- scripts/common.sh@355 -- # echo 1 00:06:05.302 23:39:53 thread -- scripts/common.sh@365 -- # ver1[v]=1 00:06:05.302 23:39:53 thread -- scripts/common.sh@366 -- # decimal 2 00:06:05.302 23:39:53 thread -- scripts/common.sh@353 -- # local d=2 00:06:05.302 23:39:53 thread -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:06:05.302 23:39:53 thread -- scripts/common.sh@355 -- # echo 2 00:06:05.302 23:39:53 thread -- scripts/common.sh@366 -- # ver2[v]=2 00:06:05.302 23:39:53 thread -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:06:05.302 23:39:53 thread -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:06:05.302 23:39:53 thread -- scripts/common.sh@368 -- # return 0 00:06:05.302 23:39:53 thread -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:06:05.302 23:39:53 thread -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:06:05.302 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:05.302 --rc genhtml_branch_coverage=1 00:06:05.302 --rc genhtml_function_coverage=1 00:06:05.302 --rc genhtml_legend=1 00:06:05.302 --rc geninfo_all_blocks=1 00:06:05.302 --rc geninfo_unexecuted_blocks=1 00:06:05.302 00:06:05.302 ' 00:06:05.302 23:39:53 thread -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:06:05.302 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:05.302 --rc genhtml_branch_coverage=1 00:06:05.302 --rc genhtml_function_coverage=1 00:06:05.302 --rc genhtml_legend=1 00:06:05.302 --rc geninfo_all_blocks=1 00:06:05.302 --rc geninfo_unexecuted_blocks=1 00:06:05.302 00:06:05.302 ' 00:06:05.302 23:39:53 thread -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:06:05.302 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:05.302 --rc genhtml_branch_coverage=1 00:06:05.302 --rc genhtml_function_coverage=1 00:06:05.302 --rc genhtml_legend=1 00:06:05.302 --rc geninfo_all_blocks=1 00:06:05.302 --rc geninfo_unexecuted_blocks=1 00:06:05.302 00:06:05.302 ' 00:06:05.302 23:39:53 thread -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:06:05.302 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:05.302 --rc genhtml_branch_coverage=1 00:06:05.302 --rc genhtml_function_coverage=1 00:06:05.302 --rc genhtml_legend=1 00:06:05.302 --rc geninfo_all_blocks=1 00:06:05.302 --rc geninfo_unexecuted_blocks=1 00:06:05.302 00:06:05.302 ' 00:06:05.302 23:39:53 thread -- thread/thread.sh@11 -- # run_test thread_poller_perf /home/vagrant/spdk_repo/spdk/test/thread/poller_perf/poller_perf -b 1000 -l 1 -t 1 00:06:05.302 23:39:53 thread -- common/autotest_common.sh@1105 -- # '[' 8 -le 1 ']' 00:06:05.302 23:39:53 thread -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:05.302 23:39:53 thread -- common/autotest_common.sh@10 -- # set +x 00:06:05.568 ************************************ 00:06:05.568 START TEST thread_poller_perf 00:06:05.568 ************************************ 00:06:05.568 23:39:53 thread.thread_poller_perf -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/thread/poller_perf/poller_perf -b 1000 -l 1 -t 1 00:06:05.568 [2024-11-26 23:39:53.475775] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:06:05.568 [2024-11-26 23:39:53.476388] [ DPDK EAL parameters: poller_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid71036 ] 00:06:05.568 [2024-11-26 23:39:53.629267] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:05.568 Running 1000 pollers for 1 seconds with 1 microseconds period. 00:06:05.568 [2024-11-26 23:39:53.653461] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:06.950 [2024-11-26T23:39:55.082Z] ====================================== 00:06:06.950 [2024-11-26T23:39:55.082Z] busy:2297627696 (cyc) 00:06:06.950 [2024-11-26T23:39:55.082Z] total_run_count: 418000 00:06:06.950 [2024-11-26T23:39:55.082Z] tsc_hz: 2290000000 (cyc) 00:06:06.950 [2024-11-26T23:39:55.082Z] ====================================== 00:06:06.950 [2024-11-26T23:39:55.082Z] poller_cost: 5496 (cyc), 2400 (nsec) 00:06:06.950 00:06:06.950 real 0m1.285s 00:06:06.950 user 0m1.106s 00:06:06.950 sys 0m0.073s 00:06:06.950 23:39:54 thread.thread_poller_perf -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:06.950 23:39:54 thread.thread_poller_perf -- common/autotest_common.sh@10 -- # set +x 00:06:06.950 ************************************ 00:06:06.950 END TEST thread_poller_perf 00:06:06.950 ************************************ 00:06:06.950 23:39:54 thread -- thread/thread.sh@12 -- # run_test thread_poller_perf /home/vagrant/spdk_repo/spdk/test/thread/poller_perf/poller_perf -b 1000 -l 0 -t 1 00:06:06.950 23:39:54 thread -- common/autotest_common.sh@1105 -- # '[' 8 -le 1 ']' 00:06:06.950 23:39:54 thread -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:06.950 23:39:54 thread -- common/autotest_common.sh@10 -- # set +x 00:06:06.950 ************************************ 00:06:06.950 START TEST thread_poller_perf 00:06:06.950 ************************************ 00:06:06.950 23:39:54 thread.thread_poller_perf -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/thread/poller_perf/poller_perf -b 1000 -l 0 -t 1 00:06:06.950 [2024-11-26 23:39:54.827817] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:06:06.950 [2024-11-26 23:39:54.827959] [ DPDK EAL parameters: poller_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid71067 ] 00:06:06.950 [2024-11-26 23:39:54.980943] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:06.950 Running 1000 pollers for 1 seconds with 0 microseconds period. 00:06:06.950 [2024-11-26 23:39:55.006578] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:08.328 [2024-11-26T23:39:56.460Z] ====================================== 00:06:08.328 [2024-11-26T23:39:56.460Z] busy:2293542678 (cyc) 00:06:08.328 [2024-11-26T23:39:56.460Z] total_run_count: 5497000 00:06:08.328 [2024-11-26T23:39:56.460Z] tsc_hz: 2290000000 (cyc) 00:06:08.328 [2024-11-26T23:39:56.460Z] ====================================== 00:06:08.328 [2024-11-26T23:39:56.460Z] poller_cost: 417 (cyc), 182 (nsec) 00:06:08.328 00:06:08.328 real 0m1.282s 00:06:08.328 user 0m1.108s 00:06:08.328 sys 0m0.068s 00:06:08.328 23:39:56 thread.thread_poller_perf -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:08.328 23:39:56 thread.thread_poller_perf -- common/autotest_common.sh@10 -- # set +x 00:06:08.328 ************************************ 00:06:08.328 END TEST thread_poller_perf 00:06:08.328 ************************************ 00:06:08.328 23:39:56 thread -- thread/thread.sh@17 -- # [[ y != \y ]] 00:06:08.328 00:06:08.328 real 0m2.923s 00:06:08.328 user 0m2.380s 00:06:08.328 sys 0m0.346s 00:06:08.328 23:39:56 thread -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:08.328 23:39:56 thread -- common/autotest_common.sh@10 -- # set +x 00:06:08.328 ************************************ 00:06:08.328 END TEST thread 00:06:08.328 ************************************ 00:06:08.328 23:39:56 -- spdk/autotest.sh@171 -- # [[ 0 -eq 1 ]] 00:06:08.328 23:39:56 -- spdk/autotest.sh@176 -- # run_test app_cmdline /home/vagrant/spdk_repo/spdk/test/app/cmdline.sh 00:06:08.328 23:39:56 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:06:08.328 23:39:56 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:08.328 23:39:56 -- common/autotest_common.sh@10 -- # set +x 00:06:08.328 ************************************ 00:06:08.328 START TEST app_cmdline 00:06:08.328 ************************************ 00:06:08.328 23:39:56 app_cmdline -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/app/cmdline.sh 00:06:08.328 * Looking for test storage... 00:06:08.328 * Found test storage at /home/vagrant/spdk_repo/spdk/test/app 00:06:08.328 23:39:56 app_cmdline -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:06:08.328 23:39:56 app_cmdline -- common/autotest_common.sh@1693 -- # lcov --version 00:06:08.328 23:39:56 app_cmdline -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:06:08.328 23:39:56 app_cmdline -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:06:08.328 23:39:56 app_cmdline -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:06:08.328 23:39:56 app_cmdline -- scripts/common.sh@333 -- # local ver1 ver1_l 00:06:08.328 23:39:56 app_cmdline -- scripts/common.sh@334 -- # local ver2 ver2_l 00:06:08.328 23:39:56 app_cmdline -- scripts/common.sh@336 -- # IFS=.-: 00:06:08.328 23:39:56 app_cmdline -- scripts/common.sh@336 -- # read -ra ver1 00:06:08.328 23:39:56 app_cmdline -- scripts/common.sh@337 -- # IFS=.-: 00:06:08.328 23:39:56 app_cmdline -- scripts/common.sh@337 -- # read -ra ver2 00:06:08.328 23:39:56 app_cmdline -- scripts/common.sh@338 -- # local 'op=<' 00:06:08.328 23:39:56 app_cmdline -- scripts/common.sh@340 -- # ver1_l=2 00:06:08.328 23:39:56 app_cmdline -- scripts/common.sh@341 -- # ver2_l=1 00:06:08.328 23:39:56 app_cmdline -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:06:08.328 23:39:56 app_cmdline -- scripts/common.sh@344 -- # case "$op" in 00:06:08.328 23:39:56 app_cmdline -- scripts/common.sh@345 -- # : 1 00:06:08.328 23:39:56 app_cmdline -- scripts/common.sh@364 -- # (( v = 0 )) 00:06:08.328 23:39:56 app_cmdline -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:06:08.328 23:39:56 app_cmdline -- scripts/common.sh@365 -- # decimal 1 00:06:08.328 23:39:56 app_cmdline -- scripts/common.sh@353 -- # local d=1 00:06:08.328 23:39:56 app_cmdline -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:06:08.328 23:39:56 app_cmdline -- scripts/common.sh@355 -- # echo 1 00:06:08.328 23:39:56 app_cmdline -- scripts/common.sh@365 -- # ver1[v]=1 00:06:08.328 23:39:56 app_cmdline -- scripts/common.sh@366 -- # decimal 2 00:06:08.328 23:39:56 app_cmdline -- scripts/common.sh@353 -- # local d=2 00:06:08.328 23:39:56 app_cmdline -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:06:08.328 23:39:56 app_cmdline -- scripts/common.sh@355 -- # echo 2 00:06:08.328 23:39:56 app_cmdline -- scripts/common.sh@366 -- # ver2[v]=2 00:06:08.328 23:39:56 app_cmdline -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:06:08.328 23:39:56 app_cmdline -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:06:08.328 23:39:56 app_cmdline -- scripts/common.sh@368 -- # return 0 00:06:08.328 23:39:56 app_cmdline -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:06:08.328 23:39:56 app_cmdline -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:06:08.328 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:08.328 --rc genhtml_branch_coverage=1 00:06:08.328 --rc genhtml_function_coverage=1 00:06:08.328 --rc genhtml_legend=1 00:06:08.328 --rc geninfo_all_blocks=1 00:06:08.328 --rc geninfo_unexecuted_blocks=1 00:06:08.328 00:06:08.328 ' 00:06:08.328 23:39:56 app_cmdline -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:06:08.328 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:08.328 --rc genhtml_branch_coverage=1 00:06:08.328 --rc genhtml_function_coverage=1 00:06:08.328 --rc genhtml_legend=1 00:06:08.328 --rc geninfo_all_blocks=1 00:06:08.328 --rc geninfo_unexecuted_blocks=1 00:06:08.328 00:06:08.328 ' 00:06:08.328 23:39:56 app_cmdline -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:06:08.328 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:08.328 --rc genhtml_branch_coverage=1 00:06:08.328 --rc genhtml_function_coverage=1 00:06:08.328 --rc genhtml_legend=1 00:06:08.328 --rc geninfo_all_blocks=1 00:06:08.328 --rc geninfo_unexecuted_blocks=1 00:06:08.328 00:06:08.328 ' 00:06:08.328 23:39:56 app_cmdline -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:06:08.328 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:08.328 --rc genhtml_branch_coverage=1 00:06:08.328 --rc genhtml_function_coverage=1 00:06:08.328 --rc genhtml_legend=1 00:06:08.328 --rc geninfo_all_blocks=1 00:06:08.328 --rc geninfo_unexecuted_blocks=1 00:06:08.328 00:06:08.328 ' 00:06:08.328 23:39:56 app_cmdline -- app/cmdline.sh@14 -- # trap 'killprocess $spdk_tgt_pid' EXIT 00:06:08.328 23:39:56 app_cmdline -- app/cmdline.sh@17 -- # spdk_tgt_pid=71151 00:06:08.328 23:39:56 app_cmdline -- app/cmdline.sh@16 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt --rpcs-allowed spdk_get_version,rpc_get_methods 00:06:08.328 23:39:56 app_cmdline -- app/cmdline.sh@18 -- # waitforlisten 71151 00:06:08.328 23:39:56 app_cmdline -- common/autotest_common.sh@835 -- # '[' -z 71151 ']' 00:06:08.328 23:39:56 app_cmdline -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:08.328 23:39:56 app_cmdline -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:08.328 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:08.328 23:39:56 app_cmdline -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:08.328 23:39:56 app_cmdline -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:08.328 23:39:56 app_cmdline -- common/autotest_common.sh@10 -- # set +x 00:06:08.587 [2024-11-26 23:39:56.504541] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:06:08.587 [2024-11-26 23:39:56.504661] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid71151 ] 00:06:08.587 [2024-11-26 23:39:56.658178] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:08.587 [2024-11-26 23:39:56.682201] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:09.524 23:39:57 app_cmdline -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:09.524 23:39:57 app_cmdline -- common/autotest_common.sh@868 -- # return 0 00:06:09.524 23:39:57 app_cmdline -- app/cmdline.sh@20 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py spdk_get_version 00:06:09.524 { 00:06:09.524 "version": "SPDK v25.01-pre git sha1 2f2acf4eb", 00:06:09.524 "fields": { 00:06:09.524 "major": 25, 00:06:09.524 "minor": 1, 00:06:09.524 "patch": 0, 00:06:09.524 "suffix": "-pre", 00:06:09.524 "commit": "2f2acf4eb" 00:06:09.524 } 00:06:09.524 } 00:06:09.524 23:39:57 app_cmdline -- app/cmdline.sh@22 -- # expected_methods=() 00:06:09.524 23:39:57 app_cmdline -- app/cmdline.sh@23 -- # expected_methods+=("rpc_get_methods") 00:06:09.524 23:39:57 app_cmdline -- app/cmdline.sh@24 -- # expected_methods+=("spdk_get_version") 00:06:09.524 23:39:57 app_cmdline -- app/cmdline.sh@26 -- # methods=($(rpc_cmd rpc_get_methods | jq -r ".[]" | sort)) 00:06:09.524 23:39:57 app_cmdline -- app/cmdline.sh@26 -- # rpc_cmd rpc_get_methods 00:06:09.525 23:39:57 app_cmdline -- app/cmdline.sh@26 -- # jq -r '.[]' 00:06:09.525 23:39:57 app_cmdline -- app/cmdline.sh@26 -- # sort 00:06:09.525 23:39:57 app_cmdline -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:09.525 23:39:57 app_cmdline -- common/autotest_common.sh@10 -- # set +x 00:06:09.525 23:39:57 app_cmdline -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:09.525 23:39:57 app_cmdline -- app/cmdline.sh@27 -- # (( 2 == 2 )) 00:06:09.525 23:39:57 app_cmdline -- app/cmdline.sh@28 -- # [[ rpc_get_methods spdk_get_version == \r\p\c\_\g\e\t\_\m\e\t\h\o\d\s\ \s\p\d\k\_\g\e\t\_\v\e\r\s\i\o\n ]] 00:06:09.525 23:39:57 app_cmdline -- app/cmdline.sh@30 -- # NOT /home/vagrant/spdk_repo/spdk/scripts/rpc.py env_dpdk_get_mem_stats 00:06:09.525 23:39:57 app_cmdline -- common/autotest_common.sh@652 -- # local es=0 00:06:09.525 23:39:57 app_cmdline -- common/autotest_common.sh@654 -- # valid_exec_arg /home/vagrant/spdk_repo/spdk/scripts/rpc.py env_dpdk_get_mem_stats 00:06:09.525 23:39:57 app_cmdline -- common/autotest_common.sh@640 -- # local arg=/home/vagrant/spdk_repo/spdk/scripts/rpc.py 00:06:09.525 23:39:57 app_cmdline -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:06:09.525 23:39:57 app_cmdline -- common/autotest_common.sh@644 -- # type -t /home/vagrant/spdk_repo/spdk/scripts/rpc.py 00:06:09.525 23:39:57 app_cmdline -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:06:09.525 23:39:57 app_cmdline -- common/autotest_common.sh@646 -- # type -P /home/vagrant/spdk_repo/spdk/scripts/rpc.py 00:06:09.525 23:39:57 app_cmdline -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:06:09.525 23:39:57 app_cmdline -- common/autotest_common.sh@646 -- # arg=/home/vagrant/spdk_repo/spdk/scripts/rpc.py 00:06:09.525 23:39:57 app_cmdline -- common/autotest_common.sh@646 -- # [[ -x /home/vagrant/spdk_repo/spdk/scripts/rpc.py ]] 00:06:09.525 23:39:57 app_cmdline -- common/autotest_common.sh@655 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py env_dpdk_get_mem_stats 00:06:09.783 request: 00:06:09.783 { 00:06:09.783 "method": "env_dpdk_get_mem_stats", 00:06:09.783 "req_id": 1 00:06:09.783 } 00:06:09.783 Got JSON-RPC error response 00:06:09.783 response: 00:06:09.783 { 00:06:09.783 "code": -32601, 00:06:09.783 "message": "Method not found" 00:06:09.783 } 00:06:09.783 23:39:57 app_cmdline -- common/autotest_common.sh@655 -- # es=1 00:06:09.783 23:39:57 app_cmdline -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:06:09.783 23:39:57 app_cmdline -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:06:09.783 23:39:57 app_cmdline -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:06:09.783 23:39:57 app_cmdline -- app/cmdline.sh@1 -- # killprocess 71151 00:06:09.783 23:39:57 app_cmdline -- common/autotest_common.sh@954 -- # '[' -z 71151 ']' 00:06:09.783 23:39:57 app_cmdline -- common/autotest_common.sh@958 -- # kill -0 71151 00:06:09.783 23:39:57 app_cmdline -- common/autotest_common.sh@959 -- # uname 00:06:09.783 23:39:57 app_cmdline -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:06:09.783 23:39:57 app_cmdline -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 71151 00:06:09.783 23:39:57 app_cmdline -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:06:09.784 23:39:57 app_cmdline -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:06:09.784 killing process with pid 71151 00:06:09.784 23:39:57 app_cmdline -- common/autotest_common.sh@972 -- # echo 'killing process with pid 71151' 00:06:09.784 23:39:57 app_cmdline -- common/autotest_common.sh@973 -- # kill 71151 00:06:09.784 23:39:57 app_cmdline -- common/autotest_common.sh@978 -- # wait 71151 00:06:10.043 00:06:10.043 real 0m1.951s 00:06:10.043 user 0m2.191s 00:06:10.043 sys 0m0.516s 00:06:10.043 23:39:58 app_cmdline -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:10.043 23:39:58 app_cmdline -- common/autotest_common.sh@10 -- # set +x 00:06:10.043 ************************************ 00:06:10.043 END TEST app_cmdline 00:06:10.043 ************************************ 00:06:10.301 23:39:58 -- spdk/autotest.sh@177 -- # run_test version /home/vagrant/spdk_repo/spdk/test/app/version.sh 00:06:10.301 23:39:58 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:06:10.301 23:39:58 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:10.301 23:39:58 -- common/autotest_common.sh@10 -- # set +x 00:06:10.301 ************************************ 00:06:10.301 START TEST version 00:06:10.301 ************************************ 00:06:10.301 23:39:58 version -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/app/version.sh 00:06:10.301 * Looking for test storage... 00:06:10.301 * Found test storage at /home/vagrant/spdk_repo/spdk/test/app 00:06:10.301 23:39:58 version -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:06:10.301 23:39:58 version -- common/autotest_common.sh@1693 -- # lcov --version 00:06:10.301 23:39:58 version -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:06:10.301 23:39:58 version -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:06:10.301 23:39:58 version -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:06:10.301 23:39:58 version -- scripts/common.sh@333 -- # local ver1 ver1_l 00:06:10.301 23:39:58 version -- scripts/common.sh@334 -- # local ver2 ver2_l 00:06:10.301 23:39:58 version -- scripts/common.sh@336 -- # IFS=.-: 00:06:10.301 23:39:58 version -- scripts/common.sh@336 -- # read -ra ver1 00:06:10.301 23:39:58 version -- scripts/common.sh@337 -- # IFS=.-: 00:06:10.302 23:39:58 version -- scripts/common.sh@337 -- # read -ra ver2 00:06:10.302 23:39:58 version -- scripts/common.sh@338 -- # local 'op=<' 00:06:10.302 23:39:58 version -- scripts/common.sh@340 -- # ver1_l=2 00:06:10.302 23:39:58 version -- scripts/common.sh@341 -- # ver2_l=1 00:06:10.302 23:39:58 version -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:06:10.302 23:39:58 version -- scripts/common.sh@344 -- # case "$op" in 00:06:10.302 23:39:58 version -- scripts/common.sh@345 -- # : 1 00:06:10.302 23:39:58 version -- scripts/common.sh@364 -- # (( v = 0 )) 00:06:10.302 23:39:58 version -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:06:10.302 23:39:58 version -- scripts/common.sh@365 -- # decimal 1 00:06:10.302 23:39:58 version -- scripts/common.sh@353 -- # local d=1 00:06:10.302 23:39:58 version -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:06:10.302 23:39:58 version -- scripts/common.sh@355 -- # echo 1 00:06:10.302 23:39:58 version -- scripts/common.sh@365 -- # ver1[v]=1 00:06:10.302 23:39:58 version -- scripts/common.sh@366 -- # decimal 2 00:06:10.302 23:39:58 version -- scripts/common.sh@353 -- # local d=2 00:06:10.302 23:39:58 version -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:06:10.302 23:39:58 version -- scripts/common.sh@355 -- # echo 2 00:06:10.302 23:39:58 version -- scripts/common.sh@366 -- # ver2[v]=2 00:06:10.302 23:39:58 version -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:06:10.302 23:39:58 version -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:06:10.302 23:39:58 version -- scripts/common.sh@368 -- # return 0 00:06:10.302 23:39:58 version -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:06:10.302 23:39:58 version -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:06:10.302 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:10.302 --rc genhtml_branch_coverage=1 00:06:10.302 --rc genhtml_function_coverage=1 00:06:10.302 --rc genhtml_legend=1 00:06:10.302 --rc geninfo_all_blocks=1 00:06:10.302 --rc geninfo_unexecuted_blocks=1 00:06:10.302 00:06:10.302 ' 00:06:10.561 23:39:58 version -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:06:10.561 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:10.561 --rc genhtml_branch_coverage=1 00:06:10.561 --rc genhtml_function_coverage=1 00:06:10.561 --rc genhtml_legend=1 00:06:10.561 --rc geninfo_all_blocks=1 00:06:10.561 --rc geninfo_unexecuted_blocks=1 00:06:10.561 00:06:10.561 ' 00:06:10.561 23:39:58 version -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:06:10.561 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:10.561 --rc genhtml_branch_coverage=1 00:06:10.561 --rc genhtml_function_coverage=1 00:06:10.561 --rc genhtml_legend=1 00:06:10.561 --rc geninfo_all_blocks=1 00:06:10.561 --rc geninfo_unexecuted_blocks=1 00:06:10.561 00:06:10.561 ' 00:06:10.561 23:39:58 version -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:06:10.561 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:10.561 --rc genhtml_branch_coverage=1 00:06:10.561 --rc genhtml_function_coverage=1 00:06:10.561 --rc genhtml_legend=1 00:06:10.561 --rc geninfo_all_blocks=1 00:06:10.561 --rc geninfo_unexecuted_blocks=1 00:06:10.561 00:06:10.561 ' 00:06:10.561 23:39:58 version -- app/version.sh@17 -- # get_header_version major 00:06:10.561 23:39:58 version -- app/version.sh@13 -- # grep -E '^#define SPDK_VERSION_MAJOR[[:space:]]+' /home/vagrant/spdk_repo/spdk/include/spdk/version.h 00:06:10.561 23:39:58 version -- app/version.sh@14 -- # cut -f2 00:06:10.561 23:39:58 version -- app/version.sh@14 -- # tr -d '"' 00:06:10.561 23:39:58 version -- app/version.sh@17 -- # major=25 00:06:10.561 23:39:58 version -- app/version.sh@18 -- # get_header_version minor 00:06:10.562 23:39:58 version -- app/version.sh@13 -- # grep -E '^#define SPDK_VERSION_MINOR[[:space:]]+' /home/vagrant/spdk_repo/spdk/include/spdk/version.h 00:06:10.562 23:39:58 version -- app/version.sh@14 -- # cut -f2 00:06:10.562 23:39:58 version -- app/version.sh@14 -- # tr -d '"' 00:06:10.562 23:39:58 version -- app/version.sh@18 -- # minor=1 00:06:10.562 23:39:58 version -- app/version.sh@19 -- # get_header_version patch 00:06:10.562 23:39:58 version -- app/version.sh@13 -- # grep -E '^#define SPDK_VERSION_PATCH[[:space:]]+' /home/vagrant/spdk_repo/spdk/include/spdk/version.h 00:06:10.562 23:39:58 version -- app/version.sh@14 -- # cut -f2 00:06:10.562 23:39:58 version -- app/version.sh@14 -- # tr -d '"' 00:06:10.562 23:39:58 version -- app/version.sh@19 -- # patch=0 00:06:10.562 23:39:58 version -- app/version.sh@20 -- # get_header_version suffix 00:06:10.562 23:39:58 version -- app/version.sh@13 -- # grep -E '^#define SPDK_VERSION_SUFFIX[[:space:]]+' /home/vagrant/spdk_repo/spdk/include/spdk/version.h 00:06:10.562 23:39:58 version -- app/version.sh@14 -- # cut -f2 00:06:10.562 23:39:58 version -- app/version.sh@14 -- # tr -d '"' 00:06:10.562 23:39:58 version -- app/version.sh@20 -- # suffix=-pre 00:06:10.562 23:39:58 version -- app/version.sh@22 -- # version=25.1 00:06:10.562 23:39:58 version -- app/version.sh@25 -- # (( patch != 0 )) 00:06:10.562 23:39:58 version -- app/version.sh@28 -- # version=25.1rc0 00:06:10.562 23:39:58 version -- app/version.sh@30 -- # PYTHONPATH=:/home/vagrant/spdk_repo/spdk/python:/home/vagrant/spdk_repo/spdk/test/rpc_plugins:/home/vagrant/spdk_repo/spdk/python:/home/vagrant/spdk_repo/spdk/test/rpc_plugins:/home/vagrant/spdk_repo/spdk/python 00:06:10.562 23:39:58 version -- app/version.sh@30 -- # python3 -c 'import spdk; print(spdk.__version__)' 00:06:10.562 23:39:58 version -- app/version.sh@30 -- # py_version=25.1rc0 00:06:10.562 23:39:58 version -- app/version.sh@31 -- # [[ 25.1rc0 == \2\5\.\1\r\c\0 ]] 00:06:10.562 ************************************ 00:06:10.562 END TEST version 00:06:10.562 ************************************ 00:06:10.562 00:06:10.562 real 0m0.322s 00:06:10.562 user 0m0.197s 00:06:10.562 sys 0m0.186s 00:06:10.562 23:39:58 version -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:10.562 23:39:58 version -- common/autotest_common.sh@10 -- # set +x 00:06:10.562 23:39:58 -- spdk/autotest.sh@179 -- # '[' 0 -eq 1 ']' 00:06:10.562 23:39:58 -- spdk/autotest.sh@188 -- # [[ 1 -eq 1 ]] 00:06:10.562 23:39:58 -- spdk/autotest.sh@189 -- # run_test bdev_raid /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh 00:06:10.562 23:39:58 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:06:10.562 23:39:58 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:10.562 23:39:58 -- common/autotest_common.sh@10 -- # set +x 00:06:10.562 ************************************ 00:06:10.562 START TEST bdev_raid 00:06:10.562 ************************************ 00:06:10.562 23:39:58 bdev_raid -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh 00:06:10.821 * Looking for test storage... 00:06:10.822 * Found test storage at /home/vagrant/spdk_repo/spdk/test/bdev 00:06:10.822 23:39:58 bdev_raid -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:06:10.822 23:39:58 bdev_raid -- common/autotest_common.sh@1693 -- # lcov --version 00:06:10.822 23:39:58 bdev_raid -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:06:10.822 23:39:58 bdev_raid -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:06:10.822 23:39:58 bdev_raid -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:06:10.822 23:39:58 bdev_raid -- scripts/common.sh@333 -- # local ver1 ver1_l 00:06:10.822 23:39:58 bdev_raid -- scripts/common.sh@334 -- # local ver2 ver2_l 00:06:10.822 23:39:58 bdev_raid -- scripts/common.sh@336 -- # IFS=.-: 00:06:10.822 23:39:58 bdev_raid -- scripts/common.sh@336 -- # read -ra ver1 00:06:10.822 23:39:58 bdev_raid -- scripts/common.sh@337 -- # IFS=.-: 00:06:10.822 23:39:58 bdev_raid -- scripts/common.sh@337 -- # read -ra ver2 00:06:10.822 23:39:58 bdev_raid -- scripts/common.sh@338 -- # local 'op=<' 00:06:10.822 23:39:58 bdev_raid -- scripts/common.sh@340 -- # ver1_l=2 00:06:10.822 23:39:58 bdev_raid -- scripts/common.sh@341 -- # ver2_l=1 00:06:10.822 23:39:58 bdev_raid -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:06:10.822 23:39:58 bdev_raid -- scripts/common.sh@344 -- # case "$op" in 00:06:10.822 23:39:58 bdev_raid -- scripts/common.sh@345 -- # : 1 00:06:10.822 23:39:58 bdev_raid -- scripts/common.sh@364 -- # (( v = 0 )) 00:06:10.822 23:39:58 bdev_raid -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:06:10.822 23:39:58 bdev_raid -- scripts/common.sh@365 -- # decimal 1 00:06:10.822 23:39:58 bdev_raid -- scripts/common.sh@353 -- # local d=1 00:06:10.822 23:39:58 bdev_raid -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:06:10.822 23:39:58 bdev_raid -- scripts/common.sh@355 -- # echo 1 00:06:10.822 23:39:58 bdev_raid -- scripts/common.sh@365 -- # ver1[v]=1 00:06:10.822 23:39:58 bdev_raid -- scripts/common.sh@366 -- # decimal 2 00:06:10.822 23:39:58 bdev_raid -- scripts/common.sh@353 -- # local d=2 00:06:10.822 23:39:58 bdev_raid -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:06:10.822 23:39:58 bdev_raid -- scripts/common.sh@355 -- # echo 2 00:06:10.822 23:39:58 bdev_raid -- scripts/common.sh@366 -- # ver2[v]=2 00:06:10.822 23:39:58 bdev_raid -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:06:10.822 23:39:58 bdev_raid -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:06:10.822 23:39:58 bdev_raid -- scripts/common.sh@368 -- # return 0 00:06:10.822 23:39:58 bdev_raid -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:06:10.822 23:39:58 bdev_raid -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:06:10.822 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:10.822 --rc genhtml_branch_coverage=1 00:06:10.822 --rc genhtml_function_coverage=1 00:06:10.822 --rc genhtml_legend=1 00:06:10.822 --rc geninfo_all_blocks=1 00:06:10.822 --rc geninfo_unexecuted_blocks=1 00:06:10.822 00:06:10.822 ' 00:06:10.822 23:39:58 bdev_raid -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:06:10.822 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:10.822 --rc genhtml_branch_coverage=1 00:06:10.822 --rc genhtml_function_coverage=1 00:06:10.822 --rc genhtml_legend=1 00:06:10.822 --rc geninfo_all_blocks=1 00:06:10.822 --rc geninfo_unexecuted_blocks=1 00:06:10.822 00:06:10.822 ' 00:06:10.822 23:39:58 bdev_raid -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:06:10.822 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:10.822 --rc genhtml_branch_coverage=1 00:06:10.822 --rc genhtml_function_coverage=1 00:06:10.822 --rc genhtml_legend=1 00:06:10.822 --rc geninfo_all_blocks=1 00:06:10.822 --rc geninfo_unexecuted_blocks=1 00:06:10.822 00:06:10.822 ' 00:06:10.822 23:39:58 bdev_raid -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:06:10.822 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:10.822 --rc genhtml_branch_coverage=1 00:06:10.822 --rc genhtml_function_coverage=1 00:06:10.822 --rc genhtml_legend=1 00:06:10.822 --rc geninfo_all_blocks=1 00:06:10.822 --rc geninfo_unexecuted_blocks=1 00:06:10.822 00:06:10.822 ' 00:06:10.822 23:39:58 bdev_raid -- bdev/bdev_raid.sh@12 -- # source /home/vagrant/spdk_repo/spdk/test/bdev/nbd_common.sh 00:06:10.822 23:39:58 bdev_raid -- bdev/nbd_common.sh@6 -- # set -e 00:06:10.822 23:39:58 bdev_raid -- bdev/bdev_raid.sh@14 -- # rpc_py=rpc_cmd 00:06:10.822 23:39:58 bdev_raid -- bdev/bdev_raid.sh@946 -- # mkdir -p /raidtest 00:06:10.822 23:39:58 bdev_raid -- bdev/bdev_raid.sh@947 -- # trap 'cleanup; exit 1' EXIT 00:06:10.822 23:39:58 bdev_raid -- bdev/bdev_raid.sh@949 -- # base_blocklen=512 00:06:10.822 23:39:58 bdev_raid -- bdev/bdev_raid.sh@951 -- # run_test raid1_resize_data_offset_test raid_resize_data_offset_test 00:06:10.822 23:39:58 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:06:10.822 23:39:58 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:10.822 23:39:58 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:06:10.822 ************************************ 00:06:10.822 START TEST raid1_resize_data_offset_test 00:06:10.822 ************************************ 00:06:10.822 23:39:58 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@1129 -- # raid_resize_data_offset_test 00:06:10.822 23:39:58 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@917 -- # raid_pid=71316 00:06:10.822 Process raid pid: 71316 00:06:10.822 23:39:58 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@918 -- # echo 'Process raid pid: 71316' 00:06:10.822 23:39:58 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@919 -- # waitforlisten 71316 00:06:10.822 23:39:58 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@916 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:06:10.822 23:39:58 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@835 -- # '[' -z 71316 ']' 00:06:10.822 23:39:58 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:10.822 23:39:58 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:10.822 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:10.822 23:39:58 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:10.822 23:39:58 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:10.822 23:39:58 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:06:10.822 [2024-11-26 23:39:58.912230] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:06:10.822 [2024-11-26 23:39:58.912368] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:06:11.082 [2024-11-26 23:39:59.067303] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:11.082 [2024-11-26 23:39:59.092129] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:11.082 [2024-11-26 23:39:59.133460] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:06:11.082 [2024-11-26 23:39:59.133500] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:06:11.665 23:39:59 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:11.665 23:39:59 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@868 -- # return 0 00:06:11.665 23:39:59 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@922 -- # rpc_cmd bdev_malloc_create -b malloc0 64 512 -o 16 00:06:11.665 23:39:59 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:11.665 23:39:59 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:06:11.665 malloc0 00:06:11.665 23:39:59 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:11.665 23:39:59 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@923 -- # rpc_cmd bdev_malloc_create -b malloc1 64 512 -o 16 00:06:11.665 23:39:59 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:11.665 23:39:59 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:06:11.665 malloc1 00:06:11.665 23:39:59 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:11.665 23:39:59 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@924 -- # rpc_cmd bdev_null_create null0 64 512 00:06:11.665 23:39:59 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:11.665 23:39:59 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:06:11.942 null0 00:06:11.942 23:39:59 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:11.942 23:39:59 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@926 -- # rpc_cmd bdev_raid_create -n Raid -r 1 -b ''\''malloc0 malloc1 null0'\''' -s 00:06:11.942 23:39:59 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:11.942 23:39:59 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:06:11.942 [2024-11-26 23:39:59.803076] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc0 is claimed 00:06:11.942 [2024-11-26 23:39:59.804909] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:06:11.942 [2024-11-26 23:39:59.804954] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev null0 is claimed 00:06:11.942 [2024-11-26 23:39:59.805074] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:06:11.942 [2024-11-26 23:39:59.805089] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 129024, blocklen 512 00:06:11.942 [2024-11-26 23:39:59.805407] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000021f0 00:06:11.942 [2024-11-26 23:39:59.805533] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:06:11.942 [2024-11-26 23:39:59.805550] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Raid, raid_bdev 0x617000001200 00:06:11.942 [2024-11-26 23:39:59.805675] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:06:11.942 23:39:59 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:11.942 23:39:59 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@929 -- # rpc_cmd bdev_raid_get_bdevs all 00:06:11.942 23:39:59 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@929 -- # jq -r '.[].base_bdevs_list[2].data_offset' 00:06:11.942 23:39:59 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:11.942 23:39:59 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:06:11.942 23:39:59 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:11.942 23:39:59 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@929 -- # (( 2048 == 2048 )) 00:06:11.942 23:39:59 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@931 -- # rpc_cmd bdev_null_delete null0 00:06:11.942 23:39:59 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:11.942 23:39:59 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:06:11.942 [2024-11-26 23:39:59.858997] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: null0 00:06:11.942 23:39:59 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:11.942 23:39:59 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@935 -- # rpc_cmd bdev_malloc_create -b malloc2 512 512 -o 30 00:06:11.942 23:39:59 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:11.942 23:39:59 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:06:11.942 malloc2 00:06:11.942 23:39:59 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:11.942 23:39:59 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@936 -- # rpc_cmd bdev_raid_add_base_bdev Raid malloc2 00:06:11.942 23:39:59 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:11.942 23:39:59 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:06:11.942 [2024-11-26 23:39:59.980875] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:06:11.942 [2024-11-26 23:39:59.985913] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:06:11.942 23:39:59 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:11.942 [2024-11-26 23:39:59.987785] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev Raid 00:06:11.942 23:39:59 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@939 -- # rpc_cmd bdev_raid_get_bdevs all 00:06:11.942 23:39:59 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@939 -- # jq -r '.[].base_bdevs_list[2].data_offset' 00:06:11.942 23:39:59 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:11.942 23:39:59 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:06:11.942 23:40:00 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:11.942 23:40:00 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@939 -- # (( 2070 == 2070 )) 00:06:11.942 23:40:00 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@941 -- # killprocess 71316 00:06:11.942 23:40:00 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@954 -- # '[' -z 71316 ']' 00:06:11.942 23:40:00 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@958 -- # kill -0 71316 00:06:11.942 23:40:00 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@959 -- # uname 00:06:11.942 23:40:00 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:06:11.942 23:40:00 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 71316 00:06:12.202 23:40:00 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:06:12.202 23:40:00 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:06:12.202 killing process with pid 71316 00:06:12.202 23:40:00 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 71316' 00:06:12.202 23:40:00 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@973 -- # kill 71316 00:06:12.202 23:40:00 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@978 -- # wait 71316 00:06:12.202 [2024-11-26 23:40:00.081953] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:06:12.202 [2024-11-26 23:40:00.083652] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev Raid: Operation canceled 00:06:12.202 [2024-11-26 23:40:00.083711] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:06:12.202 [2024-11-26 23:40:00.083728] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: malloc2 00:06:12.202 [2024-11-26 23:40:00.089462] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:06:12.202 [2024-11-26 23:40:00.089743] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:06:12.202 [2024-11-26 23:40:00.089764] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Raid, state offline 00:06:12.202 [2024-11-26 23:40:00.297051] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:06:12.461 23:40:00 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@943 -- # return 0 00:06:12.461 00:06:12.461 real 0m1.673s 00:06:12.461 user 0m1.675s 00:06:12.461 sys 0m0.433s 00:06:12.461 23:40:00 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:12.461 23:40:00 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:06:12.461 ************************************ 00:06:12.461 END TEST raid1_resize_data_offset_test 00:06:12.461 ************************************ 00:06:12.461 23:40:00 bdev_raid -- bdev/bdev_raid.sh@953 -- # run_test raid0_resize_superblock_test raid_resize_superblock_test 0 00:06:12.461 23:40:00 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:06:12.461 23:40:00 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:12.461 23:40:00 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:06:12.461 ************************************ 00:06:12.461 START TEST raid0_resize_superblock_test 00:06:12.461 ************************************ 00:06:12.461 23:40:00 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@1129 -- # raid_resize_superblock_test 0 00:06:12.461 23:40:00 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@854 -- # local raid_level=0 00:06:12.461 23:40:00 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@857 -- # raid_pid=71367 00:06:12.461 23:40:00 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@856 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:06:12.461 Process raid pid: 71367 00:06:12.461 23:40:00 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@858 -- # echo 'Process raid pid: 71367' 00:06:12.461 23:40:00 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@859 -- # waitforlisten 71367 00:06:12.461 23:40:00 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@835 -- # '[' -z 71367 ']' 00:06:12.461 23:40:00 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:12.461 23:40:00 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:12.461 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:12.461 23:40:00 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:12.461 23:40:00 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:12.461 23:40:00 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:12.720 [2024-11-26 23:40:00.655165] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:06:12.720 [2024-11-26 23:40:00.655305] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:06:12.720 [2024-11-26 23:40:00.808756] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:12.720 [2024-11-26 23:40:00.833416] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:12.978 [2024-11-26 23:40:00.874642] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:06:12.978 [2024-11-26 23:40:00.874689] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:06:13.546 23:40:01 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:13.546 23:40:01 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@868 -- # return 0 00:06:13.546 23:40:01 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@861 -- # rpc_cmd bdev_malloc_create -b malloc0 512 512 00:06:13.546 23:40:01 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:13.546 23:40:01 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:13.546 malloc0 00:06:13.546 23:40:01 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:13.546 23:40:01 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@863 -- # rpc_cmd bdev_passthru_create -b malloc0 -p pt0 00:06:13.546 23:40:01 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:13.546 23:40:01 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:13.546 [2024-11-26 23:40:01.625456] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc0 00:06:13.546 [2024-11-26 23:40:01.625512] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:06:13.546 [2024-11-26 23:40:01.625531] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:06:13.546 [2024-11-26 23:40:01.625542] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:06:13.546 [2024-11-26 23:40:01.627724] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:06:13.546 [2024-11-26 23:40:01.627762] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt0 00:06:13.546 pt0 00:06:13.546 23:40:01 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:13.547 23:40:01 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@864 -- # rpc_cmd bdev_lvol_create_lvstore pt0 lvs0 00:06:13.547 23:40:01 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:13.547 23:40:01 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:13.814 979ac72a-e200-4982-a2ed-557929491e45 00:06:13.814 23:40:01 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:13.814 23:40:01 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@866 -- # rpc_cmd bdev_lvol_create -l lvs0 lvol0 64 00:06:13.814 23:40:01 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:13.814 23:40:01 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:13.814 3056121e-5f4b-462e-9fd9-a85ccd267f80 00:06:13.814 23:40:01 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:13.814 23:40:01 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@867 -- # rpc_cmd bdev_lvol_create -l lvs0 lvol1 64 00:06:13.814 23:40:01 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:13.814 23:40:01 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:13.814 ecb15cd0-95c4-4c10-a309-284035fead7d 00:06:13.814 23:40:01 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:13.814 23:40:01 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@869 -- # case $raid_level in 00:06:13.814 23:40:01 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@870 -- # rpc_cmd bdev_raid_create -n Raid -r 0 -z 64 -b ''\''lvs0/lvol0 lvs0/lvol1'\''' -s 00:06:13.814 23:40:01 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:13.814 23:40:01 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:13.814 [2024-11-26 23:40:01.760982] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev 3056121e-5f4b-462e-9fd9-a85ccd267f80 is claimed 00:06:13.814 [2024-11-26 23:40:01.761070] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev ecb15cd0-95c4-4c10-a309-284035fead7d is claimed 00:06:13.814 [2024-11-26 23:40:01.761179] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:06:13.814 [2024-11-26 23:40:01.761192] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 245760, blocklen 512 00:06:13.814 [2024-11-26 23:40:01.761513] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:06:13.814 [2024-11-26 23:40:01.761714] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:06:13.814 [2024-11-26 23:40:01.761739] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Raid, raid_bdev 0x617000001200 00:06:13.814 [2024-11-26 23:40:01.761887] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:06:13.814 23:40:01 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:13.814 23:40:01 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@875 -- # rpc_cmd bdev_get_bdevs -b lvs0/lvol0 00:06:13.814 23:40:01 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@875 -- # jq '.[].num_blocks' 00:06:13.814 23:40:01 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:13.814 23:40:01 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:13.814 23:40:01 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:13.814 23:40:01 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@875 -- # (( 64 == 64 )) 00:06:13.814 23:40:01 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@876 -- # jq '.[].num_blocks' 00:06:13.814 23:40:01 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@876 -- # rpc_cmd bdev_get_bdevs -b lvs0/lvol1 00:06:13.814 23:40:01 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:13.814 23:40:01 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:13.814 23:40:01 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:13.814 23:40:01 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@876 -- # (( 64 == 64 )) 00:06:13.814 23:40:01 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@879 -- # case $raid_level in 00:06:13.814 23:40:01 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@880 -- # jq '.[].num_blocks' 00:06:13.814 23:40:01 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@879 -- # case $raid_level in 00:06:13.814 23:40:01 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@880 -- # rpc_cmd bdev_get_bdevs -b Raid 00:06:13.814 23:40:01 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:13.814 23:40:01 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:13.814 [2024-11-26 23:40:01.853011] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:06:13.814 23:40:01 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:13.814 23:40:01 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@879 -- # case $raid_level in 00:06:13.814 23:40:01 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@879 -- # case $raid_level in 00:06:13.814 23:40:01 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@880 -- # (( 245760 == 245760 )) 00:06:13.814 23:40:01 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@885 -- # rpc_cmd bdev_lvol_resize lvs0/lvol0 100 00:06:13.814 23:40:01 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:13.814 23:40:01 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:13.814 [2024-11-26 23:40:01.892878] bdev_raid.c:2317:raid_bdev_resize_base_bdev: *DEBUG*: raid_bdev_resize_base_bdev 00:06:13.814 [2024-11-26 23:40:01.892907] bdev_raid.c:2330:raid_bdev_resize_base_bdev: *NOTICE*: base_bdev '3056121e-5f4b-462e-9fd9-a85ccd267f80' was resized: old size 131072, new size 204800 00:06:13.814 23:40:01 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:13.814 23:40:01 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@886 -- # rpc_cmd bdev_lvol_resize lvs0/lvol1 100 00:06:13.814 23:40:01 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:13.814 23:40:01 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:13.814 [2024-11-26 23:40:01.904797] bdev_raid.c:2317:raid_bdev_resize_base_bdev: *DEBUG*: raid_bdev_resize_base_bdev 00:06:13.814 [2024-11-26 23:40:01.904824] bdev_raid.c:2330:raid_bdev_resize_base_bdev: *NOTICE*: base_bdev 'ecb15cd0-95c4-4c10-a309-284035fead7d' was resized: old size 131072, new size 204800 00:06:13.814 [2024-11-26 23:40:01.904861] bdev_raid.c:2344:raid_bdev_resize_base_bdev: *NOTICE*: raid bdev 'Raid': block count was changed from 245760 to 393216 00:06:13.814 23:40:01 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:13.814 23:40:01 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@889 -- # rpc_cmd bdev_get_bdevs -b lvs0/lvol0 00:06:13.814 23:40:01 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@889 -- # jq '.[].num_blocks' 00:06:13.814 23:40:01 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:13.814 23:40:01 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:13.814 23:40:01 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:14.081 23:40:01 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@889 -- # (( 100 == 100 )) 00:06:14.081 23:40:01 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@890 -- # rpc_cmd bdev_get_bdevs -b lvs0/lvol1 00:06:14.081 23:40:01 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:14.081 23:40:01 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:14.081 23:40:01 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@890 -- # jq '.[].num_blocks' 00:06:14.081 23:40:01 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:14.081 23:40:01 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@890 -- # (( 100 == 100 )) 00:06:14.081 23:40:01 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@893 -- # case $raid_level in 00:06:14.081 23:40:01 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@894 -- # jq '.[].num_blocks' 00:06:14.081 23:40:01 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@893 -- # case $raid_level in 00:06:14.081 23:40:01 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@894 -- # rpc_cmd bdev_get_bdevs -b Raid 00:06:14.081 23:40:01 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:14.081 23:40:01 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:14.081 [2024-11-26 23:40:02.000716] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:06:14.081 23:40:02 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:14.081 23:40:01 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@893 -- # case $raid_level in 00:06:14.081 23:40:01 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@893 -- # case $raid_level in 00:06:14.081 23:40:02 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@894 -- # (( 393216 == 393216 )) 00:06:14.081 23:40:02 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@898 -- # rpc_cmd bdev_passthru_delete pt0 00:06:14.081 23:40:02 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:14.081 23:40:02 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:14.081 [2024-11-26 23:40:02.028484] vbdev_lvol.c: 150:vbdev_lvs_hotremove_cb: *NOTICE*: bdev pt0 being removed: closing lvstore lvs0 00:06:14.081 [2024-11-26 23:40:02.028557] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: lvs0/lvol0 00:06:14.081 [2024-11-26 23:40:02.028576] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:06:14.081 [2024-11-26 23:40:02.028587] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: lvs0/lvol1 00:06:14.081 [2024-11-26 23:40:02.028688] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:06:14.081 [2024-11-26 23:40:02.028727] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:06:14.081 [2024-11-26 23:40:02.028738] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Raid, state offline 00:06:14.081 23:40:02 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:14.081 23:40:02 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@899 -- # rpc_cmd bdev_passthru_create -b malloc0 -p pt0 00:06:14.081 23:40:02 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:14.081 23:40:02 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:14.081 [2024-11-26 23:40:02.040449] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc0 00:06:14.081 [2024-11-26 23:40:02.040502] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:06:14.081 [2024-11-26 23:40:02.040520] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:06:14.081 [2024-11-26 23:40:02.040538] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:06:14.081 [2024-11-26 23:40:02.042580] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:06:14.081 [2024-11-26 23:40:02.042617] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt0 00:06:14.081 [2024-11-26 23:40:02.043973] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev 3056121e-5f4b-462e-9fd9-a85ccd267f80 00:06:14.081 [2024-11-26 23:40:02.044052] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev 3056121e-5f4b-462e-9fd9-a85ccd267f80 is claimed 00:06:14.081 [2024-11-26 23:40:02.044150] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev ecb15cd0-95c4-4c10-a309-284035fead7d 00:06:14.081 [2024-11-26 23:40:02.044187] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev ecb15cd0-95c4-4c10-a309-284035fead7d is claimed 00:06:14.081 [2024-11-26 23:40:02.044264] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev ecb15cd0-95c4-4c10-a309-284035fead7d (2) smaller than existing raid bdev Raid (3) 00:06:14.081 [2024-11-26 23:40:02.044281] bdev_raid.c:3894:raid_bdev_examine_done: *ERROR*: Failed to examine bdev 3056121e-5f4b-462e-9fd9-a85ccd267f80: File exists 00:06:14.081 [2024-11-26 23:40:02.044316] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001580 00:06:14.081 [2024-11-26 23:40:02.044324] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 393216, blocklen 512 00:06:14.081 [2024-11-26 23:40:02.044550] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000026d0 00:06:14.081 [2024-11-26 23:40:02.044704] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001580 00:06:14.081 [2024-11-26 23:40:02.044721] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Raid, raid_bdev 0x617000001580 00:06:14.081 [2024-11-26 23:40:02.044858] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:06:14.081 pt0 00:06:14.082 23:40:02 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:14.082 23:40:02 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@900 -- # rpc_cmd bdev_wait_for_examine 00:06:14.082 23:40:02 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:14.082 23:40:02 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:14.082 23:40:02 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:14.082 23:40:02 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@904 -- # case $raid_level in 00:06:14.082 23:40:02 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@905 -- # rpc_cmd bdev_get_bdevs -b Raid 00:06:14.082 23:40:02 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@904 -- # case $raid_level in 00:06:14.082 23:40:02 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@905 -- # jq '.[].num_blocks' 00:06:14.082 23:40:02 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:14.082 23:40:02 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:14.082 [2024-11-26 23:40:02.068975] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:06:14.082 23:40:02 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:14.082 23:40:02 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@904 -- # case $raid_level in 00:06:14.082 23:40:02 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@904 -- # case $raid_level in 00:06:14.082 23:40:02 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@905 -- # (( 393216 == 393216 )) 00:06:14.082 23:40:02 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@909 -- # killprocess 71367 00:06:14.082 23:40:02 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@954 -- # '[' -z 71367 ']' 00:06:14.082 23:40:02 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@958 -- # kill -0 71367 00:06:14.082 23:40:02 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@959 -- # uname 00:06:14.082 23:40:02 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:06:14.082 23:40:02 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 71367 00:06:14.082 23:40:02 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:06:14.082 23:40:02 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:06:14.082 killing process with pid 71367 00:06:14.082 23:40:02 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 71367' 00:06:14.082 23:40:02 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@973 -- # kill 71367 00:06:14.082 [2024-11-26 23:40:02.134397] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:06:14.082 [2024-11-26 23:40:02.134466] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:06:14.082 [2024-11-26 23:40:02.134507] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:06:14.082 [2024-11-26 23:40:02.134514] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Raid, state offline 00:06:14.082 23:40:02 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@978 -- # wait 71367 00:06:14.341 [2024-11-26 23:40:02.289679] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:06:14.599 23:40:02 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@911 -- # return 0 00:06:14.599 00:06:14.599 real 0m1.922s 00:06:14.599 user 0m2.189s 00:06:14.599 sys 0m0.460s 00:06:14.599 23:40:02 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:14.599 23:40:02 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:14.599 ************************************ 00:06:14.599 END TEST raid0_resize_superblock_test 00:06:14.599 ************************************ 00:06:14.599 23:40:02 bdev_raid -- bdev/bdev_raid.sh@954 -- # run_test raid1_resize_superblock_test raid_resize_superblock_test 1 00:06:14.599 23:40:02 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:06:14.599 23:40:02 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:14.599 23:40:02 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:06:14.599 ************************************ 00:06:14.599 START TEST raid1_resize_superblock_test 00:06:14.599 ************************************ 00:06:14.599 23:40:02 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@1129 -- # raid_resize_superblock_test 1 00:06:14.599 23:40:02 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@854 -- # local raid_level=1 00:06:14.599 23:40:02 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@857 -- # raid_pid=71438 00:06:14.599 Process raid pid: 71438 00:06:14.599 23:40:02 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@856 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:06:14.599 23:40:02 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@858 -- # echo 'Process raid pid: 71438' 00:06:14.599 23:40:02 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@859 -- # waitforlisten 71438 00:06:14.599 23:40:02 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@835 -- # '[' -z 71438 ']' 00:06:14.599 23:40:02 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:14.599 23:40:02 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:14.599 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:14.599 23:40:02 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:14.599 23:40:02 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:14.599 23:40:02 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:14.599 [2024-11-26 23:40:02.646275] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:06:14.599 [2024-11-26 23:40:02.646437] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:06:14.859 [2024-11-26 23:40:02.801015] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:14.859 [2024-11-26 23:40:02.825234] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:14.859 [2024-11-26 23:40:02.866956] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:06:14.859 [2024-11-26 23:40:02.866990] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:06:15.428 23:40:03 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:15.428 23:40:03 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@868 -- # return 0 00:06:15.428 23:40:03 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@861 -- # rpc_cmd bdev_malloc_create -b malloc0 512 512 00:06:15.428 23:40:03 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:15.428 23:40:03 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:15.689 malloc0 00:06:15.689 23:40:03 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:15.689 23:40:03 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@863 -- # rpc_cmd bdev_passthru_create -b malloc0 -p pt0 00:06:15.689 23:40:03 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:15.689 23:40:03 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:15.689 [2024-11-26 23:40:03.591927] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc0 00:06:15.689 [2024-11-26 23:40:03.592019] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:06:15.689 [2024-11-26 23:40:03.592045] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:06:15.689 [2024-11-26 23:40:03.592058] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:06:15.689 [2024-11-26 23:40:03.594400] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:06:15.689 [2024-11-26 23:40:03.594446] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt0 00:06:15.689 pt0 00:06:15.689 23:40:03 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:15.689 23:40:03 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@864 -- # rpc_cmd bdev_lvol_create_lvstore pt0 lvs0 00:06:15.689 23:40:03 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:15.689 23:40:03 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:15.689 2f04f54e-d3f2-4578-958f-36fcc362bdf9 00:06:15.689 23:40:03 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:15.689 23:40:03 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@866 -- # rpc_cmd bdev_lvol_create -l lvs0 lvol0 64 00:06:15.689 23:40:03 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:15.689 23:40:03 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:15.689 2e11477a-a6d5-4143-878d-5324996351b3 00:06:15.689 23:40:03 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:15.689 23:40:03 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@867 -- # rpc_cmd bdev_lvol_create -l lvs0 lvol1 64 00:06:15.689 23:40:03 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:15.689 23:40:03 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:15.689 bb024705-4010-48ca-bd55-38633afcf038 00:06:15.689 23:40:03 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:15.689 23:40:03 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@869 -- # case $raid_level in 00:06:15.689 23:40:03 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@871 -- # rpc_cmd bdev_raid_create -n Raid -r 1 -b ''\''lvs0/lvol0 lvs0/lvol1'\''' -s 00:06:15.689 23:40:03 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:15.689 23:40:03 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:15.689 [2024-11-26 23:40:03.727371] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev 2e11477a-a6d5-4143-878d-5324996351b3 is claimed 00:06:15.689 [2024-11-26 23:40:03.727460] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev bb024705-4010-48ca-bd55-38633afcf038 is claimed 00:06:15.689 [2024-11-26 23:40:03.727568] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:06:15.689 [2024-11-26 23:40:03.727587] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 122880, blocklen 512 00:06:15.689 [2024-11-26 23:40:03.727853] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:06:15.689 [2024-11-26 23:40:03.728002] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:06:15.690 [2024-11-26 23:40:03.728019] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Raid, raid_bdev 0x617000001200 00:06:15.690 [2024-11-26 23:40:03.728156] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:06:15.690 23:40:03 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:15.690 23:40:03 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@875 -- # rpc_cmd bdev_get_bdevs -b lvs0/lvol0 00:06:15.690 23:40:03 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@875 -- # jq '.[].num_blocks' 00:06:15.690 23:40:03 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:15.690 23:40:03 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:15.690 23:40:03 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:15.690 23:40:03 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@875 -- # (( 64 == 64 )) 00:06:15.690 23:40:03 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@876 -- # jq '.[].num_blocks' 00:06:15.690 23:40:03 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@876 -- # rpc_cmd bdev_get_bdevs -b lvs0/lvol1 00:06:15.690 23:40:03 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:15.690 23:40:03 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:15.690 23:40:03 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:15.690 23:40:03 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@876 -- # (( 64 == 64 )) 00:06:15.690 23:40:03 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@879 -- # case $raid_level in 00:06:15.690 23:40:03 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@881 -- # rpc_cmd bdev_get_bdevs -b Raid 00:06:15.690 23:40:03 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:15.690 23:40:03 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:15.690 23:40:03 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@879 -- # case $raid_level in 00:06:15.690 23:40:03 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@881 -- # jq '.[].num_blocks' 00:06:15.690 [2024-11-26 23:40:03.815480] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:06:15.950 23:40:03 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:15.950 23:40:03 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@879 -- # case $raid_level in 00:06:15.950 23:40:03 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@879 -- # case $raid_level in 00:06:15.950 23:40:03 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@881 -- # (( 122880 == 122880 )) 00:06:15.950 23:40:03 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@885 -- # rpc_cmd bdev_lvol_resize lvs0/lvol0 100 00:06:15.950 23:40:03 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:15.950 23:40:03 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:15.950 [2024-11-26 23:40:03.863267] bdev_raid.c:2317:raid_bdev_resize_base_bdev: *DEBUG*: raid_bdev_resize_base_bdev 00:06:15.950 [2024-11-26 23:40:03.863296] bdev_raid.c:2330:raid_bdev_resize_base_bdev: *NOTICE*: base_bdev '2e11477a-a6d5-4143-878d-5324996351b3' was resized: old size 131072, new size 204800 00:06:15.950 23:40:03 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:15.950 23:40:03 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@886 -- # rpc_cmd bdev_lvol_resize lvs0/lvol1 100 00:06:15.950 23:40:03 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:15.950 23:40:03 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:15.950 [2024-11-26 23:40:03.875175] bdev_raid.c:2317:raid_bdev_resize_base_bdev: *DEBUG*: raid_bdev_resize_base_bdev 00:06:15.950 [2024-11-26 23:40:03.875203] bdev_raid.c:2330:raid_bdev_resize_base_bdev: *NOTICE*: base_bdev 'bb024705-4010-48ca-bd55-38633afcf038' was resized: old size 131072, new size 204800 00:06:15.950 [2024-11-26 23:40:03.875242] bdev_raid.c:2344:raid_bdev_resize_base_bdev: *NOTICE*: raid bdev 'Raid': block count was changed from 122880 to 196608 00:06:15.950 23:40:03 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:15.950 23:40:03 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@889 -- # jq '.[].num_blocks' 00:06:15.950 23:40:03 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@889 -- # rpc_cmd bdev_get_bdevs -b lvs0/lvol0 00:06:15.950 23:40:03 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:15.950 23:40:03 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:15.950 23:40:03 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:15.950 23:40:03 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@889 -- # (( 100 == 100 )) 00:06:15.950 23:40:03 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@890 -- # rpc_cmd bdev_get_bdevs -b lvs0/lvol1 00:06:15.950 23:40:03 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:15.950 23:40:03 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:15.950 23:40:03 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@890 -- # jq '.[].num_blocks' 00:06:15.950 23:40:03 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:15.950 23:40:03 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@890 -- # (( 100 == 100 )) 00:06:15.950 23:40:03 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@893 -- # case $raid_level in 00:06:15.950 23:40:03 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@895 -- # rpc_cmd bdev_get_bdevs -b Raid 00:06:15.950 23:40:03 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@893 -- # case $raid_level in 00:06:15.950 23:40:03 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@895 -- # jq '.[].num_blocks' 00:06:15.950 23:40:03 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:15.950 23:40:03 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:15.950 [2024-11-26 23:40:03.975181] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:06:15.950 23:40:03 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:15.950 23:40:03 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@893 -- # case $raid_level in 00:06:15.950 23:40:03 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@893 -- # case $raid_level in 00:06:15.950 23:40:04 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@895 -- # (( 196608 == 196608 )) 00:06:15.951 23:40:04 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@898 -- # rpc_cmd bdev_passthru_delete pt0 00:06:15.951 23:40:04 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:15.951 23:40:04 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:15.951 [2024-11-26 23:40:04.018858] vbdev_lvol.c: 150:vbdev_lvs_hotremove_cb: *NOTICE*: bdev pt0 being removed: closing lvstore lvs0 00:06:15.951 [2024-11-26 23:40:04.018924] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: lvs0/lvol0 00:06:15.951 [2024-11-26 23:40:04.018952] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: lvs0/lvol1 00:06:15.951 [2024-11-26 23:40:04.019107] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:06:15.951 [2024-11-26 23:40:04.019284] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:06:15.951 [2024-11-26 23:40:04.019366] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:06:15.951 [2024-11-26 23:40:04.019390] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Raid, state offline 00:06:15.951 23:40:04 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:15.951 23:40:04 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@899 -- # rpc_cmd bdev_passthru_create -b malloc0 -p pt0 00:06:15.951 23:40:04 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:15.951 23:40:04 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:15.951 [2024-11-26 23:40:04.030801] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc0 00:06:15.951 [2024-11-26 23:40:04.030852] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:06:15.951 [2024-11-26 23:40:04.030870] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:06:15.951 [2024-11-26 23:40:04.030880] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:06:15.951 [2024-11-26 23:40:04.033039] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:06:15.951 [2024-11-26 23:40:04.033074] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt0 00:06:15.951 [2024-11-26 23:40:04.034549] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev 2e11477a-a6d5-4143-878d-5324996351b3 00:06:15.951 [2024-11-26 23:40:04.034634] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev 2e11477a-a6d5-4143-878d-5324996351b3 is claimed 00:06:15.951 [2024-11-26 23:40:04.034730] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev bb024705-4010-48ca-bd55-38633afcf038 00:06:15.951 [2024-11-26 23:40:04.034756] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev bb024705-4010-48ca-bd55-38633afcf038 is claimed 00:06:15.951 [2024-11-26 23:40:04.034842] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev bb024705-4010-48ca-bd55-38633afcf038 (2) smaller than existing raid bdev Raid (3) 00:06:15.951 [2024-11-26 23:40:04.034869] bdev_raid.c:3894:raid_bdev_examine_done: *ERROR*: Failed to examine bdev 2e11477a-a6d5-4143-878d-5324996351b3: File exists 00:06:15.951 [2024-11-26 23:40:04.034919] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001580 00:06:15.951 [2024-11-26 23:40:04.034941] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 196608, blocklen 512 00:06:15.951 [2024-11-26 23:40:04.035186] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000026d0 00:06:15.951 [2024-11-26 23:40:04.035373] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001580 00:06:15.951 [2024-11-26 23:40:04.035390] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Raid, raid_bdev 0x617000001580 00:06:15.951 [2024-11-26 23:40:04.035534] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:06:15.951 pt0 00:06:15.951 23:40:04 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:15.951 23:40:04 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@900 -- # rpc_cmd bdev_wait_for_examine 00:06:15.951 23:40:04 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:15.951 23:40:04 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:15.951 23:40:04 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:15.951 23:40:04 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@904 -- # case $raid_level in 00:06:15.951 23:40:04 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@906 -- # rpc_cmd bdev_get_bdevs -b Raid 00:06:15.951 23:40:04 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@904 -- # case $raid_level in 00:06:15.951 23:40:04 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@906 -- # jq '.[].num_blocks' 00:06:15.951 23:40:04 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:15.951 23:40:04 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:15.951 [2024-11-26 23:40:04.059091] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:06:16.211 23:40:04 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:16.211 23:40:04 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@904 -- # case $raid_level in 00:06:16.211 23:40:04 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@904 -- # case $raid_level in 00:06:16.211 23:40:04 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@906 -- # (( 196608 == 196608 )) 00:06:16.211 23:40:04 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@909 -- # killprocess 71438 00:06:16.211 23:40:04 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@954 -- # '[' -z 71438 ']' 00:06:16.211 23:40:04 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@958 -- # kill -0 71438 00:06:16.211 23:40:04 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@959 -- # uname 00:06:16.211 23:40:04 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:06:16.211 23:40:04 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 71438 00:06:16.211 23:40:04 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:06:16.211 23:40:04 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:06:16.211 killing process with pid 71438 00:06:16.211 23:40:04 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 71438' 00:06:16.211 23:40:04 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@973 -- # kill 71438 00:06:16.211 [2024-11-26 23:40:04.140543] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:06:16.211 [2024-11-26 23:40:04.140606] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:06:16.211 [2024-11-26 23:40:04.140658] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:06:16.211 [2024-11-26 23:40:04.140666] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Raid, state offline 00:06:16.211 23:40:04 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@978 -- # wait 71438 00:06:16.211 [2024-11-26 23:40:04.299061] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:06:16.470 23:40:04 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@911 -- # return 0 00:06:16.470 00:06:16.470 real 0m1.942s 00:06:16.470 user 0m2.221s 00:06:16.470 sys 0m0.460s 00:06:16.470 23:40:04 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:16.470 ************************************ 00:06:16.470 END TEST raid1_resize_superblock_test 00:06:16.470 ************************************ 00:06:16.470 23:40:04 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:16.470 23:40:04 bdev_raid -- bdev/bdev_raid.sh@956 -- # uname -s 00:06:16.470 23:40:04 bdev_raid -- bdev/bdev_raid.sh@956 -- # '[' Linux = Linux ']' 00:06:16.470 23:40:04 bdev_raid -- bdev/bdev_raid.sh@956 -- # modprobe -n nbd 00:06:16.470 23:40:04 bdev_raid -- bdev/bdev_raid.sh@957 -- # has_nbd=true 00:06:16.470 23:40:04 bdev_raid -- bdev/bdev_raid.sh@958 -- # modprobe nbd 00:06:16.470 23:40:04 bdev_raid -- bdev/bdev_raid.sh@959 -- # run_test raid_function_test_raid0 raid_function_test raid0 00:06:16.470 23:40:04 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:06:16.470 23:40:04 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:16.470 23:40:04 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:06:16.730 ************************************ 00:06:16.730 START TEST raid_function_test_raid0 00:06:16.730 ************************************ 00:06:16.730 23:40:04 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@1129 -- # raid_function_test raid0 00:06:16.730 23:40:04 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@64 -- # local raid_level=raid0 00:06:16.730 23:40:04 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@65 -- # local nbd=/dev/nbd0 00:06:16.730 23:40:04 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@66 -- # local raid_bdev 00:06:16.730 23:40:04 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@69 -- # raid_pid=71516 00:06:16.730 23:40:04 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@68 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:06:16.730 23:40:04 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@70 -- # echo 'Process raid pid: 71516' 00:06:16.730 Process raid pid: 71516 00:06:16.730 23:40:04 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@71 -- # waitforlisten 71516 00:06:16.730 23:40:04 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@835 -- # '[' -z 71516 ']' 00:06:16.730 23:40:04 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:16.730 23:40:04 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:16.730 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:16.730 23:40:04 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:16.730 23:40:04 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:16.730 23:40:04 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@10 -- # set +x 00:06:16.730 [2024-11-26 23:40:04.682975] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:06:16.730 [2024-11-26 23:40:04.683113] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:06:16.730 [2024-11-26 23:40:04.835577] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:16.990 [2024-11-26 23:40:04.860085] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:16.990 [2024-11-26 23:40:04.901144] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:06:16.990 [2024-11-26 23:40:04.901184] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:06:17.560 23:40:05 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:17.560 23:40:05 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@868 -- # return 0 00:06:17.560 23:40:05 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@73 -- # rpc_cmd bdev_malloc_create 32 512 -b Base_1 00:06:17.560 23:40:05 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:17.560 23:40:05 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@10 -- # set +x 00:06:17.560 Base_1 00:06:17.560 23:40:05 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:17.560 23:40:05 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@74 -- # rpc_cmd bdev_malloc_create 32 512 -b Base_2 00:06:17.560 23:40:05 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:17.560 23:40:05 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@10 -- # set +x 00:06:17.560 Base_2 00:06:17.560 23:40:05 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:17.560 23:40:05 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@75 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''Base_1 Base_2'\''' -n raid 00:06:17.560 23:40:05 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:17.560 23:40:05 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@10 -- # set +x 00:06:17.560 [2024-11-26 23:40:05.548041] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev Base_1 is claimed 00:06:17.560 [2024-11-26 23:40:05.549833] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev Base_2 is claimed 00:06:17.560 [2024-11-26 23:40:05.549897] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:06:17.560 [2024-11-26 23:40:05.549909] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 131072, blocklen 512 00:06:17.560 [2024-11-26 23:40:05.550176] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:06:17.560 [2024-11-26 23:40:05.550307] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:06:17.560 [2024-11-26 23:40:05.550328] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid, raid_bdev 0x617000001200 00:06:17.560 [2024-11-26 23:40:05.550471] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:06:17.560 23:40:05 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:17.560 23:40:05 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@77 -- # rpc_cmd bdev_raid_get_bdevs online 00:06:17.560 23:40:05 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:17.560 23:40:05 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@77 -- # jq -r '.[0]["name"] | select(.)' 00:06:17.560 23:40:05 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@10 -- # set +x 00:06:17.560 23:40:05 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:17.560 23:40:05 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@77 -- # raid_bdev=raid 00:06:17.560 23:40:05 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@78 -- # '[' raid = '' ']' 00:06:17.560 23:40:05 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@83 -- # nbd_start_disks /var/tmp/spdk.sock raid /dev/nbd0 00:06:17.560 23:40:05 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:06:17.561 23:40:05 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@10 -- # bdev_list=('raid') 00:06:17.561 23:40:05 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@10 -- # local bdev_list 00:06:17.561 23:40:05 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:06:17.561 23:40:05 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@11 -- # local nbd_list 00:06:17.561 23:40:05 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@12 -- # local i 00:06:17.561 23:40:05 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:06:17.561 23:40:05 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:06:17.561 23:40:05 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid /dev/nbd0 00:06:17.821 [2024-11-26 23:40:05.775717] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:06:17.821 /dev/nbd0 00:06:17.821 23:40:05 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:06:17.821 23:40:05 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:06:17.821 23:40:05 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:06:17.821 23:40:05 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@873 -- # local i 00:06:17.821 23:40:05 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:06:17.821 23:40:05 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:06:17.821 23:40:05 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:06:17.821 23:40:05 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@877 -- # break 00:06:17.821 23:40:05 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:06:17.821 23:40:05 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:06:17.821 23:40:05 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:06:17.821 1+0 records in 00:06:17.821 1+0 records out 00:06:17.821 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000483744 s, 8.5 MB/s 00:06:17.821 23:40:05 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:06:17.821 23:40:05 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@890 -- # size=4096 00:06:17.821 23:40:05 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:06:17.821 23:40:05 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:06:17.821 23:40:05 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@893 -- # return 0 00:06:17.821 23:40:05 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:06:17.821 23:40:05 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:06:17.821 23:40:05 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@84 -- # nbd_get_count /var/tmp/spdk.sock 00:06:17.821 23:40:05 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk.sock 00:06:17.821 23:40:05 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_get_disks 00:06:18.090 23:40:06 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[ 00:06:18.090 { 00:06:18.090 "nbd_device": "/dev/nbd0", 00:06:18.090 "bdev_name": "raid" 00:06:18.090 } 00:06:18.090 ]' 00:06:18.090 23:40:06 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@64 -- # echo '[ 00:06:18.090 { 00:06:18.090 "nbd_device": "/dev/nbd0", 00:06:18.090 "bdev_name": "raid" 00:06:18.090 } 00:06:18.090 ]' 00:06:18.090 23:40:06 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:06:18.090 23:40:06 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@64 -- # nbd_disks_name=/dev/nbd0 00:06:18.090 23:40:06 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:06:18.090 23:40:06 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@65 -- # echo /dev/nbd0 00:06:18.090 23:40:06 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@65 -- # count=1 00:06:18.090 23:40:06 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@66 -- # echo 1 00:06:18.090 23:40:06 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@84 -- # count=1 00:06:18.090 23:40:06 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@85 -- # '[' 1 -ne 1 ']' 00:06:18.090 23:40:06 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@89 -- # raid_unmap_data_verify /dev/nbd0 00:06:18.090 23:40:06 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@17 -- # hash blkdiscard 00:06:18.090 23:40:06 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@18 -- # local nbd=/dev/nbd0 00:06:18.090 23:40:06 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@19 -- # local blksize 00:06:18.090 23:40:06 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@20 -- # grep -v LOG-SEC 00:06:18.090 23:40:06 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@20 -- # lsblk -o LOG-SEC /dev/nbd0 00:06:18.090 23:40:06 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@20 -- # cut -d ' ' -f 5 00:06:18.090 23:40:06 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@20 -- # blksize=512 00:06:18.090 23:40:06 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@21 -- # local rw_blk_num=4096 00:06:18.090 23:40:06 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@22 -- # local rw_len=2097152 00:06:18.090 23:40:06 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@23 -- # unmap_blk_offs=('0' '1028' '321') 00:06:18.090 23:40:06 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@23 -- # local unmap_blk_offs 00:06:18.090 23:40:06 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@24 -- # unmap_blk_nums=('128' '2035' '456') 00:06:18.090 23:40:06 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@24 -- # local unmap_blk_nums 00:06:18.090 23:40:06 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@25 -- # local unmap_off 00:06:18.090 23:40:06 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@26 -- # local unmap_len 00:06:18.090 23:40:06 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@29 -- # dd if=/dev/urandom of=/raidtest/raidrandtest bs=512 count=4096 00:06:18.090 4096+0 records in 00:06:18.090 4096+0 records out 00:06:18.090 2097152 bytes (2.1 MB, 2.0 MiB) copied, 0.0365563 s, 57.4 MB/s 00:06:18.090 23:40:06 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@30 -- # dd if=/raidtest/raidrandtest of=/dev/nbd0 bs=512 count=4096 oflag=direct 00:06:18.353 4096+0 records in 00:06:18.353 4096+0 records out 00:06:18.353 2097152 bytes (2.1 MB, 2.0 MiB) copied, 0.22442 s, 9.3 MB/s 00:06:18.353 23:40:06 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@31 -- # blockdev --flushbufs /dev/nbd0 00:06:18.353 23:40:06 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@34 -- # cmp -b -n 2097152 /raidtest/raidrandtest /dev/nbd0 00:06:18.353 23:40:06 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@36 -- # (( i = 0 )) 00:06:18.353 23:40:06 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@36 -- # (( i < 3 )) 00:06:18.353 23:40:06 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@37 -- # unmap_off=0 00:06:18.353 23:40:06 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@38 -- # unmap_len=65536 00:06:18.353 23:40:06 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@41 -- # dd if=/dev/zero of=/raidtest/raidrandtest bs=512 seek=0 count=128 conv=notrunc 00:06:18.353 128+0 records in 00:06:18.353 128+0 records out 00:06:18.353 65536 bytes (66 kB, 64 KiB) copied, 0.00182916 s, 35.8 MB/s 00:06:18.353 23:40:06 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@44 -- # blkdiscard -o 0 -l 65536 /dev/nbd0 00:06:18.353 23:40:06 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@45 -- # blockdev --flushbufs /dev/nbd0 00:06:18.353 23:40:06 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@48 -- # cmp -b -n 2097152 /raidtest/raidrandtest /dev/nbd0 00:06:18.353 23:40:06 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@36 -- # (( i++ )) 00:06:18.353 23:40:06 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@36 -- # (( i < 3 )) 00:06:18.353 23:40:06 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@37 -- # unmap_off=526336 00:06:18.353 23:40:06 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@38 -- # unmap_len=1041920 00:06:18.353 23:40:06 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@41 -- # dd if=/dev/zero of=/raidtest/raidrandtest bs=512 seek=1028 count=2035 conv=notrunc 00:06:18.353 2035+0 records in 00:06:18.354 2035+0 records out 00:06:18.354 1041920 bytes (1.0 MB, 1018 KiB) copied, 0.0147579 s, 70.6 MB/s 00:06:18.354 23:40:06 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@44 -- # blkdiscard -o 526336 -l 1041920 /dev/nbd0 00:06:18.354 23:40:06 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@45 -- # blockdev --flushbufs /dev/nbd0 00:06:18.354 23:40:06 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@48 -- # cmp -b -n 2097152 /raidtest/raidrandtest /dev/nbd0 00:06:18.354 23:40:06 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@36 -- # (( i++ )) 00:06:18.354 23:40:06 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@36 -- # (( i < 3 )) 00:06:18.354 23:40:06 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@37 -- # unmap_off=164352 00:06:18.354 23:40:06 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@38 -- # unmap_len=233472 00:06:18.354 23:40:06 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@41 -- # dd if=/dev/zero of=/raidtest/raidrandtest bs=512 seek=321 count=456 conv=notrunc 00:06:18.354 456+0 records in 00:06:18.354 456+0 records out 00:06:18.354 233472 bytes (233 kB, 228 KiB) copied, 0.00368657 s, 63.3 MB/s 00:06:18.614 23:40:06 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@44 -- # blkdiscard -o 164352 -l 233472 /dev/nbd0 00:06:18.614 23:40:06 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@45 -- # blockdev --flushbufs /dev/nbd0 00:06:18.614 23:40:06 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@48 -- # cmp -b -n 2097152 /raidtest/raidrandtest /dev/nbd0 00:06:18.614 23:40:06 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@36 -- # (( i++ )) 00:06:18.614 23:40:06 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@36 -- # (( i < 3 )) 00:06:18.614 23:40:06 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@52 -- # return 0 00:06:18.614 23:40:06 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@91 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:06:18.614 23:40:06 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:06:18.614 23:40:06 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:06:18.614 23:40:06 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@50 -- # local nbd_list 00:06:18.614 23:40:06 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@51 -- # local i 00:06:18.614 23:40:06 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:06:18.614 23:40:06 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:06:18.614 23:40:06 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:06:18.614 [2024-11-26 23:40:06.722110] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:06:18.614 23:40:06 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:06:18.614 23:40:06 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:06:18.614 23:40:06 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:06:18.614 23:40:06 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:06:18.614 23:40:06 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:06:18.614 23:40:06 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@41 -- # break 00:06:18.614 23:40:06 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@45 -- # return 0 00:06:18.614 23:40:06 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@92 -- # nbd_get_count /var/tmp/spdk.sock 00:06:18.614 23:40:06 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk.sock 00:06:18.614 23:40:06 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_get_disks 00:06:18.873 23:40:06 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:06:18.873 23:40:06 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@64 -- # echo '[]' 00:06:18.873 23:40:06 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:06:18.873 23:40:06 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:06:18.873 23:40:06 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@65 -- # echo '' 00:06:18.873 23:40:06 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:06:18.873 23:40:06 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@65 -- # true 00:06:18.873 23:40:06 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@65 -- # count=0 00:06:18.873 23:40:06 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@66 -- # echo 0 00:06:18.873 23:40:06 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@92 -- # count=0 00:06:18.873 23:40:06 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@93 -- # '[' 0 -ne 0 ']' 00:06:18.873 23:40:06 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@97 -- # killprocess 71516 00:06:18.873 23:40:06 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@954 -- # '[' -z 71516 ']' 00:06:18.873 23:40:06 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@958 -- # kill -0 71516 00:06:18.873 23:40:06 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@959 -- # uname 00:06:18.874 23:40:06 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:06:18.874 23:40:06 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 71516 00:06:19.133 killing process with pid 71516 00:06:19.133 23:40:07 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:06:19.133 23:40:07 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:06:19.133 23:40:07 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@972 -- # echo 'killing process with pid 71516' 00:06:19.133 23:40:07 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@973 -- # kill 71516 00:06:19.133 [2024-11-26 23:40:07.033253] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:06:19.133 23:40:07 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@978 -- # wait 71516 00:06:19.133 [2024-11-26 23:40:07.033378] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:06:19.133 [2024-11-26 23:40:07.033432] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:06:19.133 [2024-11-26 23:40:07.033451] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid, state offline 00:06:19.133 [2024-11-26 23:40:07.056085] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:06:19.393 23:40:07 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@99 -- # return 0 00:06:19.393 00:06:19.393 real 0m2.663s 00:06:19.393 user 0m3.263s 00:06:19.393 sys 0m0.941s 00:06:19.393 23:40:07 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:19.393 23:40:07 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@10 -- # set +x 00:06:19.393 ************************************ 00:06:19.393 END TEST raid_function_test_raid0 00:06:19.393 ************************************ 00:06:19.393 23:40:07 bdev_raid -- bdev/bdev_raid.sh@960 -- # run_test raid_function_test_concat raid_function_test concat 00:06:19.393 23:40:07 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:06:19.393 23:40:07 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:19.393 23:40:07 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:06:19.393 ************************************ 00:06:19.393 START TEST raid_function_test_concat 00:06:19.393 ************************************ 00:06:19.393 23:40:07 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@1129 -- # raid_function_test concat 00:06:19.394 23:40:07 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@64 -- # local raid_level=concat 00:06:19.394 23:40:07 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@65 -- # local nbd=/dev/nbd0 00:06:19.394 23:40:07 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@66 -- # local raid_bdev 00:06:19.394 23:40:07 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@69 -- # raid_pid=71630 00:06:19.394 23:40:07 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@68 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:06:19.394 Process raid pid: 71630 00:06:19.394 23:40:07 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@70 -- # echo 'Process raid pid: 71630' 00:06:19.394 23:40:07 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@71 -- # waitforlisten 71630 00:06:19.394 23:40:07 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@835 -- # '[' -z 71630 ']' 00:06:19.394 23:40:07 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:19.394 23:40:07 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:19.394 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:19.394 23:40:07 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:19.394 23:40:07 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:19.394 23:40:07 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@10 -- # set +x 00:06:19.394 [2024-11-26 23:40:07.412374] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:06:19.394 [2024-11-26 23:40:07.412510] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:06:19.654 [2024-11-26 23:40:07.568219] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:19.654 [2024-11-26 23:40:07.593456] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:19.654 [2024-11-26 23:40:07.634422] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:06:19.654 [2024-11-26 23:40:07.634464] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:06:20.229 23:40:08 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:20.229 23:40:08 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@868 -- # return 0 00:06:20.229 23:40:08 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@73 -- # rpc_cmd bdev_malloc_create 32 512 -b Base_1 00:06:20.229 23:40:08 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:20.229 23:40:08 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@10 -- # set +x 00:06:20.229 Base_1 00:06:20.229 23:40:08 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:20.229 23:40:08 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@74 -- # rpc_cmd bdev_malloc_create 32 512 -b Base_2 00:06:20.229 23:40:08 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:20.229 23:40:08 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@10 -- # set +x 00:06:20.229 Base_2 00:06:20.229 23:40:08 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:20.229 23:40:08 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@75 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''Base_1 Base_2'\''' -n raid 00:06:20.229 23:40:08 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:20.229 23:40:08 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@10 -- # set +x 00:06:20.229 [2024-11-26 23:40:08.281804] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev Base_1 is claimed 00:06:20.229 [2024-11-26 23:40:08.283622] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev Base_2 is claimed 00:06:20.229 [2024-11-26 23:40:08.283693] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:06:20.229 [2024-11-26 23:40:08.283710] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 131072, blocklen 512 00:06:20.229 [2024-11-26 23:40:08.283980] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:06:20.229 [2024-11-26 23:40:08.284135] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:06:20.229 [2024-11-26 23:40:08.284149] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid, raid_bdev 0x617000001200 00:06:20.229 [2024-11-26 23:40:08.284280] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:06:20.229 23:40:08 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:20.229 23:40:08 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@77 -- # rpc_cmd bdev_raid_get_bdevs online 00:06:20.229 23:40:08 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@77 -- # jq -r '.[0]["name"] | select(.)' 00:06:20.229 23:40:08 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:20.229 23:40:08 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@10 -- # set +x 00:06:20.229 23:40:08 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:20.229 23:40:08 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@77 -- # raid_bdev=raid 00:06:20.229 23:40:08 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@78 -- # '[' raid = '' ']' 00:06:20.229 23:40:08 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@83 -- # nbd_start_disks /var/tmp/spdk.sock raid /dev/nbd0 00:06:20.229 23:40:08 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:06:20.229 23:40:08 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@10 -- # bdev_list=('raid') 00:06:20.229 23:40:08 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@10 -- # local bdev_list 00:06:20.229 23:40:08 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:06:20.229 23:40:08 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@11 -- # local nbd_list 00:06:20.229 23:40:08 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@12 -- # local i 00:06:20.229 23:40:08 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:06:20.229 23:40:08 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:06:20.229 23:40:08 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid /dev/nbd0 00:06:20.490 [2024-11-26 23:40:08.517514] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:06:20.490 /dev/nbd0 00:06:20.490 23:40:08 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:06:20.490 23:40:08 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:06:20.490 23:40:08 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:06:20.490 23:40:08 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@873 -- # local i 00:06:20.490 23:40:08 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:06:20.490 23:40:08 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:06:20.490 23:40:08 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:06:20.490 23:40:08 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@877 -- # break 00:06:20.490 23:40:08 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:06:20.490 23:40:08 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:06:20.490 23:40:08 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:06:20.490 1+0 records in 00:06:20.490 1+0 records out 00:06:20.490 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000431057 s, 9.5 MB/s 00:06:20.490 23:40:08 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:06:20.490 23:40:08 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@890 -- # size=4096 00:06:20.490 23:40:08 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:06:20.490 23:40:08 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:06:20.490 23:40:08 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@893 -- # return 0 00:06:20.490 23:40:08 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:06:20.490 23:40:08 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:06:20.490 23:40:08 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@84 -- # nbd_get_count /var/tmp/spdk.sock 00:06:20.490 23:40:08 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk.sock 00:06:20.490 23:40:08 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_get_disks 00:06:20.750 23:40:08 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[ 00:06:20.750 { 00:06:20.750 "nbd_device": "/dev/nbd0", 00:06:20.750 "bdev_name": "raid" 00:06:20.750 } 00:06:20.750 ]' 00:06:20.750 23:40:08 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:06:20.750 23:40:08 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@64 -- # echo '[ 00:06:20.750 { 00:06:20.750 "nbd_device": "/dev/nbd0", 00:06:20.750 "bdev_name": "raid" 00:06:20.750 } 00:06:20.750 ]' 00:06:20.750 23:40:08 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@64 -- # nbd_disks_name=/dev/nbd0 00:06:20.750 23:40:08 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:06:20.750 23:40:08 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@65 -- # echo /dev/nbd0 00:06:20.750 23:40:08 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@65 -- # count=1 00:06:20.750 23:40:08 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@66 -- # echo 1 00:06:20.750 23:40:08 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@84 -- # count=1 00:06:20.751 23:40:08 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@85 -- # '[' 1 -ne 1 ']' 00:06:20.751 23:40:08 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@89 -- # raid_unmap_data_verify /dev/nbd0 00:06:20.751 23:40:08 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@17 -- # hash blkdiscard 00:06:20.751 23:40:08 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@18 -- # local nbd=/dev/nbd0 00:06:20.751 23:40:08 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@19 -- # local blksize 00:06:20.751 23:40:08 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@20 -- # lsblk -o LOG-SEC /dev/nbd0 00:06:20.751 23:40:08 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@20 -- # grep -v LOG-SEC 00:06:20.751 23:40:08 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@20 -- # cut -d ' ' -f 5 00:06:20.751 23:40:08 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@20 -- # blksize=512 00:06:20.751 23:40:08 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@21 -- # local rw_blk_num=4096 00:06:20.751 23:40:08 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@22 -- # local rw_len=2097152 00:06:20.751 23:40:08 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@23 -- # unmap_blk_offs=('0' '1028' '321') 00:06:20.751 23:40:08 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@23 -- # local unmap_blk_offs 00:06:20.751 23:40:08 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@24 -- # unmap_blk_nums=('128' '2035' '456') 00:06:20.751 23:40:08 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@24 -- # local unmap_blk_nums 00:06:20.751 23:40:08 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@25 -- # local unmap_off 00:06:20.751 23:40:08 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@26 -- # local unmap_len 00:06:20.751 23:40:08 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@29 -- # dd if=/dev/urandom of=/raidtest/raidrandtest bs=512 count=4096 00:06:20.751 4096+0 records in 00:06:20.751 4096+0 records out 00:06:20.751 2097152 bytes (2.1 MB, 2.0 MiB) copied, 0.0333621 s, 62.9 MB/s 00:06:20.751 23:40:08 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@30 -- # dd if=/raidtest/raidrandtest of=/dev/nbd0 bs=512 count=4096 oflag=direct 00:06:21.041 4096+0 records in 00:06:21.041 4096+0 records out 00:06:21.041 2097152 bytes (2.1 MB, 2.0 MiB) copied, 0.178351 s, 11.8 MB/s 00:06:21.041 23:40:09 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@31 -- # blockdev --flushbufs /dev/nbd0 00:06:21.041 23:40:09 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@34 -- # cmp -b -n 2097152 /raidtest/raidrandtest /dev/nbd0 00:06:21.041 23:40:09 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@36 -- # (( i = 0 )) 00:06:21.041 23:40:09 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@36 -- # (( i < 3 )) 00:06:21.041 23:40:09 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@37 -- # unmap_off=0 00:06:21.041 23:40:09 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@38 -- # unmap_len=65536 00:06:21.041 23:40:09 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@41 -- # dd if=/dev/zero of=/raidtest/raidrandtest bs=512 seek=0 count=128 conv=notrunc 00:06:21.041 128+0 records in 00:06:21.041 128+0 records out 00:06:21.041 65536 bytes (66 kB, 64 KiB) copied, 0.00107598 s, 60.9 MB/s 00:06:21.041 23:40:09 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@44 -- # blkdiscard -o 0 -l 65536 /dev/nbd0 00:06:21.041 23:40:09 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@45 -- # blockdev --flushbufs /dev/nbd0 00:06:21.041 23:40:09 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@48 -- # cmp -b -n 2097152 /raidtest/raidrandtest /dev/nbd0 00:06:21.041 23:40:09 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@36 -- # (( i++ )) 00:06:21.041 23:40:09 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@36 -- # (( i < 3 )) 00:06:21.041 23:40:09 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@37 -- # unmap_off=526336 00:06:21.041 23:40:09 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@38 -- # unmap_len=1041920 00:06:21.041 23:40:09 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@41 -- # dd if=/dev/zero of=/raidtest/raidrandtest bs=512 seek=1028 count=2035 conv=notrunc 00:06:21.041 2035+0 records in 00:06:21.041 2035+0 records out 00:06:21.041 1041920 bytes (1.0 MB, 1018 KiB) copied, 0.0122718 s, 84.9 MB/s 00:06:21.041 23:40:09 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@44 -- # blkdiscard -o 526336 -l 1041920 /dev/nbd0 00:06:21.041 23:40:09 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@45 -- # blockdev --flushbufs /dev/nbd0 00:06:21.041 23:40:09 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@48 -- # cmp -b -n 2097152 /raidtest/raidrandtest /dev/nbd0 00:06:21.041 23:40:09 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@36 -- # (( i++ )) 00:06:21.041 23:40:09 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@36 -- # (( i < 3 )) 00:06:21.041 23:40:09 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@37 -- # unmap_off=164352 00:06:21.041 23:40:09 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@38 -- # unmap_len=233472 00:06:21.041 23:40:09 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@41 -- # dd if=/dev/zero of=/raidtest/raidrandtest bs=512 seek=321 count=456 conv=notrunc 00:06:21.041 456+0 records in 00:06:21.041 456+0 records out 00:06:21.041 233472 bytes (233 kB, 228 KiB) copied, 0.00359489 s, 64.9 MB/s 00:06:21.041 23:40:09 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@44 -- # blkdiscard -o 164352 -l 233472 /dev/nbd0 00:06:21.041 23:40:09 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@45 -- # blockdev --flushbufs /dev/nbd0 00:06:21.041 23:40:09 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@48 -- # cmp -b -n 2097152 /raidtest/raidrandtest /dev/nbd0 00:06:21.041 23:40:09 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@36 -- # (( i++ )) 00:06:21.041 23:40:09 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@36 -- # (( i < 3 )) 00:06:21.041 23:40:09 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@52 -- # return 0 00:06:21.041 23:40:09 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@91 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:06:21.041 23:40:09 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:06:21.041 23:40:09 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:06:21.041 23:40:09 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@50 -- # local nbd_list 00:06:21.041 23:40:09 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@51 -- # local i 00:06:21.041 23:40:09 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:06:21.041 23:40:09 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:06:21.305 23:40:09 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:06:21.305 [2024-11-26 23:40:09.352322] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:06:21.305 23:40:09 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:06:21.305 23:40:09 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:06:21.305 23:40:09 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:06:21.305 23:40:09 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:06:21.305 23:40:09 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:06:21.305 23:40:09 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@41 -- # break 00:06:21.305 23:40:09 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@45 -- # return 0 00:06:21.305 23:40:09 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@92 -- # nbd_get_count /var/tmp/spdk.sock 00:06:21.305 23:40:09 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk.sock 00:06:21.305 23:40:09 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_get_disks 00:06:21.564 23:40:09 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:06:21.564 23:40:09 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@64 -- # echo '[]' 00:06:21.564 23:40:09 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:06:21.564 23:40:09 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:06:21.564 23:40:09 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@65 -- # echo '' 00:06:21.564 23:40:09 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:06:21.564 23:40:09 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@65 -- # true 00:06:21.564 23:40:09 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@65 -- # count=0 00:06:21.564 23:40:09 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@66 -- # echo 0 00:06:21.564 23:40:09 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@92 -- # count=0 00:06:21.564 23:40:09 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@93 -- # '[' 0 -ne 0 ']' 00:06:21.564 23:40:09 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@97 -- # killprocess 71630 00:06:21.564 23:40:09 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@954 -- # '[' -z 71630 ']' 00:06:21.564 23:40:09 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@958 -- # kill -0 71630 00:06:21.564 23:40:09 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@959 -- # uname 00:06:21.564 23:40:09 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:06:21.564 23:40:09 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 71630 00:06:21.564 23:40:09 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:06:21.564 23:40:09 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:06:21.564 23:40:09 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@972 -- # echo 'killing process with pid 71630' 00:06:21.564 killing process with pid 71630 00:06:21.564 23:40:09 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@973 -- # kill 71630 00:06:21.564 [2024-11-26 23:40:09.661818] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:06:21.564 [2024-11-26 23:40:09.661940] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:06:21.564 23:40:09 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@978 -- # wait 71630 00:06:21.564 [2024-11-26 23:40:09.661998] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:06:21.564 [2024-11-26 23:40:09.662010] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid, state offline 00:06:21.564 [2024-11-26 23:40:09.684197] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:06:21.823 23:40:09 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@99 -- # return 0 00:06:21.823 00:06:21.823 real 0m2.561s 00:06:21.823 user 0m3.190s 00:06:21.823 sys 0m0.855s 00:06:21.823 23:40:09 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:21.823 23:40:09 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@10 -- # set +x 00:06:21.823 ************************************ 00:06:21.823 END TEST raid_function_test_concat 00:06:21.823 ************************************ 00:06:21.823 23:40:09 bdev_raid -- bdev/bdev_raid.sh@963 -- # run_test raid0_resize_test raid_resize_test 0 00:06:21.823 23:40:09 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:06:21.823 23:40:09 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:21.823 23:40:09 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:06:22.083 ************************************ 00:06:22.083 START TEST raid0_resize_test 00:06:22.083 ************************************ 00:06:22.083 23:40:09 bdev_raid.raid0_resize_test -- common/autotest_common.sh@1129 -- # raid_resize_test 0 00:06:22.083 23:40:09 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@332 -- # local raid_level=0 00:06:22.083 23:40:09 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@333 -- # local blksize=512 00:06:22.083 23:40:09 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@334 -- # local bdev_size_mb=32 00:06:22.083 23:40:09 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@335 -- # local new_bdev_size_mb=64 00:06:22.083 23:40:09 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@336 -- # local blkcnt 00:06:22.083 23:40:09 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@337 -- # local raid_size_mb 00:06:22.083 23:40:09 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@338 -- # local new_raid_size_mb 00:06:22.083 23:40:09 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@339 -- # local expected_size 00:06:22.083 23:40:09 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@342 -- # raid_pid=71742 00:06:22.083 23:40:09 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@341 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:06:22.083 Process raid pid: 71742 00:06:22.083 23:40:09 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@343 -- # echo 'Process raid pid: 71742' 00:06:22.083 23:40:09 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@344 -- # waitforlisten 71742 00:06:22.083 23:40:09 bdev_raid.raid0_resize_test -- common/autotest_common.sh@835 -- # '[' -z 71742 ']' 00:06:22.083 23:40:09 bdev_raid.raid0_resize_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:22.083 23:40:09 bdev_raid.raid0_resize_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:22.083 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:22.083 23:40:09 bdev_raid.raid0_resize_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:22.083 23:40:09 bdev_raid.raid0_resize_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:22.083 23:40:09 bdev_raid.raid0_resize_test -- common/autotest_common.sh@10 -- # set +x 00:06:22.083 [2024-11-26 23:40:10.044458] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:06:22.083 [2024-11-26 23:40:10.044586] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:06:22.083 [2024-11-26 23:40:10.199666] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:22.343 [2024-11-26 23:40:10.224346] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:22.343 [2024-11-26 23:40:10.265928] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:06:22.343 [2024-11-26 23:40:10.265958] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:06:22.912 23:40:10 bdev_raid.raid0_resize_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:22.912 23:40:10 bdev_raid.raid0_resize_test -- common/autotest_common.sh@868 -- # return 0 00:06:22.912 23:40:10 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@346 -- # rpc_cmd bdev_null_create Base_1 32 512 00:06:22.912 23:40:10 bdev_raid.raid0_resize_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:22.912 23:40:10 bdev_raid.raid0_resize_test -- common/autotest_common.sh@10 -- # set +x 00:06:22.912 Base_1 00:06:22.912 23:40:10 bdev_raid.raid0_resize_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:22.912 23:40:10 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@347 -- # rpc_cmd bdev_null_create Base_2 32 512 00:06:22.912 23:40:10 bdev_raid.raid0_resize_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:22.912 23:40:10 bdev_raid.raid0_resize_test -- common/autotest_common.sh@10 -- # set +x 00:06:22.912 Base_2 00:06:22.912 23:40:10 bdev_raid.raid0_resize_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:22.912 23:40:10 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@349 -- # '[' 0 -eq 0 ']' 00:06:22.912 23:40:10 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@350 -- # rpc_cmd bdev_raid_create -z 64 -r 0 -b ''\''Base_1 Base_2'\''' -n Raid 00:06:22.912 23:40:10 bdev_raid.raid0_resize_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:22.912 23:40:10 bdev_raid.raid0_resize_test -- common/autotest_common.sh@10 -- # set +x 00:06:22.912 [2024-11-26 23:40:10.896030] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev Base_1 is claimed 00:06:22.912 [2024-11-26 23:40:10.897868] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev Base_2 is claimed 00:06:22.912 [2024-11-26 23:40:10.897935] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:06:22.912 [2024-11-26 23:40:10.897947] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 131072, blocklen 512 00:06:22.912 [2024-11-26 23:40:10.898205] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000021f0 00:06:22.912 [2024-11-26 23:40:10.898336] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:06:22.912 [2024-11-26 23:40:10.898360] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Raid, raid_bdev 0x617000001200 00:06:22.912 [2024-11-26 23:40:10.898470] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:06:22.912 23:40:10 bdev_raid.raid0_resize_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:22.912 23:40:10 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@356 -- # rpc_cmd bdev_null_resize Base_1 64 00:06:22.912 23:40:10 bdev_raid.raid0_resize_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:22.912 23:40:10 bdev_raid.raid0_resize_test -- common/autotest_common.sh@10 -- # set +x 00:06:22.912 [2024-11-26 23:40:10.908022] bdev_raid.c:2317:raid_bdev_resize_base_bdev: *DEBUG*: raid_bdev_resize_base_bdev 00:06:22.912 [2024-11-26 23:40:10.908051] bdev_raid.c:2330:raid_bdev_resize_base_bdev: *NOTICE*: base_bdev 'Base_1' was resized: old size 65536, new size 131072 00:06:22.912 true 00:06:22.912 23:40:10 bdev_raid.raid0_resize_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:22.912 23:40:10 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@359 -- # rpc_cmd bdev_get_bdevs -b Raid 00:06:22.912 23:40:10 bdev_raid.raid0_resize_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:22.913 23:40:10 bdev_raid.raid0_resize_test -- common/autotest_common.sh@10 -- # set +x 00:06:22.913 23:40:10 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@359 -- # jq '.[].num_blocks' 00:06:22.913 [2024-11-26 23:40:10.920167] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:06:22.913 23:40:10 bdev_raid.raid0_resize_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:22.913 23:40:10 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@359 -- # blkcnt=131072 00:06:22.913 23:40:10 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@360 -- # raid_size_mb=64 00:06:22.913 23:40:10 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@361 -- # '[' 0 -eq 0 ']' 00:06:22.913 23:40:10 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@362 -- # expected_size=64 00:06:22.913 23:40:10 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@366 -- # '[' 64 '!=' 64 ']' 00:06:22.913 23:40:10 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@372 -- # rpc_cmd bdev_null_resize Base_2 64 00:06:22.913 23:40:10 bdev_raid.raid0_resize_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:22.913 23:40:10 bdev_raid.raid0_resize_test -- common/autotest_common.sh@10 -- # set +x 00:06:22.913 [2024-11-26 23:40:10.967881] bdev_raid.c:2317:raid_bdev_resize_base_bdev: *DEBUG*: raid_bdev_resize_base_bdev 00:06:22.913 [2024-11-26 23:40:10.967906] bdev_raid.c:2330:raid_bdev_resize_base_bdev: *NOTICE*: base_bdev 'Base_2' was resized: old size 65536, new size 131072 00:06:22.913 [2024-11-26 23:40:10.967929] bdev_raid.c:2344:raid_bdev_resize_base_bdev: *NOTICE*: raid bdev 'Raid': block count was changed from 131072 to 262144 00:06:22.913 true 00:06:22.913 23:40:10 bdev_raid.raid0_resize_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:22.913 23:40:10 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@375 -- # jq '.[].num_blocks' 00:06:22.913 23:40:10 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@375 -- # rpc_cmd bdev_get_bdevs -b Raid 00:06:22.913 23:40:10 bdev_raid.raid0_resize_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:22.913 23:40:10 bdev_raid.raid0_resize_test -- common/autotest_common.sh@10 -- # set +x 00:06:22.913 [2024-11-26 23:40:10.984050] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:06:22.913 23:40:11 bdev_raid.raid0_resize_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:22.913 23:40:11 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@375 -- # blkcnt=262144 00:06:22.913 23:40:11 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@376 -- # raid_size_mb=128 00:06:22.913 23:40:11 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@377 -- # '[' 0 -eq 0 ']' 00:06:22.913 23:40:11 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@378 -- # expected_size=128 00:06:22.913 23:40:11 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@382 -- # '[' 128 '!=' 128 ']' 00:06:22.913 23:40:11 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@387 -- # killprocess 71742 00:06:22.913 23:40:11 bdev_raid.raid0_resize_test -- common/autotest_common.sh@954 -- # '[' -z 71742 ']' 00:06:22.913 23:40:11 bdev_raid.raid0_resize_test -- common/autotest_common.sh@958 -- # kill -0 71742 00:06:22.913 23:40:11 bdev_raid.raid0_resize_test -- common/autotest_common.sh@959 -- # uname 00:06:22.913 23:40:11 bdev_raid.raid0_resize_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:06:22.913 23:40:11 bdev_raid.raid0_resize_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 71742 00:06:23.172 23:40:11 bdev_raid.raid0_resize_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:06:23.172 23:40:11 bdev_raid.raid0_resize_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:06:23.172 killing process with pid 71742 00:06:23.172 23:40:11 bdev_raid.raid0_resize_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 71742' 00:06:23.172 23:40:11 bdev_raid.raid0_resize_test -- common/autotest_common.sh@973 -- # kill 71742 00:06:23.172 [2024-11-26 23:40:11.051657] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:06:23.172 [2024-11-26 23:40:11.051779] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:06:23.172 [2024-11-26 23:40:11.051833] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to fr 23:40:11 bdev_raid.raid0_resize_test -- common/autotest_common.sh@978 -- # wait 71742 00:06:23.172 ee all in destruct 00:06:23.172 [2024-11-26 23:40:11.051847] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Raid, state offline 00:06:23.172 [2024-11-26 23:40:11.053328] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:06:23.172 23:40:11 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@389 -- # return 0 00:06:23.172 00:06:23.172 real 0m1.300s 00:06:23.172 user 0m1.469s 00:06:23.172 sys 0m0.279s 00:06:23.172 23:40:11 bdev_raid.raid0_resize_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:23.172 23:40:11 bdev_raid.raid0_resize_test -- common/autotest_common.sh@10 -- # set +x 00:06:23.172 ************************************ 00:06:23.172 END TEST raid0_resize_test 00:06:23.172 ************************************ 00:06:23.432 23:40:11 bdev_raid -- bdev/bdev_raid.sh@964 -- # run_test raid1_resize_test raid_resize_test 1 00:06:23.432 23:40:11 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:06:23.432 23:40:11 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:23.432 23:40:11 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:06:23.432 ************************************ 00:06:23.432 START TEST raid1_resize_test 00:06:23.432 ************************************ 00:06:23.432 23:40:11 bdev_raid.raid1_resize_test -- common/autotest_common.sh@1129 -- # raid_resize_test 1 00:06:23.432 23:40:11 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@332 -- # local raid_level=1 00:06:23.432 23:40:11 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@333 -- # local blksize=512 00:06:23.432 23:40:11 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@334 -- # local bdev_size_mb=32 00:06:23.432 23:40:11 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@335 -- # local new_bdev_size_mb=64 00:06:23.432 23:40:11 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@336 -- # local blkcnt 00:06:23.432 23:40:11 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@337 -- # local raid_size_mb 00:06:23.432 23:40:11 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@338 -- # local new_raid_size_mb 00:06:23.432 23:40:11 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@339 -- # local expected_size 00:06:23.432 23:40:11 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@342 -- # raid_pid=71791 00:06:23.432 23:40:11 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@341 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:06:23.432 23:40:11 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@343 -- # echo 'Process raid pid: 71791' 00:06:23.432 Process raid pid: 71791 00:06:23.432 23:40:11 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@344 -- # waitforlisten 71791 00:06:23.432 23:40:11 bdev_raid.raid1_resize_test -- common/autotest_common.sh@835 -- # '[' -z 71791 ']' 00:06:23.432 23:40:11 bdev_raid.raid1_resize_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:23.432 23:40:11 bdev_raid.raid1_resize_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:23.432 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:23.432 23:40:11 bdev_raid.raid1_resize_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:23.432 23:40:11 bdev_raid.raid1_resize_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:23.432 23:40:11 bdev_raid.raid1_resize_test -- common/autotest_common.sh@10 -- # set +x 00:06:23.432 [2024-11-26 23:40:11.410316] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:06:23.432 [2024-11-26 23:40:11.410469] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:06:23.693 [2024-11-26 23:40:11.565169] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:23.693 [2024-11-26 23:40:11.589627] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:23.693 [2024-11-26 23:40:11.630984] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:06:23.693 [2024-11-26 23:40:11.631023] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:06:24.262 23:40:12 bdev_raid.raid1_resize_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:24.262 23:40:12 bdev_raid.raid1_resize_test -- common/autotest_common.sh@868 -- # return 0 00:06:24.262 23:40:12 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@346 -- # rpc_cmd bdev_null_create Base_1 32 512 00:06:24.262 23:40:12 bdev_raid.raid1_resize_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:24.262 23:40:12 bdev_raid.raid1_resize_test -- common/autotest_common.sh@10 -- # set +x 00:06:24.262 Base_1 00:06:24.262 23:40:12 bdev_raid.raid1_resize_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:24.262 23:40:12 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@347 -- # rpc_cmd bdev_null_create Base_2 32 512 00:06:24.262 23:40:12 bdev_raid.raid1_resize_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:24.262 23:40:12 bdev_raid.raid1_resize_test -- common/autotest_common.sh@10 -- # set +x 00:06:24.262 Base_2 00:06:24.262 23:40:12 bdev_raid.raid1_resize_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:24.262 23:40:12 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@349 -- # '[' 1 -eq 0 ']' 00:06:24.262 23:40:12 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@352 -- # rpc_cmd bdev_raid_create -r 1 -b ''\''Base_1 Base_2'\''' -n Raid 00:06:24.262 23:40:12 bdev_raid.raid1_resize_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:24.262 23:40:12 bdev_raid.raid1_resize_test -- common/autotest_common.sh@10 -- # set +x 00:06:24.262 [2024-11-26 23:40:12.256625] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev Base_1 is claimed 00:06:24.262 [2024-11-26 23:40:12.258396] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev Base_2 is claimed 00:06:24.262 [2024-11-26 23:40:12.258469] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:06:24.262 [2024-11-26 23:40:12.258480] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:06:24.262 [2024-11-26 23:40:12.258737] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000021f0 00:06:24.262 [2024-11-26 23:40:12.258854] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:06:24.262 [2024-11-26 23:40:12.258869] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Raid, raid_bdev 0x617000001200 00:06:24.262 [2024-11-26 23:40:12.258979] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:06:24.262 23:40:12 bdev_raid.raid1_resize_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:24.262 23:40:12 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@356 -- # rpc_cmd bdev_null_resize Base_1 64 00:06:24.262 23:40:12 bdev_raid.raid1_resize_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:24.262 23:40:12 bdev_raid.raid1_resize_test -- common/autotest_common.sh@10 -- # set +x 00:06:24.262 [2024-11-26 23:40:12.268603] bdev_raid.c:2317:raid_bdev_resize_base_bdev: *DEBUG*: raid_bdev_resize_base_bdev 00:06:24.262 [2024-11-26 23:40:12.268639] bdev_raid.c:2330:raid_bdev_resize_base_bdev: *NOTICE*: base_bdev 'Base_1' was resized: old size 65536, new size 131072 00:06:24.262 true 00:06:24.262 23:40:12 bdev_raid.raid1_resize_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:24.262 23:40:12 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@359 -- # rpc_cmd bdev_get_bdevs -b Raid 00:06:24.262 23:40:12 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@359 -- # jq '.[].num_blocks' 00:06:24.262 23:40:12 bdev_raid.raid1_resize_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:24.262 23:40:12 bdev_raid.raid1_resize_test -- common/autotest_common.sh@10 -- # set +x 00:06:24.262 [2024-11-26 23:40:12.284756] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:06:24.262 23:40:12 bdev_raid.raid1_resize_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:24.262 23:40:12 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@359 -- # blkcnt=65536 00:06:24.262 23:40:12 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@360 -- # raid_size_mb=32 00:06:24.262 23:40:12 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@361 -- # '[' 1 -eq 0 ']' 00:06:24.262 23:40:12 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@364 -- # expected_size=32 00:06:24.262 23:40:12 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@366 -- # '[' 32 '!=' 32 ']' 00:06:24.262 23:40:12 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@372 -- # rpc_cmd bdev_null_resize Base_2 64 00:06:24.262 23:40:12 bdev_raid.raid1_resize_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:24.262 23:40:12 bdev_raid.raid1_resize_test -- common/autotest_common.sh@10 -- # set +x 00:06:24.262 [2024-11-26 23:40:12.332473] bdev_raid.c:2317:raid_bdev_resize_base_bdev: *DEBUG*: raid_bdev_resize_base_bdev 00:06:24.262 [2024-11-26 23:40:12.332500] bdev_raid.c:2330:raid_bdev_resize_base_bdev: *NOTICE*: base_bdev 'Base_2' was resized: old size 65536, new size 131072 00:06:24.262 [2024-11-26 23:40:12.332527] bdev_raid.c:2344:raid_bdev_resize_base_bdev: *NOTICE*: raid bdev 'Raid': block count was changed from 65536 to 131072 00:06:24.262 true 00:06:24.262 23:40:12 bdev_raid.raid1_resize_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:24.262 23:40:12 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@375 -- # jq '.[].num_blocks' 00:06:24.262 23:40:12 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@375 -- # rpc_cmd bdev_get_bdevs -b Raid 00:06:24.262 23:40:12 bdev_raid.raid1_resize_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:24.262 23:40:12 bdev_raid.raid1_resize_test -- common/autotest_common.sh@10 -- # set +x 00:06:24.262 [2024-11-26 23:40:12.348625] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:06:24.262 23:40:12 bdev_raid.raid1_resize_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:24.262 23:40:12 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@375 -- # blkcnt=131072 00:06:24.262 23:40:12 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@376 -- # raid_size_mb=64 00:06:24.262 23:40:12 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@377 -- # '[' 1 -eq 0 ']' 00:06:24.262 23:40:12 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@380 -- # expected_size=64 00:06:24.262 23:40:12 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@382 -- # '[' 64 '!=' 64 ']' 00:06:24.262 23:40:12 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@387 -- # killprocess 71791 00:06:24.262 23:40:12 bdev_raid.raid1_resize_test -- common/autotest_common.sh@954 -- # '[' -z 71791 ']' 00:06:24.262 23:40:12 bdev_raid.raid1_resize_test -- common/autotest_common.sh@958 -- # kill -0 71791 00:06:24.262 23:40:12 bdev_raid.raid1_resize_test -- common/autotest_common.sh@959 -- # uname 00:06:24.262 23:40:12 bdev_raid.raid1_resize_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:06:24.262 23:40:12 bdev_raid.raid1_resize_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 71791 00:06:24.522 23:40:12 bdev_raid.raid1_resize_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:06:24.522 23:40:12 bdev_raid.raid1_resize_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:06:24.522 killing process with pid 71791 00:06:24.522 23:40:12 bdev_raid.raid1_resize_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 71791' 00:06:24.522 23:40:12 bdev_raid.raid1_resize_test -- common/autotest_common.sh@973 -- # kill 71791 00:06:24.522 [2024-11-26 23:40:12.418303] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:06:24.522 [2024-11-26 23:40:12.418407] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:06:24.522 23:40:12 bdev_raid.raid1_resize_test -- common/autotest_common.sh@978 -- # wait 71791 00:06:24.522 [2024-11-26 23:40:12.418823] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:06:24.522 [2024-11-26 23:40:12.418846] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Raid, state offline 00:06:24.522 [2024-11-26 23:40:12.419963] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:06:24.522 23:40:12 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@389 -- # return 0 00:06:24.522 00:06:24.522 real 0m1.300s 00:06:24.522 user 0m1.460s 00:06:24.522 sys 0m0.289s 00:06:24.522 23:40:12 bdev_raid.raid1_resize_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:24.522 23:40:12 bdev_raid.raid1_resize_test -- common/autotest_common.sh@10 -- # set +x 00:06:24.522 ************************************ 00:06:24.522 END TEST raid1_resize_test 00:06:24.522 ************************************ 00:06:24.783 23:40:12 bdev_raid -- bdev/bdev_raid.sh@966 -- # for n in {2..4} 00:06:24.783 23:40:12 bdev_raid -- bdev/bdev_raid.sh@967 -- # for level in raid0 concat raid1 00:06:24.783 23:40:12 bdev_raid -- bdev/bdev_raid.sh@968 -- # run_test raid_state_function_test raid_state_function_test raid0 2 false 00:06:24.783 23:40:12 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:06:24.783 23:40:12 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:24.783 23:40:12 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:06:24.783 ************************************ 00:06:24.783 START TEST raid_state_function_test 00:06:24.783 ************************************ 00:06:24.783 23:40:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1129 -- # raid_state_function_test raid0 2 false 00:06:24.783 23:40:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=raid0 00:06:24.783 23:40:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=2 00:06:24.783 23:40:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:06:24.783 23:40:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:06:24.783 23:40:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:06:24.783 23:40:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:06:24.783 23:40:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:06:24.783 23:40:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:06:24.783 23:40:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:06:24.783 23:40:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:06:24.783 23:40:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:06:24.783 23:40:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:06:24.783 23:40:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:06:24.783 23:40:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:06:24.783 23:40:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:06:24.783 23:40:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:06:24.783 23:40:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:06:24.783 23:40:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:06:24.783 23:40:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' raid0 '!=' raid1 ']' 00:06:24.783 23:40:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:06:24.783 23:40:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:06:24.783 23:40:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:06:24.783 23:40:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:06:24.783 23:40:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=71838 00:06:24.783 23:40:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:06:24.783 Process raid pid: 71838 00:06:24.783 23:40:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 71838' 00:06:24.783 23:40:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 71838 00:06:24.783 23:40:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@835 -- # '[' -z 71838 ']' 00:06:24.783 23:40:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:24.783 23:40:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:24.783 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:24.783 23:40:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:24.783 23:40:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:24.783 23:40:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:06:24.783 [2024-11-26 23:40:12.784181] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:06:24.783 [2024-11-26 23:40:12.784328] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:06:25.043 [2024-11-26 23:40:12.939903] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:25.043 [2024-11-26 23:40:12.964762] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:25.043 [2024-11-26 23:40:13.006203] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:06:25.043 [2024-11-26 23:40:13.006244] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:06:25.615 23:40:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:25.615 23:40:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@868 -- # return 0 00:06:25.615 23:40:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:06:25.615 23:40:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:25.615 23:40:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:06:25.615 [2024-11-26 23:40:13.611948] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:06:25.615 [2024-11-26 23:40:13.612002] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:06:25.615 [2024-11-26 23:40:13.612012] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:06:25.615 [2024-11-26 23:40:13.612022] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:06:25.616 23:40:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:25.616 23:40:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 2 00:06:25.616 23:40:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:06:25.616 23:40:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:06:25.616 23:40:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:06:25.616 23:40:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:06:25.616 23:40:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:06:25.616 23:40:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:06:25.616 23:40:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:06:25.616 23:40:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:06:25.616 23:40:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:06:25.616 23:40:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:06:25.616 23:40:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:25.616 23:40:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:06:25.616 23:40:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:06:25.616 23:40:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:25.616 23:40:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:06:25.616 "name": "Existed_Raid", 00:06:25.616 "uuid": "00000000-0000-0000-0000-000000000000", 00:06:25.616 "strip_size_kb": 64, 00:06:25.616 "state": "configuring", 00:06:25.616 "raid_level": "raid0", 00:06:25.616 "superblock": false, 00:06:25.616 "num_base_bdevs": 2, 00:06:25.616 "num_base_bdevs_discovered": 0, 00:06:25.616 "num_base_bdevs_operational": 2, 00:06:25.616 "base_bdevs_list": [ 00:06:25.616 { 00:06:25.616 "name": "BaseBdev1", 00:06:25.616 "uuid": "00000000-0000-0000-0000-000000000000", 00:06:25.616 "is_configured": false, 00:06:25.616 "data_offset": 0, 00:06:25.616 "data_size": 0 00:06:25.616 }, 00:06:25.616 { 00:06:25.616 "name": "BaseBdev2", 00:06:25.616 "uuid": "00000000-0000-0000-0000-000000000000", 00:06:25.616 "is_configured": false, 00:06:25.616 "data_offset": 0, 00:06:25.616 "data_size": 0 00:06:25.616 } 00:06:25.616 ] 00:06:25.616 }' 00:06:25.616 23:40:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:06:25.616 23:40:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:06:26.188 23:40:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:06:26.188 23:40:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:26.188 23:40:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:06:26.188 [2024-11-26 23:40:14.043177] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:06:26.188 [2024-11-26 23:40:14.043222] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:06:26.188 23:40:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:26.188 23:40:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:06:26.188 23:40:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:26.188 23:40:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:06:26.188 [2024-11-26 23:40:14.051156] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:06:26.188 [2024-11-26 23:40:14.051200] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:06:26.188 [2024-11-26 23:40:14.051209] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:06:26.188 [2024-11-26 23:40:14.051227] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:06:26.188 23:40:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:26.188 23:40:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:06:26.188 23:40:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:26.188 23:40:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:06:26.188 [2024-11-26 23:40:14.067830] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:06:26.188 BaseBdev1 00:06:26.188 23:40:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:26.188 23:40:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:06:26.188 23:40:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:06:26.188 23:40:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:06:26.188 23:40:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:06:26.188 23:40:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:06:26.188 23:40:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:06:26.188 23:40:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:06:26.188 23:40:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:26.188 23:40:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:06:26.188 23:40:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:26.188 23:40:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:06:26.188 23:40:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:26.188 23:40:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:06:26.188 [ 00:06:26.188 { 00:06:26.188 "name": "BaseBdev1", 00:06:26.188 "aliases": [ 00:06:26.188 "0a5dfcf0-775a-45aa-83ef-ac97a198e867" 00:06:26.188 ], 00:06:26.188 "product_name": "Malloc disk", 00:06:26.188 "block_size": 512, 00:06:26.188 "num_blocks": 65536, 00:06:26.188 "uuid": "0a5dfcf0-775a-45aa-83ef-ac97a198e867", 00:06:26.188 "assigned_rate_limits": { 00:06:26.188 "rw_ios_per_sec": 0, 00:06:26.188 "rw_mbytes_per_sec": 0, 00:06:26.188 "r_mbytes_per_sec": 0, 00:06:26.188 "w_mbytes_per_sec": 0 00:06:26.188 }, 00:06:26.188 "claimed": true, 00:06:26.188 "claim_type": "exclusive_write", 00:06:26.188 "zoned": false, 00:06:26.188 "supported_io_types": { 00:06:26.188 "read": true, 00:06:26.188 "write": true, 00:06:26.188 "unmap": true, 00:06:26.189 "flush": true, 00:06:26.189 "reset": true, 00:06:26.189 "nvme_admin": false, 00:06:26.189 "nvme_io": false, 00:06:26.189 "nvme_io_md": false, 00:06:26.189 "write_zeroes": true, 00:06:26.189 "zcopy": true, 00:06:26.189 "get_zone_info": false, 00:06:26.189 "zone_management": false, 00:06:26.189 "zone_append": false, 00:06:26.189 "compare": false, 00:06:26.189 "compare_and_write": false, 00:06:26.189 "abort": true, 00:06:26.189 "seek_hole": false, 00:06:26.189 "seek_data": false, 00:06:26.189 "copy": true, 00:06:26.189 "nvme_iov_md": false 00:06:26.189 }, 00:06:26.189 "memory_domains": [ 00:06:26.189 { 00:06:26.189 "dma_device_id": "system", 00:06:26.189 "dma_device_type": 1 00:06:26.189 }, 00:06:26.189 { 00:06:26.189 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:06:26.189 "dma_device_type": 2 00:06:26.189 } 00:06:26.189 ], 00:06:26.189 "driver_specific": {} 00:06:26.189 } 00:06:26.189 ] 00:06:26.189 23:40:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:26.189 23:40:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:06:26.189 23:40:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 2 00:06:26.189 23:40:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:06:26.189 23:40:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:06:26.189 23:40:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:06:26.189 23:40:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:06:26.189 23:40:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:06:26.189 23:40:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:06:26.189 23:40:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:06:26.189 23:40:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:06:26.189 23:40:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:06:26.189 23:40:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:06:26.189 23:40:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:06:26.189 23:40:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:26.189 23:40:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:06:26.189 23:40:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:26.189 23:40:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:06:26.189 "name": "Existed_Raid", 00:06:26.189 "uuid": "00000000-0000-0000-0000-000000000000", 00:06:26.189 "strip_size_kb": 64, 00:06:26.189 "state": "configuring", 00:06:26.189 "raid_level": "raid0", 00:06:26.189 "superblock": false, 00:06:26.189 "num_base_bdevs": 2, 00:06:26.189 "num_base_bdevs_discovered": 1, 00:06:26.189 "num_base_bdevs_operational": 2, 00:06:26.189 "base_bdevs_list": [ 00:06:26.189 { 00:06:26.189 "name": "BaseBdev1", 00:06:26.189 "uuid": "0a5dfcf0-775a-45aa-83ef-ac97a198e867", 00:06:26.189 "is_configured": true, 00:06:26.189 "data_offset": 0, 00:06:26.189 "data_size": 65536 00:06:26.189 }, 00:06:26.189 { 00:06:26.189 "name": "BaseBdev2", 00:06:26.189 "uuid": "00000000-0000-0000-0000-000000000000", 00:06:26.189 "is_configured": false, 00:06:26.189 "data_offset": 0, 00:06:26.189 "data_size": 0 00:06:26.189 } 00:06:26.189 ] 00:06:26.189 }' 00:06:26.189 23:40:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:06:26.189 23:40:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:06:26.448 23:40:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:06:26.448 23:40:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:26.448 23:40:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:06:26.449 [2024-11-26 23:40:14.547078] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:06:26.449 [2024-11-26 23:40:14.547133] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:06:26.449 23:40:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:26.449 23:40:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:06:26.449 23:40:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:26.449 23:40:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:06:26.449 [2024-11-26 23:40:14.555084] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:06:26.449 [2024-11-26 23:40:14.556951] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:06:26.449 [2024-11-26 23:40:14.556990] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:06:26.449 23:40:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:26.449 23:40:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:06:26.449 23:40:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:06:26.449 23:40:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 2 00:06:26.449 23:40:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:06:26.449 23:40:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:06:26.449 23:40:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:06:26.449 23:40:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:06:26.449 23:40:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:06:26.449 23:40:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:06:26.449 23:40:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:06:26.449 23:40:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:06:26.449 23:40:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:06:26.449 23:40:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:06:26.449 23:40:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:06:26.449 23:40:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:26.449 23:40:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:06:26.708 23:40:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:26.708 23:40:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:06:26.708 "name": "Existed_Raid", 00:06:26.708 "uuid": "00000000-0000-0000-0000-000000000000", 00:06:26.708 "strip_size_kb": 64, 00:06:26.708 "state": "configuring", 00:06:26.708 "raid_level": "raid0", 00:06:26.708 "superblock": false, 00:06:26.708 "num_base_bdevs": 2, 00:06:26.708 "num_base_bdevs_discovered": 1, 00:06:26.708 "num_base_bdevs_operational": 2, 00:06:26.708 "base_bdevs_list": [ 00:06:26.708 { 00:06:26.708 "name": "BaseBdev1", 00:06:26.708 "uuid": "0a5dfcf0-775a-45aa-83ef-ac97a198e867", 00:06:26.708 "is_configured": true, 00:06:26.708 "data_offset": 0, 00:06:26.708 "data_size": 65536 00:06:26.708 }, 00:06:26.708 { 00:06:26.708 "name": "BaseBdev2", 00:06:26.708 "uuid": "00000000-0000-0000-0000-000000000000", 00:06:26.708 "is_configured": false, 00:06:26.708 "data_offset": 0, 00:06:26.708 "data_size": 0 00:06:26.708 } 00:06:26.708 ] 00:06:26.708 }' 00:06:26.708 23:40:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:06:26.708 23:40:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:06:26.968 23:40:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:06:26.968 23:40:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:26.968 23:40:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:06:26.969 [2024-11-26 23:40:14.961204] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:06:26.969 [2024-11-26 23:40:14.961253] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:06:26.969 [2024-11-26 23:40:14.961261] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 131072, blocklen 512 00:06:26.969 [2024-11-26 23:40:14.961536] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:06:26.969 [2024-11-26 23:40:14.961721] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:06:26.969 [2024-11-26 23:40:14.961742] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:06:26.969 [2024-11-26 23:40:14.961972] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:06:26.969 BaseBdev2 00:06:26.969 23:40:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:26.969 23:40:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:06:26.969 23:40:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:06:26.969 23:40:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:06:26.969 23:40:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:06:26.969 23:40:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:06:26.969 23:40:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:06:26.969 23:40:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:06:26.969 23:40:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:26.969 23:40:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:06:26.969 23:40:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:26.969 23:40:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:06:26.969 23:40:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:26.969 23:40:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:06:26.969 [ 00:06:26.969 { 00:06:26.969 "name": "BaseBdev2", 00:06:26.969 "aliases": [ 00:06:26.969 "d52cdb2a-2ee6-4e64-a8a2-91a2ff17a347" 00:06:26.969 ], 00:06:26.969 "product_name": "Malloc disk", 00:06:26.969 "block_size": 512, 00:06:26.969 "num_blocks": 65536, 00:06:26.969 "uuid": "d52cdb2a-2ee6-4e64-a8a2-91a2ff17a347", 00:06:26.969 "assigned_rate_limits": { 00:06:26.969 "rw_ios_per_sec": 0, 00:06:26.969 "rw_mbytes_per_sec": 0, 00:06:26.969 "r_mbytes_per_sec": 0, 00:06:26.969 "w_mbytes_per_sec": 0 00:06:26.969 }, 00:06:26.969 "claimed": true, 00:06:26.969 "claim_type": "exclusive_write", 00:06:26.969 "zoned": false, 00:06:26.969 "supported_io_types": { 00:06:26.969 "read": true, 00:06:26.969 "write": true, 00:06:26.969 "unmap": true, 00:06:26.969 "flush": true, 00:06:26.969 "reset": true, 00:06:26.969 "nvme_admin": false, 00:06:26.969 "nvme_io": false, 00:06:26.969 "nvme_io_md": false, 00:06:26.969 "write_zeroes": true, 00:06:26.969 "zcopy": true, 00:06:26.969 "get_zone_info": false, 00:06:26.969 "zone_management": false, 00:06:26.969 "zone_append": false, 00:06:26.969 "compare": false, 00:06:26.969 "compare_and_write": false, 00:06:26.969 "abort": true, 00:06:26.969 "seek_hole": false, 00:06:26.969 "seek_data": false, 00:06:26.969 "copy": true, 00:06:26.969 "nvme_iov_md": false 00:06:26.969 }, 00:06:26.969 "memory_domains": [ 00:06:26.969 { 00:06:26.969 "dma_device_id": "system", 00:06:26.969 "dma_device_type": 1 00:06:26.969 }, 00:06:26.969 { 00:06:26.969 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:06:26.969 "dma_device_type": 2 00:06:26.969 } 00:06:26.969 ], 00:06:26.969 "driver_specific": {} 00:06:26.969 } 00:06:26.969 ] 00:06:26.969 23:40:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:26.969 23:40:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:06:26.969 23:40:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:06:26.969 23:40:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:06:26.969 23:40:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid0 64 2 00:06:26.969 23:40:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:06:26.969 23:40:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:06:26.969 23:40:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:06:26.969 23:40:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:06:26.969 23:40:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:06:26.969 23:40:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:06:26.969 23:40:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:06:26.969 23:40:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:06:26.969 23:40:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:06:26.969 23:40:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:06:26.969 23:40:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:06:26.969 23:40:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:26.969 23:40:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:06:26.969 23:40:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:26.969 23:40:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:06:26.969 "name": "Existed_Raid", 00:06:26.969 "uuid": "ac57b01b-488c-4cd2-b4b7-58761997a350", 00:06:26.969 "strip_size_kb": 64, 00:06:26.969 "state": "online", 00:06:26.969 "raid_level": "raid0", 00:06:26.969 "superblock": false, 00:06:26.969 "num_base_bdevs": 2, 00:06:26.969 "num_base_bdevs_discovered": 2, 00:06:26.969 "num_base_bdevs_operational": 2, 00:06:26.969 "base_bdevs_list": [ 00:06:26.969 { 00:06:26.969 "name": "BaseBdev1", 00:06:26.969 "uuid": "0a5dfcf0-775a-45aa-83ef-ac97a198e867", 00:06:26.969 "is_configured": true, 00:06:26.969 "data_offset": 0, 00:06:26.969 "data_size": 65536 00:06:26.969 }, 00:06:26.969 { 00:06:26.969 "name": "BaseBdev2", 00:06:26.969 "uuid": "d52cdb2a-2ee6-4e64-a8a2-91a2ff17a347", 00:06:26.969 "is_configured": true, 00:06:26.969 "data_offset": 0, 00:06:26.969 "data_size": 65536 00:06:26.969 } 00:06:26.969 ] 00:06:26.969 }' 00:06:26.969 23:40:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:06:26.969 23:40:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:06:27.538 23:40:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:06:27.538 23:40:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:06:27.538 23:40:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:06:27.538 23:40:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:06:27.538 23:40:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:06:27.538 23:40:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:06:27.538 23:40:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:06:27.538 23:40:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:27.538 23:40:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:06:27.538 23:40:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:06:27.538 [2024-11-26 23:40:15.436738] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:06:27.538 23:40:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:27.538 23:40:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:06:27.538 "name": "Existed_Raid", 00:06:27.538 "aliases": [ 00:06:27.538 "ac57b01b-488c-4cd2-b4b7-58761997a350" 00:06:27.538 ], 00:06:27.538 "product_name": "Raid Volume", 00:06:27.538 "block_size": 512, 00:06:27.538 "num_blocks": 131072, 00:06:27.538 "uuid": "ac57b01b-488c-4cd2-b4b7-58761997a350", 00:06:27.538 "assigned_rate_limits": { 00:06:27.538 "rw_ios_per_sec": 0, 00:06:27.538 "rw_mbytes_per_sec": 0, 00:06:27.538 "r_mbytes_per_sec": 0, 00:06:27.538 "w_mbytes_per_sec": 0 00:06:27.538 }, 00:06:27.538 "claimed": false, 00:06:27.538 "zoned": false, 00:06:27.538 "supported_io_types": { 00:06:27.538 "read": true, 00:06:27.538 "write": true, 00:06:27.538 "unmap": true, 00:06:27.538 "flush": true, 00:06:27.538 "reset": true, 00:06:27.538 "nvme_admin": false, 00:06:27.538 "nvme_io": false, 00:06:27.538 "nvme_io_md": false, 00:06:27.538 "write_zeroes": true, 00:06:27.538 "zcopy": false, 00:06:27.538 "get_zone_info": false, 00:06:27.538 "zone_management": false, 00:06:27.538 "zone_append": false, 00:06:27.538 "compare": false, 00:06:27.538 "compare_and_write": false, 00:06:27.538 "abort": false, 00:06:27.538 "seek_hole": false, 00:06:27.538 "seek_data": false, 00:06:27.538 "copy": false, 00:06:27.538 "nvme_iov_md": false 00:06:27.538 }, 00:06:27.538 "memory_domains": [ 00:06:27.538 { 00:06:27.538 "dma_device_id": "system", 00:06:27.538 "dma_device_type": 1 00:06:27.538 }, 00:06:27.538 { 00:06:27.538 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:06:27.538 "dma_device_type": 2 00:06:27.538 }, 00:06:27.538 { 00:06:27.538 "dma_device_id": "system", 00:06:27.538 "dma_device_type": 1 00:06:27.538 }, 00:06:27.538 { 00:06:27.538 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:06:27.538 "dma_device_type": 2 00:06:27.538 } 00:06:27.538 ], 00:06:27.538 "driver_specific": { 00:06:27.538 "raid": { 00:06:27.538 "uuid": "ac57b01b-488c-4cd2-b4b7-58761997a350", 00:06:27.538 "strip_size_kb": 64, 00:06:27.538 "state": "online", 00:06:27.538 "raid_level": "raid0", 00:06:27.538 "superblock": false, 00:06:27.539 "num_base_bdevs": 2, 00:06:27.539 "num_base_bdevs_discovered": 2, 00:06:27.539 "num_base_bdevs_operational": 2, 00:06:27.539 "base_bdevs_list": [ 00:06:27.539 { 00:06:27.539 "name": "BaseBdev1", 00:06:27.539 "uuid": "0a5dfcf0-775a-45aa-83ef-ac97a198e867", 00:06:27.539 "is_configured": true, 00:06:27.539 "data_offset": 0, 00:06:27.539 "data_size": 65536 00:06:27.539 }, 00:06:27.539 { 00:06:27.539 "name": "BaseBdev2", 00:06:27.539 "uuid": "d52cdb2a-2ee6-4e64-a8a2-91a2ff17a347", 00:06:27.539 "is_configured": true, 00:06:27.539 "data_offset": 0, 00:06:27.539 "data_size": 65536 00:06:27.539 } 00:06:27.539 ] 00:06:27.539 } 00:06:27.539 } 00:06:27.539 }' 00:06:27.539 23:40:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:06:27.539 23:40:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:06:27.539 BaseBdev2' 00:06:27.539 23:40:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:06:27.539 23:40:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:06:27.539 23:40:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:06:27.539 23:40:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:06:27.539 23:40:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:27.539 23:40:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:06:27.539 23:40:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:06:27.539 23:40:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:27.539 23:40:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:06:27.539 23:40:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:06:27.539 23:40:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:06:27.539 23:40:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:06:27.539 23:40:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:27.539 23:40:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:06:27.539 23:40:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:06:27.539 23:40:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:27.539 23:40:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:06:27.539 23:40:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:06:27.539 23:40:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:06:27.539 23:40:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:27.539 23:40:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:06:27.539 [2024-11-26 23:40:15.660085] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:06:27.539 [2024-11-26 23:40:15.660118] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:06:27.539 [2024-11-26 23:40:15.660180] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:06:27.799 23:40:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:27.799 23:40:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:06:27.799 23:40:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy raid0 00:06:27.799 23:40:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:06:27.799 23:40:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # return 1 00:06:27.799 23:40:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:06:27.799 23:40:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline raid0 64 1 00:06:27.799 23:40:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:06:27.799 23:40:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:06:27.799 23:40:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:06:27.799 23:40:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:06:27.799 23:40:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:06:27.799 23:40:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:06:27.799 23:40:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:06:27.799 23:40:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:06:27.799 23:40:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:06:27.799 23:40:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:06:27.799 23:40:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:27.799 23:40:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:06:27.799 23:40:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:06:27.800 23:40:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:27.800 23:40:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:06:27.800 "name": "Existed_Raid", 00:06:27.800 "uuid": "ac57b01b-488c-4cd2-b4b7-58761997a350", 00:06:27.800 "strip_size_kb": 64, 00:06:27.800 "state": "offline", 00:06:27.800 "raid_level": "raid0", 00:06:27.800 "superblock": false, 00:06:27.800 "num_base_bdevs": 2, 00:06:27.800 "num_base_bdevs_discovered": 1, 00:06:27.800 "num_base_bdevs_operational": 1, 00:06:27.800 "base_bdevs_list": [ 00:06:27.800 { 00:06:27.800 "name": null, 00:06:27.800 "uuid": "00000000-0000-0000-0000-000000000000", 00:06:27.800 "is_configured": false, 00:06:27.800 "data_offset": 0, 00:06:27.800 "data_size": 65536 00:06:27.800 }, 00:06:27.800 { 00:06:27.800 "name": "BaseBdev2", 00:06:27.800 "uuid": "d52cdb2a-2ee6-4e64-a8a2-91a2ff17a347", 00:06:27.800 "is_configured": true, 00:06:27.800 "data_offset": 0, 00:06:27.800 "data_size": 65536 00:06:27.800 } 00:06:27.800 ] 00:06:27.800 }' 00:06:27.800 23:40:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:06:27.800 23:40:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:06:28.061 23:40:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:06:28.061 23:40:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:06:28.061 23:40:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:06:28.061 23:40:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:28.061 23:40:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:06:28.061 23:40:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:06:28.061 23:40:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:28.061 23:40:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:06:28.061 23:40:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:06:28.061 23:40:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:06:28.061 23:40:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:28.061 23:40:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:06:28.061 [2024-11-26 23:40:16.138259] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:06:28.061 [2024-11-26 23:40:16.138318] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:06:28.061 23:40:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:28.061 23:40:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:06:28.061 23:40:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:06:28.061 23:40:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:06:28.061 23:40:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:28.061 23:40:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:06:28.061 23:40:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:06:28.061 23:40:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:28.321 23:40:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:06:28.321 23:40:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:06:28.321 23:40:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 2 -gt 2 ']' 00:06:28.321 23:40:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 71838 00:06:28.321 23:40:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # '[' -z 71838 ']' 00:06:28.321 23:40:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@958 -- # kill -0 71838 00:06:28.321 23:40:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # uname 00:06:28.321 23:40:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:06:28.321 23:40:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 71838 00:06:28.321 23:40:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:06:28.321 23:40:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:06:28.321 killing process with pid 71838 00:06:28.321 23:40:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 71838' 00:06:28.321 23:40:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@973 -- # kill 71838 00:06:28.321 [2024-11-26 23:40:16.242815] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:06:28.321 23:40:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@978 -- # wait 71838 00:06:28.321 [2024-11-26 23:40:16.243797] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:06:28.321 23:40:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:06:28.321 00:06:28.321 real 0m3.749s 00:06:28.321 user 0m5.967s 00:06:28.321 sys 0m0.698s 00:06:28.321 23:40:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:28.321 23:40:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:06:28.321 ************************************ 00:06:28.321 END TEST raid_state_function_test 00:06:28.321 ************************************ 00:06:28.582 23:40:16 bdev_raid -- bdev/bdev_raid.sh@969 -- # run_test raid_state_function_test_sb raid_state_function_test raid0 2 true 00:06:28.582 23:40:16 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:06:28.582 23:40:16 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:28.582 23:40:16 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:06:28.582 ************************************ 00:06:28.582 START TEST raid_state_function_test_sb 00:06:28.582 ************************************ 00:06:28.582 23:40:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1129 -- # raid_state_function_test raid0 2 true 00:06:28.582 23:40:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=raid0 00:06:28.582 23:40:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=2 00:06:28.582 23:40:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:06:28.582 23:40:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:06:28.582 23:40:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:06:28.582 23:40:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:06:28.582 23:40:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:06:28.582 23:40:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:06:28.582 23:40:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:06:28.582 23:40:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:06:28.582 23:40:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:06:28.582 23:40:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:06:28.582 23:40:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:06:28.582 23:40:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:06:28.582 23:40:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:06:28.582 23:40:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:06:28.582 23:40:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:06:28.582 23:40:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:06:28.582 23:40:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' raid0 '!=' raid1 ']' 00:06:28.582 23:40:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:06:28.582 23:40:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:06:28.582 23:40:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:06:28.582 23:40:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:06:28.582 23:40:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=72075 00:06:28.582 23:40:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:06:28.582 23:40:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 72075' 00:06:28.582 Process raid pid: 72075 00:06:28.582 23:40:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 72075 00:06:28.582 23:40:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@835 -- # '[' -z 72075 ']' 00:06:28.582 23:40:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:28.582 23:40:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:28.582 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:28.582 23:40:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:28.582 23:40:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:28.582 23:40:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:06:28.582 [2024-11-26 23:40:16.602026] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:06:28.582 [2024-11-26 23:40:16.602151] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:06:28.841 [2024-11-26 23:40:16.758969] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:28.841 [2024-11-26 23:40:16.783182] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:28.841 [2024-11-26 23:40:16.824284] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:06:28.842 [2024-11-26 23:40:16.824325] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:06:29.412 23:40:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:29.412 23:40:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@868 -- # return 0 00:06:29.412 23:40:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:06:29.412 23:40:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:29.412 23:40:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:06:29.412 [2024-11-26 23:40:17.426201] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:06:29.412 [2024-11-26 23:40:17.426253] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:06:29.412 [2024-11-26 23:40:17.426265] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:06:29.412 [2024-11-26 23:40:17.426278] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:06:29.412 23:40:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:29.412 23:40:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 2 00:06:29.412 23:40:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:06:29.412 23:40:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:06:29.412 23:40:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:06:29.412 23:40:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:06:29.412 23:40:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:06:29.412 23:40:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:06:29.412 23:40:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:06:29.413 23:40:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:06:29.413 23:40:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:06:29.413 23:40:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:06:29.413 23:40:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:06:29.413 23:40:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:29.413 23:40:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:06:29.413 23:40:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:29.413 23:40:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:06:29.413 "name": "Existed_Raid", 00:06:29.413 "uuid": "87708111-0b34-40ea-8e9a-33f0011a9ec2", 00:06:29.413 "strip_size_kb": 64, 00:06:29.413 "state": "configuring", 00:06:29.413 "raid_level": "raid0", 00:06:29.413 "superblock": true, 00:06:29.413 "num_base_bdevs": 2, 00:06:29.413 "num_base_bdevs_discovered": 0, 00:06:29.413 "num_base_bdevs_operational": 2, 00:06:29.413 "base_bdevs_list": [ 00:06:29.413 { 00:06:29.413 "name": "BaseBdev1", 00:06:29.413 "uuid": "00000000-0000-0000-0000-000000000000", 00:06:29.413 "is_configured": false, 00:06:29.413 "data_offset": 0, 00:06:29.413 "data_size": 0 00:06:29.413 }, 00:06:29.413 { 00:06:29.413 "name": "BaseBdev2", 00:06:29.413 "uuid": "00000000-0000-0000-0000-000000000000", 00:06:29.413 "is_configured": false, 00:06:29.413 "data_offset": 0, 00:06:29.413 "data_size": 0 00:06:29.413 } 00:06:29.413 ] 00:06:29.413 }' 00:06:29.413 23:40:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:06:29.413 23:40:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:06:29.983 23:40:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:06:29.983 23:40:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:29.983 23:40:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:06:29.983 [2024-11-26 23:40:17.885321] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:06:29.983 [2024-11-26 23:40:17.885390] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:06:29.983 23:40:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:29.983 23:40:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:06:29.983 23:40:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:29.983 23:40:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:06:29.983 [2024-11-26 23:40:17.893321] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:06:29.983 [2024-11-26 23:40:17.893391] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:06:29.983 [2024-11-26 23:40:17.893403] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:06:29.983 [2024-11-26 23:40:17.893428] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:06:29.983 23:40:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:29.983 23:40:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:06:29.983 23:40:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:29.983 23:40:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:06:29.983 [2024-11-26 23:40:17.910293] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:06:29.983 BaseBdev1 00:06:29.983 23:40:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:29.983 23:40:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:06:29.983 23:40:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:06:29.983 23:40:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:06:29.983 23:40:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:06:29.983 23:40:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:06:29.983 23:40:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:06:29.983 23:40:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:06:29.983 23:40:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:29.983 23:40:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:06:29.983 23:40:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:29.983 23:40:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:06:29.983 23:40:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:29.983 23:40:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:06:29.983 [ 00:06:29.983 { 00:06:29.983 "name": "BaseBdev1", 00:06:29.983 "aliases": [ 00:06:29.983 "9366f437-6fa4-48ff-b1ad-6fdf3e4e8efe" 00:06:29.983 ], 00:06:29.983 "product_name": "Malloc disk", 00:06:29.983 "block_size": 512, 00:06:29.983 "num_blocks": 65536, 00:06:29.983 "uuid": "9366f437-6fa4-48ff-b1ad-6fdf3e4e8efe", 00:06:29.983 "assigned_rate_limits": { 00:06:29.983 "rw_ios_per_sec": 0, 00:06:29.983 "rw_mbytes_per_sec": 0, 00:06:29.983 "r_mbytes_per_sec": 0, 00:06:29.983 "w_mbytes_per_sec": 0 00:06:29.983 }, 00:06:29.983 "claimed": true, 00:06:29.983 "claim_type": "exclusive_write", 00:06:29.983 "zoned": false, 00:06:29.983 "supported_io_types": { 00:06:29.983 "read": true, 00:06:29.983 "write": true, 00:06:29.983 "unmap": true, 00:06:29.983 "flush": true, 00:06:29.983 "reset": true, 00:06:29.983 "nvme_admin": false, 00:06:29.983 "nvme_io": false, 00:06:29.983 "nvme_io_md": false, 00:06:29.983 "write_zeroes": true, 00:06:29.983 "zcopy": true, 00:06:29.983 "get_zone_info": false, 00:06:29.983 "zone_management": false, 00:06:29.983 "zone_append": false, 00:06:29.983 "compare": false, 00:06:29.983 "compare_and_write": false, 00:06:29.983 "abort": true, 00:06:29.983 "seek_hole": false, 00:06:29.983 "seek_data": false, 00:06:29.983 "copy": true, 00:06:29.983 "nvme_iov_md": false 00:06:29.983 }, 00:06:29.983 "memory_domains": [ 00:06:29.983 { 00:06:29.983 "dma_device_id": "system", 00:06:29.983 "dma_device_type": 1 00:06:29.983 }, 00:06:29.983 { 00:06:29.983 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:06:29.983 "dma_device_type": 2 00:06:29.983 } 00:06:29.983 ], 00:06:29.983 "driver_specific": {} 00:06:29.983 } 00:06:29.983 ] 00:06:29.983 23:40:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:29.983 23:40:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:06:29.984 23:40:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 2 00:06:29.984 23:40:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:06:29.984 23:40:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:06:29.984 23:40:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:06:29.984 23:40:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:06:29.984 23:40:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:06:29.984 23:40:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:06:29.984 23:40:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:06:29.984 23:40:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:06:29.984 23:40:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:06:29.984 23:40:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:06:29.984 23:40:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:06:29.984 23:40:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:29.984 23:40:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:06:29.984 23:40:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:29.984 23:40:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:06:29.984 "name": "Existed_Raid", 00:06:29.984 "uuid": "7b48b4d3-2fc3-40b9-8103-eed7f373911a", 00:06:29.984 "strip_size_kb": 64, 00:06:29.984 "state": "configuring", 00:06:29.984 "raid_level": "raid0", 00:06:29.984 "superblock": true, 00:06:29.984 "num_base_bdevs": 2, 00:06:29.984 "num_base_bdevs_discovered": 1, 00:06:29.984 "num_base_bdevs_operational": 2, 00:06:29.984 "base_bdevs_list": [ 00:06:29.984 { 00:06:29.984 "name": "BaseBdev1", 00:06:29.984 "uuid": "9366f437-6fa4-48ff-b1ad-6fdf3e4e8efe", 00:06:29.984 "is_configured": true, 00:06:29.984 "data_offset": 2048, 00:06:29.984 "data_size": 63488 00:06:29.984 }, 00:06:29.984 { 00:06:29.984 "name": "BaseBdev2", 00:06:29.984 "uuid": "00000000-0000-0000-0000-000000000000", 00:06:29.984 "is_configured": false, 00:06:29.984 "data_offset": 0, 00:06:29.984 "data_size": 0 00:06:29.984 } 00:06:29.984 ] 00:06:29.984 }' 00:06:29.984 23:40:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:06:29.984 23:40:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:06:30.243 23:40:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:06:30.243 23:40:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:30.243 23:40:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:06:30.243 [2024-11-26 23:40:18.313649] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:06:30.243 [2024-11-26 23:40:18.313695] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:06:30.243 23:40:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:30.243 23:40:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:06:30.243 23:40:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:30.243 23:40:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:06:30.243 [2024-11-26 23:40:18.321668] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:06:30.243 [2024-11-26 23:40:18.323683] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:06:30.243 [2024-11-26 23:40:18.323761] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:06:30.243 23:40:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:30.243 23:40:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:06:30.243 23:40:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:06:30.243 23:40:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 2 00:06:30.243 23:40:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:06:30.243 23:40:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:06:30.243 23:40:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:06:30.243 23:40:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:06:30.243 23:40:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:06:30.243 23:40:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:06:30.243 23:40:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:06:30.243 23:40:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:06:30.243 23:40:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:06:30.243 23:40:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:06:30.243 23:40:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:30.243 23:40:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:06:30.243 23:40:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:06:30.243 23:40:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:30.243 23:40:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:06:30.243 "name": "Existed_Raid", 00:06:30.243 "uuid": "0331f2e3-d4b1-47a1-97cc-b376a5bc5daa", 00:06:30.243 "strip_size_kb": 64, 00:06:30.243 "state": "configuring", 00:06:30.243 "raid_level": "raid0", 00:06:30.243 "superblock": true, 00:06:30.243 "num_base_bdevs": 2, 00:06:30.243 "num_base_bdevs_discovered": 1, 00:06:30.243 "num_base_bdevs_operational": 2, 00:06:30.243 "base_bdevs_list": [ 00:06:30.243 { 00:06:30.243 "name": "BaseBdev1", 00:06:30.243 "uuid": "9366f437-6fa4-48ff-b1ad-6fdf3e4e8efe", 00:06:30.243 "is_configured": true, 00:06:30.243 "data_offset": 2048, 00:06:30.243 "data_size": 63488 00:06:30.243 }, 00:06:30.243 { 00:06:30.243 "name": "BaseBdev2", 00:06:30.243 "uuid": "00000000-0000-0000-0000-000000000000", 00:06:30.243 "is_configured": false, 00:06:30.243 "data_offset": 0, 00:06:30.243 "data_size": 0 00:06:30.243 } 00:06:30.243 ] 00:06:30.243 }' 00:06:30.243 23:40:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:06:30.243 23:40:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:06:30.812 23:40:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:06:30.812 23:40:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:30.812 23:40:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:06:30.812 [2024-11-26 23:40:18.715857] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:06:30.812 [2024-11-26 23:40:18.716113] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:06:30.812 [2024-11-26 23:40:18.716163] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:06:30.812 BaseBdev2 00:06:30.812 [2024-11-26 23:40:18.716492] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:06:30.812 23:40:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:30.812 [2024-11-26 23:40:18.716744] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:06:30.812 [2024-11-26 23:40:18.716778] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:06:30.812 [2024-11-26 23:40:18.716920] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:06:30.812 23:40:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:06:30.812 23:40:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:06:30.812 23:40:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:06:30.812 23:40:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:06:30.812 23:40:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:06:30.812 23:40:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:06:30.812 23:40:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:06:30.812 23:40:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:30.812 23:40:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:06:30.812 23:40:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:30.812 23:40:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:06:30.812 23:40:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:30.812 23:40:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:06:30.812 [ 00:06:30.812 { 00:06:30.812 "name": "BaseBdev2", 00:06:30.812 "aliases": [ 00:06:30.812 "ea29252d-3f7b-4058-913b-89bcfbb87d6b" 00:06:30.812 ], 00:06:30.812 "product_name": "Malloc disk", 00:06:30.812 "block_size": 512, 00:06:30.812 "num_blocks": 65536, 00:06:30.812 "uuid": "ea29252d-3f7b-4058-913b-89bcfbb87d6b", 00:06:30.812 "assigned_rate_limits": { 00:06:30.812 "rw_ios_per_sec": 0, 00:06:30.812 "rw_mbytes_per_sec": 0, 00:06:30.812 "r_mbytes_per_sec": 0, 00:06:30.812 "w_mbytes_per_sec": 0 00:06:30.812 }, 00:06:30.812 "claimed": true, 00:06:30.812 "claim_type": "exclusive_write", 00:06:30.812 "zoned": false, 00:06:30.812 "supported_io_types": { 00:06:30.812 "read": true, 00:06:30.812 "write": true, 00:06:30.812 "unmap": true, 00:06:30.812 "flush": true, 00:06:30.812 "reset": true, 00:06:30.812 "nvme_admin": false, 00:06:30.812 "nvme_io": false, 00:06:30.812 "nvme_io_md": false, 00:06:30.812 "write_zeroes": true, 00:06:30.812 "zcopy": true, 00:06:30.812 "get_zone_info": false, 00:06:30.812 "zone_management": false, 00:06:30.812 "zone_append": false, 00:06:30.812 "compare": false, 00:06:30.812 "compare_and_write": false, 00:06:30.812 "abort": true, 00:06:30.813 "seek_hole": false, 00:06:30.813 "seek_data": false, 00:06:30.813 "copy": true, 00:06:30.813 "nvme_iov_md": false 00:06:30.813 }, 00:06:30.813 "memory_domains": [ 00:06:30.813 { 00:06:30.813 "dma_device_id": "system", 00:06:30.813 "dma_device_type": 1 00:06:30.813 }, 00:06:30.813 { 00:06:30.813 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:06:30.813 "dma_device_type": 2 00:06:30.813 } 00:06:30.813 ], 00:06:30.813 "driver_specific": {} 00:06:30.813 } 00:06:30.813 ] 00:06:30.813 23:40:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:30.813 23:40:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:06:30.813 23:40:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:06:30.813 23:40:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:06:30.813 23:40:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid0 64 2 00:06:30.813 23:40:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:06:30.813 23:40:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:06:30.813 23:40:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:06:30.813 23:40:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:06:30.813 23:40:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:06:30.813 23:40:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:06:30.813 23:40:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:06:30.813 23:40:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:06:30.813 23:40:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:06:30.813 23:40:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:06:30.813 23:40:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:06:30.813 23:40:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:30.813 23:40:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:06:30.813 23:40:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:30.813 23:40:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:06:30.813 "name": "Existed_Raid", 00:06:30.813 "uuid": "0331f2e3-d4b1-47a1-97cc-b376a5bc5daa", 00:06:30.813 "strip_size_kb": 64, 00:06:30.813 "state": "online", 00:06:30.813 "raid_level": "raid0", 00:06:30.813 "superblock": true, 00:06:30.813 "num_base_bdevs": 2, 00:06:30.813 "num_base_bdevs_discovered": 2, 00:06:30.813 "num_base_bdevs_operational": 2, 00:06:30.813 "base_bdevs_list": [ 00:06:30.813 { 00:06:30.813 "name": "BaseBdev1", 00:06:30.813 "uuid": "9366f437-6fa4-48ff-b1ad-6fdf3e4e8efe", 00:06:30.813 "is_configured": true, 00:06:30.813 "data_offset": 2048, 00:06:30.813 "data_size": 63488 00:06:30.813 }, 00:06:30.813 { 00:06:30.813 "name": "BaseBdev2", 00:06:30.813 "uuid": "ea29252d-3f7b-4058-913b-89bcfbb87d6b", 00:06:30.813 "is_configured": true, 00:06:30.813 "data_offset": 2048, 00:06:30.813 "data_size": 63488 00:06:30.813 } 00:06:30.813 ] 00:06:30.813 }' 00:06:30.813 23:40:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:06:30.813 23:40:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:06:31.073 23:40:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:06:31.073 23:40:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:06:31.073 23:40:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:06:31.073 23:40:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:06:31.073 23:40:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:06:31.073 23:40:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:06:31.073 23:40:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:06:31.073 23:40:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:06:31.073 23:40:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:31.073 23:40:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:06:31.073 [2024-11-26 23:40:19.183396] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:06:31.073 23:40:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:31.333 23:40:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:06:31.333 "name": "Existed_Raid", 00:06:31.333 "aliases": [ 00:06:31.333 "0331f2e3-d4b1-47a1-97cc-b376a5bc5daa" 00:06:31.333 ], 00:06:31.333 "product_name": "Raid Volume", 00:06:31.333 "block_size": 512, 00:06:31.333 "num_blocks": 126976, 00:06:31.333 "uuid": "0331f2e3-d4b1-47a1-97cc-b376a5bc5daa", 00:06:31.333 "assigned_rate_limits": { 00:06:31.333 "rw_ios_per_sec": 0, 00:06:31.333 "rw_mbytes_per_sec": 0, 00:06:31.333 "r_mbytes_per_sec": 0, 00:06:31.333 "w_mbytes_per_sec": 0 00:06:31.333 }, 00:06:31.333 "claimed": false, 00:06:31.333 "zoned": false, 00:06:31.333 "supported_io_types": { 00:06:31.333 "read": true, 00:06:31.333 "write": true, 00:06:31.333 "unmap": true, 00:06:31.333 "flush": true, 00:06:31.333 "reset": true, 00:06:31.333 "nvme_admin": false, 00:06:31.333 "nvme_io": false, 00:06:31.333 "nvme_io_md": false, 00:06:31.333 "write_zeroes": true, 00:06:31.333 "zcopy": false, 00:06:31.333 "get_zone_info": false, 00:06:31.333 "zone_management": false, 00:06:31.333 "zone_append": false, 00:06:31.333 "compare": false, 00:06:31.333 "compare_and_write": false, 00:06:31.333 "abort": false, 00:06:31.333 "seek_hole": false, 00:06:31.333 "seek_data": false, 00:06:31.333 "copy": false, 00:06:31.333 "nvme_iov_md": false 00:06:31.333 }, 00:06:31.333 "memory_domains": [ 00:06:31.333 { 00:06:31.333 "dma_device_id": "system", 00:06:31.333 "dma_device_type": 1 00:06:31.333 }, 00:06:31.333 { 00:06:31.333 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:06:31.333 "dma_device_type": 2 00:06:31.333 }, 00:06:31.333 { 00:06:31.333 "dma_device_id": "system", 00:06:31.333 "dma_device_type": 1 00:06:31.333 }, 00:06:31.333 { 00:06:31.333 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:06:31.333 "dma_device_type": 2 00:06:31.333 } 00:06:31.333 ], 00:06:31.333 "driver_specific": { 00:06:31.333 "raid": { 00:06:31.333 "uuid": "0331f2e3-d4b1-47a1-97cc-b376a5bc5daa", 00:06:31.333 "strip_size_kb": 64, 00:06:31.333 "state": "online", 00:06:31.333 "raid_level": "raid0", 00:06:31.333 "superblock": true, 00:06:31.333 "num_base_bdevs": 2, 00:06:31.333 "num_base_bdevs_discovered": 2, 00:06:31.333 "num_base_bdevs_operational": 2, 00:06:31.333 "base_bdevs_list": [ 00:06:31.333 { 00:06:31.333 "name": "BaseBdev1", 00:06:31.333 "uuid": "9366f437-6fa4-48ff-b1ad-6fdf3e4e8efe", 00:06:31.333 "is_configured": true, 00:06:31.333 "data_offset": 2048, 00:06:31.333 "data_size": 63488 00:06:31.333 }, 00:06:31.333 { 00:06:31.333 "name": "BaseBdev2", 00:06:31.333 "uuid": "ea29252d-3f7b-4058-913b-89bcfbb87d6b", 00:06:31.333 "is_configured": true, 00:06:31.333 "data_offset": 2048, 00:06:31.333 "data_size": 63488 00:06:31.333 } 00:06:31.333 ] 00:06:31.333 } 00:06:31.333 } 00:06:31.333 }' 00:06:31.333 23:40:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:06:31.333 23:40:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:06:31.333 BaseBdev2' 00:06:31.333 23:40:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:06:31.333 23:40:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:06:31.333 23:40:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:06:31.333 23:40:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:06:31.333 23:40:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:06:31.333 23:40:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:31.333 23:40:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:06:31.333 23:40:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:31.333 23:40:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:06:31.333 23:40:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:06:31.333 23:40:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:06:31.333 23:40:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:06:31.333 23:40:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:06:31.333 23:40:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:31.333 23:40:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:06:31.333 23:40:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:31.333 23:40:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:06:31.333 23:40:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:06:31.333 23:40:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:06:31.333 23:40:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:31.333 23:40:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:06:31.333 [2024-11-26 23:40:19.414738] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:06:31.333 [2024-11-26 23:40:19.414766] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:06:31.333 [2024-11-26 23:40:19.414831] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:06:31.333 23:40:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:31.333 23:40:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:06:31.333 23:40:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy raid0 00:06:31.333 23:40:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:06:31.333 23:40:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # return 1 00:06:31.333 23:40:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:06:31.333 23:40:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline raid0 64 1 00:06:31.333 23:40:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:06:31.333 23:40:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:06:31.333 23:40:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:06:31.333 23:40:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:06:31.333 23:40:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:06:31.333 23:40:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:06:31.333 23:40:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:06:31.333 23:40:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:06:31.333 23:40:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:06:31.333 23:40:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:06:31.333 23:40:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:06:31.333 23:40:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:31.333 23:40:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:06:31.333 23:40:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:31.594 23:40:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:06:31.594 "name": "Existed_Raid", 00:06:31.594 "uuid": "0331f2e3-d4b1-47a1-97cc-b376a5bc5daa", 00:06:31.594 "strip_size_kb": 64, 00:06:31.594 "state": "offline", 00:06:31.594 "raid_level": "raid0", 00:06:31.594 "superblock": true, 00:06:31.594 "num_base_bdevs": 2, 00:06:31.594 "num_base_bdevs_discovered": 1, 00:06:31.594 "num_base_bdevs_operational": 1, 00:06:31.594 "base_bdevs_list": [ 00:06:31.594 { 00:06:31.594 "name": null, 00:06:31.594 "uuid": "00000000-0000-0000-0000-000000000000", 00:06:31.594 "is_configured": false, 00:06:31.594 "data_offset": 0, 00:06:31.594 "data_size": 63488 00:06:31.594 }, 00:06:31.594 { 00:06:31.594 "name": "BaseBdev2", 00:06:31.594 "uuid": "ea29252d-3f7b-4058-913b-89bcfbb87d6b", 00:06:31.594 "is_configured": true, 00:06:31.594 "data_offset": 2048, 00:06:31.594 "data_size": 63488 00:06:31.594 } 00:06:31.594 ] 00:06:31.594 }' 00:06:31.594 23:40:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:06:31.594 23:40:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:06:31.885 23:40:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:06:31.885 23:40:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:06:31.885 23:40:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:06:31.885 23:40:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:06:31.885 23:40:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:31.885 23:40:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:06:31.885 23:40:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:31.885 23:40:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:06:31.885 23:40:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:06:31.885 23:40:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:06:31.885 23:40:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:31.885 23:40:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:06:31.885 [2024-11-26 23:40:19.932979] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:06:31.885 [2024-11-26 23:40:19.933070] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:06:31.885 23:40:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:31.885 23:40:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:06:31.885 23:40:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:06:31.885 23:40:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:06:31.885 23:40:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:31.885 23:40:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:06:31.885 23:40:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:06:31.885 23:40:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:31.885 23:40:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:06:31.885 23:40:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:06:31.885 23:40:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 2 -gt 2 ']' 00:06:31.885 23:40:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 72075 00:06:31.885 23:40:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # '[' -z 72075 ']' 00:06:31.886 23:40:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@958 -- # kill -0 72075 00:06:31.886 23:40:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # uname 00:06:31.886 23:40:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:06:31.886 23:40:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 72075 00:06:32.145 killing process with pid 72075 00:06:32.145 23:40:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:06:32.145 23:40:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:06:32.145 23:40:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@972 -- # echo 'killing process with pid 72075' 00:06:32.145 23:40:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@973 -- # kill 72075 00:06:32.145 23:40:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@978 -- # wait 72075 00:06:32.145 [2024-11-26 23:40:20.035268] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:06:32.145 [2024-11-26 23:40:20.036259] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:06:32.145 ************************************ 00:06:32.145 END TEST raid_state_function_test_sb 00:06:32.145 ************************************ 00:06:32.145 23:40:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:06:32.145 00:06:32.145 real 0m3.733s 00:06:32.145 user 0m5.951s 00:06:32.145 sys 0m0.681s 00:06:32.145 23:40:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:32.145 23:40:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:06:32.405 23:40:20 bdev_raid -- bdev/bdev_raid.sh@970 -- # run_test raid_superblock_test raid_superblock_test raid0 2 00:06:32.405 23:40:20 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:06:32.405 23:40:20 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:32.405 23:40:20 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:06:32.405 ************************************ 00:06:32.405 START TEST raid_superblock_test 00:06:32.405 ************************************ 00:06:32.405 23:40:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1129 -- # raid_superblock_test raid0 2 00:06:32.405 23:40:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=raid0 00:06:32.405 23:40:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=2 00:06:32.405 23:40:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:06:32.405 23:40:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:06:32.405 23:40:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:06:32.405 23:40:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:06:32.405 23:40:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:06:32.405 23:40:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:06:32.405 23:40:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:06:32.405 23:40:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:06:32.405 23:40:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:06:32.405 23:40:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:06:32.405 23:40:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:06:32.405 23:40:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' raid0 '!=' raid1 ']' 00:06:32.405 23:40:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@405 -- # strip_size=64 00:06:32.405 23:40:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@406 -- # strip_size_create_arg='-z 64' 00:06:32.405 23:40:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=72316 00:06:32.405 23:40:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:06:32.405 23:40:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 72316 00:06:32.405 23:40:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@835 -- # '[' -z 72316 ']' 00:06:32.405 23:40:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:32.405 23:40:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:32.405 23:40:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:32.405 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:32.405 23:40:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:32.405 23:40:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:32.405 [2024-11-26 23:40:20.394761] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:06:32.405 [2024-11-26 23:40:20.394968] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid72316 ] 00:06:32.664 [2024-11-26 23:40:20.548550] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:32.664 [2024-11-26 23:40:20.573026] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:32.664 [2024-11-26 23:40:20.614890] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:06:32.664 [2024-11-26 23:40:20.615003] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:06:33.233 23:40:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:33.233 23:40:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@868 -- # return 0 00:06:33.233 23:40:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:06:33.233 23:40:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:06:33.233 23:40:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:06:33.233 23:40:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:06:33.233 23:40:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:06:33.233 23:40:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:06:33.233 23:40:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:06:33.233 23:40:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:06:33.233 23:40:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:06:33.233 23:40:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:33.233 23:40:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:33.233 malloc1 00:06:33.233 23:40:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:33.233 23:40:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:06:33.233 23:40:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:33.233 23:40:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:33.233 [2024-11-26 23:40:21.245848] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:06:33.233 [2024-11-26 23:40:21.245934] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:06:33.233 [2024-11-26 23:40:21.245964] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:06:33.233 [2024-11-26 23:40:21.245981] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:06:33.233 [2024-11-26 23:40:21.248142] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:06:33.233 [2024-11-26 23:40:21.248187] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:06:33.233 pt1 00:06:33.233 23:40:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:33.233 23:40:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:06:33.233 23:40:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:06:33.233 23:40:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:06:33.233 23:40:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:06:33.233 23:40:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:06:33.233 23:40:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:06:33.233 23:40:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:06:33.233 23:40:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:06:33.233 23:40:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:06:33.233 23:40:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:33.233 23:40:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:33.234 malloc2 00:06:33.234 23:40:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:33.234 23:40:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:06:33.234 23:40:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:33.234 23:40:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:33.234 [2024-11-26 23:40:21.266279] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:06:33.234 [2024-11-26 23:40:21.266355] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:06:33.234 [2024-11-26 23:40:21.266397] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:06:33.234 [2024-11-26 23:40:21.266429] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:06:33.234 [2024-11-26 23:40:21.268513] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:06:33.234 [2024-11-26 23:40:21.268549] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:06:33.234 pt2 00:06:33.234 23:40:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:33.234 23:40:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:06:33.234 23:40:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:06:33.234 23:40:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''pt1 pt2'\''' -n raid_bdev1 -s 00:06:33.234 23:40:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:33.234 23:40:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:33.234 [2024-11-26 23:40:21.274302] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:06:33.234 [2024-11-26 23:40:21.276128] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:06:33.234 [2024-11-26 23:40:21.276264] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:06:33.234 [2024-11-26 23:40:21.276277] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:06:33.234 [2024-11-26 23:40:21.276564] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:06:33.234 [2024-11-26 23:40:21.276760] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:06:33.234 [2024-11-26 23:40:21.276773] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:06:33.234 [2024-11-26 23:40:21.276905] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:06:33.234 23:40:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:33.234 23:40:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 2 00:06:33.234 23:40:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:06:33.234 23:40:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:06:33.234 23:40:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:06:33.234 23:40:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:06:33.234 23:40:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:06:33.234 23:40:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:06:33.234 23:40:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:06:33.234 23:40:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:06:33.234 23:40:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:06:33.234 23:40:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:06:33.234 23:40:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:33.234 23:40:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:33.234 23:40:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:06:33.234 23:40:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:33.234 23:40:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:06:33.234 "name": "raid_bdev1", 00:06:33.234 "uuid": "cbb83ce8-e2f1-4ae9-9801-c10d5a166758", 00:06:33.234 "strip_size_kb": 64, 00:06:33.234 "state": "online", 00:06:33.234 "raid_level": "raid0", 00:06:33.234 "superblock": true, 00:06:33.234 "num_base_bdevs": 2, 00:06:33.234 "num_base_bdevs_discovered": 2, 00:06:33.234 "num_base_bdevs_operational": 2, 00:06:33.234 "base_bdevs_list": [ 00:06:33.234 { 00:06:33.234 "name": "pt1", 00:06:33.234 "uuid": "00000000-0000-0000-0000-000000000001", 00:06:33.234 "is_configured": true, 00:06:33.234 "data_offset": 2048, 00:06:33.234 "data_size": 63488 00:06:33.234 }, 00:06:33.234 { 00:06:33.234 "name": "pt2", 00:06:33.234 "uuid": "00000000-0000-0000-0000-000000000002", 00:06:33.234 "is_configured": true, 00:06:33.234 "data_offset": 2048, 00:06:33.234 "data_size": 63488 00:06:33.234 } 00:06:33.234 ] 00:06:33.234 }' 00:06:33.234 23:40:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:06:33.234 23:40:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:33.804 23:40:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:06:33.804 23:40:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:06:33.804 23:40:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:06:33.804 23:40:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:06:33.804 23:40:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:06:33.804 23:40:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:06:33.804 23:40:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:06:33.804 23:40:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:33.804 23:40:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:33.804 23:40:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:06:33.804 [2024-11-26 23:40:21.701864] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:06:33.804 23:40:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:33.804 23:40:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:06:33.804 "name": "raid_bdev1", 00:06:33.804 "aliases": [ 00:06:33.804 "cbb83ce8-e2f1-4ae9-9801-c10d5a166758" 00:06:33.804 ], 00:06:33.804 "product_name": "Raid Volume", 00:06:33.804 "block_size": 512, 00:06:33.804 "num_blocks": 126976, 00:06:33.804 "uuid": "cbb83ce8-e2f1-4ae9-9801-c10d5a166758", 00:06:33.804 "assigned_rate_limits": { 00:06:33.804 "rw_ios_per_sec": 0, 00:06:33.804 "rw_mbytes_per_sec": 0, 00:06:33.804 "r_mbytes_per_sec": 0, 00:06:33.804 "w_mbytes_per_sec": 0 00:06:33.804 }, 00:06:33.804 "claimed": false, 00:06:33.804 "zoned": false, 00:06:33.804 "supported_io_types": { 00:06:33.804 "read": true, 00:06:33.804 "write": true, 00:06:33.804 "unmap": true, 00:06:33.804 "flush": true, 00:06:33.804 "reset": true, 00:06:33.804 "nvme_admin": false, 00:06:33.804 "nvme_io": false, 00:06:33.804 "nvme_io_md": false, 00:06:33.804 "write_zeroes": true, 00:06:33.804 "zcopy": false, 00:06:33.804 "get_zone_info": false, 00:06:33.804 "zone_management": false, 00:06:33.804 "zone_append": false, 00:06:33.804 "compare": false, 00:06:33.804 "compare_and_write": false, 00:06:33.804 "abort": false, 00:06:33.804 "seek_hole": false, 00:06:33.804 "seek_data": false, 00:06:33.804 "copy": false, 00:06:33.804 "nvme_iov_md": false 00:06:33.804 }, 00:06:33.804 "memory_domains": [ 00:06:33.804 { 00:06:33.804 "dma_device_id": "system", 00:06:33.804 "dma_device_type": 1 00:06:33.804 }, 00:06:33.804 { 00:06:33.804 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:06:33.804 "dma_device_type": 2 00:06:33.804 }, 00:06:33.804 { 00:06:33.804 "dma_device_id": "system", 00:06:33.804 "dma_device_type": 1 00:06:33.804 }, 00:06:33.804 { 00:06:33.804 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:06:33.804 "dma_device_type": 2 00:06:33.804 } 00:06:33.804 ], 00:06:33.804 "driver_specific": { 00:06:33.804 "raid": { 00:06:33.804 "uuid": "cbb83ce8-e2f1-4ae9-9801-c10d5a166758", 00:06:33.804 "strip_size_kb": 64, 00:06:33.804 "state": "online", 00:06:33.804 "raid_level": "raid0", 00:06:33.804 "superblock": true, 00:06:33.804 "num_base_bdevs": 2, 00:06:33.804 "num_base_bdevs_discovered": 2, 00:06:33.804 "num_base_bdevs_operational": 2, 00:06:33.804 "base_bdevs_list": [ 00:06:33.804 { 00:06:33.804 "name": "pt1", 00:06:33.804 "uuid": "00000000-0000-0000-0000-000000000001", 00:06:33.804 "is_configured": true, 00:06:33.804 "data_offset": 2048, 00:06:33.804 "data_size": 63488 00:06:33.804 }, 00:06:33.804 { 00:06:33.804 "name": "pt2", 00:06:33.804 "uuid": "00000000-0000-0000-0000-000000000002", 00:06:33.804 "is_configured": true, 00:06:33.804 "data_offset": 2048, 00:06:33.804 "data_size": 63488 00:06:33.804 } 00:06:33.804 ] 00:06:33.804 } 00:06:33.804 } 00:06:33.804 }' 00:06:33.804 23:40:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:06:33.804 23:40:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:06:33.804 pt2' 00:06:33.804 23:40:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:06:33.804 23:40:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:06:33.804 23:40:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:06:33.804 23:40:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:06:33.804 23:40:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:06:33.804 23:40:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:33.804 23:40:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:33.804 23:40:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:33.804 23:40:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:06:33.804 23:40:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:06:33.804 23:40:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:06:33.804 23:40:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:06:33.804 23:40:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:33.804 23:40:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:33.804 23:40:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:06:33.804 23:40:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:33.804 23:40:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:06:33.804 23:40:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:06:33.804 23:40:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:06:33.804 23:40:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:06:33.804 23:40:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:33.804 23:40:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:33.804 [2024-11-26 23:40:21.893491] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:06:33.804 23:40:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:33.804 23:40:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=cbb83ce8-e2f1-4ae9-9801-c10d5a166758 00:06:33.804 23:40:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z cbb83ce8-e2f1-4ae9-9801-c10d5a166758 ']' 00:06:33.804 23:40:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:06:33.804 23:40:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:33.804 23:40:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:33.804 [2024-11-26 23:40:21.925145] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:06:33.804 [2024-11-26 23:40:21.925212] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:06:33.804 [2024-11-26 23:40:21.925299] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:06:33.804 [2024-11-26 23:40:21.925365] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:06:33.804 [2024-11-26 23:40:21.925375] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:06:33.804 23:40:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:34.065 23:40:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:06:34.065 23:40:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:06:34.065 23:40:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:34.065 23:40:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:34.065 23:40:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:34.065 23:40:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:06:34.065 23:40:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:06:34.065 23:40:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:06:34.065 23:40:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:06:34.065 23:40:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:34.065 23:40:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:34.065 23:40:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:34.065 23:40:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:06:34.065 23:40:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:06:34.065 23:40:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:34.065 23:40:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:34.065 23:40:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:34.065 23:40:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:06:34.065 23:40:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:06:34.065 23:40:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:34.065 23:40:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:34.065 23:40:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:34.065 23:40:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:06:34.065 23:40:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:06:34.065 23:40:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@652 -- # local es=0 00:06:34.065 23:40:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:06:34.065 23:40:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:06:34.065 23:40:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:06:34.065 23:40:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:06:34.065 23:40:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:06:34.065 23:40:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:06:34.065 23:40:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:34.065 23:40:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:34.065 [2024-11-26 23:40:22.048945] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:06:34.065 [2024-11-26 23:40:22.050872] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:06:34.065 [2024-11-26 23:40:22.051006] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:06:34.065 [2024-11-26 23:40:22.051099] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:06:34.065 [2024-11-26 23:40:22.051201] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:06:34.065 [2024-11-26 23:40:22.051301] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state configuring 00:06:34.065 request: 00:06:34.065 { 00:06:34.065 "name": "raid_bdev1", 00:06:34.065 "raid_level": "raid0", 00:06:34.065 "base_bdevs": [ 00:06:34.065 "malloc1", 00:06:34.065 "malloc2" 00:06:34.065 ], 00:06:34.065 "strip_size_kb": 64, 00:06:34.065 "superblock": false, 00:06:34.065 "method": "bdev_raid_create", 00:06:34.065 "req_id": 1 00:06:34.065 } 00:06:34.065 Got JSON-RPC error response 00:06:34.065 response: 00:06:34.065 { 00:06:34.065 "code": -17, 00:06:34.065 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:06:34.065 } 00:06:34.065 23:40:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:06:34.065 23:40:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # es=1 00:06:34.065 23:40:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:06:34.065 23:40:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:06:34.065 23:40:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:06:34.065 23:40:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:06:34.065 23:40:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:34.065 23:40:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:34.065 23:40:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:06:34.065 23:40:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:34.065 23:40:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:06:34.065 23:40:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:06:34.065 23:40:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:06:34.065 23:40:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:34.065 23:40:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:34.065 [2024-11-26 23:40:22.104837] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:06:34.065 [2024-11-26 23:40:22.104894] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:06:34.065 [2024-11-26 23:40:22.104925] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:06:34.065 [2024-11-26 23:40:22.104936] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:06:34.065 [2024-11-26 23:40:22.107095] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:06:34.065 pt1 00:06:34.065 [2024-11-26 23:40:22.107185] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:06:34.065 [2024-11-26 23:40:22.107298] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:06:34.065 [2024-11-26 23:40:22.107358] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:06:34.065 23:40:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:34.065 23:40:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid0 64 2 00:06:34.065 23:40:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:06:34.065 23:40:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:06:34.065 23:40:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:06:34.065 23:40:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:06:34.065 23:40:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:06:34.065 23:40:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:06:34.065 23:40:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:06:34.066 23:40:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:06:34.066 23:40:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:06:34.066 23:40:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:06:34.066 23:40:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:06:34.066 23:40:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:34.066 23:40:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:34.066 23:40:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:34.066 23:40:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:06:34.066 "name": "raid_bdev1", 00:06:34.066 "uuid": "cbb83ce8-e2f1-4ae9-9801-c10d5a166758", 00:06:34.066 "strip_size_kb": 64, 00:06:34.066 "state": "configuring", 00:06:34.066 "raid_level": "raid0", 00:06:34.066 "superblock": true, 00:06:34.066 "num_base_bdevs": 2, 00:06:34.066 "num_base_bdevs_discovered": 1, 00:06:34.066 "num_base_bdevs_operational": 2, 00:06:34.066 "base_bdevs_list": [ 00:06:34.066 { 00:06:34.066 "name": "pt1", 00:06:34.066 "uuid": "00000000-0000-0000-0000-000000000001", 00:06:34.066 "is_configured": true, 00:06:34.066 "data_offset": 2048, 00:06:34.066 "data_size": 63488 00:06:34.066 }, 00:06:34.066 { 00:06:34.066 "name": null, 00:06:34.066 "uuid": "00000000-0000-0000-0000-000000000002", 00:06:34.066 "is_configured": false, 00:06:34.066 "data_offset": 2048, 00:06:34.066 "data_size": 63488 00:06:34.066 } 00:06:34.066 ] 00:06:34.066 }' 00:06:34.066 23:40:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:06:34.066 23:40:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:34.634 23:40:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 2 -gt 2 ']' 00:06:34.634 23:40:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:06:34.634 23:40:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:06:34.634 23:40:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:06:34.634 23:40:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:34.634 23:40:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:34.634 [2024-11-26 23:40:22.508191] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:06:34.634 [2024-11-26 23:40:22.508258] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:06:34.634 [2024-11-26 23:40:22.508289] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:06:34.634 [2024-11-26 23:40:22.508306] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:06:34.634 [2024-11-26 23:40:22.508794] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:06:34.634 [2024-11-26 23:40:22.508827] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:06:34.634 [2024-11-26 23:40:22.508931] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:06:34.634 [2024-11-26 23:40:22.508970] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:06:34.634 [2024-11-26 23:40:22.509096] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:06:34.634 [2024-11-26 23:40:22.509114] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:06:34.634 [2024-11-26 23:40:22.509406] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:06:34.634 [2024-11-26 23:40:22.509551] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:06:34.634 [2024-11-26 23:40:22.509575] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:06:34.634 [2024-11-26 23:40:22.509710] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:06:34.634 pt2 00:06:34.634 23:40:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:34.634 23:40:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:06:34.634 23:40:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:06:34.634 23:40:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 2 00:06:34.634 23:40:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:06:34.634 23:40:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:06:34.634 23:40:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:06:34.634 23:40:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:06:34.634 23:40:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:06:34.634 23:40:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:06:34.634 23:40:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:06:34.634 23:40:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:06:34.634 23:40:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:06:34.634 23:40:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:06:34.634 23:40:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:34.634 23:40:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:34.634 23:40:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:06:34.634 23:40:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:34.634 23:40:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:06:34.634 "name": "raid_bdev1", 00:06:34.634 "uuid": "cbb83ce8-e2f1-4ae9-9801-c10d5a166758", 00:06:34.634 "strip_size_kb": 64, 00:06:34.634 "state": "online", 00:06:34.635 "raid_level": "raid0", 00:06:34.635 "superblock": true, 00:06:34.635 "num_base_bdevs": 2, 00:06:34.635 "num_base_bdevs_discovered": 2, 00:06:34.635 "num_base_bdevs_operational": 2, 00:06:34.635 "base_bdevs_list": [ 00:06:34.635 { 00:06:34.635 "name": "pt1", 00:06:34.635 "uuid": "00000000-0000-0000-0000-000000000001", 00:06:34.635 "is_configured": true, 00:06:34.635 "data_offset": 2048, 00:06:34.635 "data_size": 63488 00:06:34.635 }, 00:06:34.635 { 00:06:34.635 "name": "pt2", 00:06:34.635 "uuid": "00000000-0000-0000-0000-000000000002", 00:06:34.635 "is_configured": true, 00:06:34.635 "data_offset": 2048, 00:06:34.635 "data_size": 63488 00:06:34.635 } 00:06:34.635 ] 00:06:34.635 }' 00:06:34.635 23:40:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:06:34.635 23:40:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:34.893 23:40:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:06:34.893 23:40:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:06:34.893 23:40:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:06:34.893 23:40:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:06:34.893 23:40:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:06:34.893 23:40:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:06:34.893 23:40:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:06:34.894 23:40:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:06:34.894 23:40:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:34.894 23:40:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:34.894 [2024-11-26 23:40:22.903784] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:06:34.894 23:40:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:34.894 23:40:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:06:34.894 "name": "raid_bdev1", 00:06:34.894 "aliases": [ 00:06:34.894 "cbb83ce8-e2f1-4ae9-9801-c10d5a166758" 00:06:34.894 ], 00:06:34.894 "product_name": "Raid Volume", 00:06:34.894 "block_size": 512, 00:06:34.894 "num_blocks": 126976, 00:06:34.894 "uuid": "cbb83ce8-e2f1-4ae9-9801-c10d5a166758", 00:06:34.894 "assigned_rate_limits": { 00:06:34.894 "rw_ios_per_sec": 0, 00:06:34.894 "rw_mbytes_per_sec": 0, 00:06:34.894 "r_mbytes_per_sec": 0, 00:06:34.894 "w_mbytes_per_sec": 0 00:06:34.894 }, 00:06:34.894 "claimed": false, 00:06:34.894 "zoned": false, 00:06:34.894 "supported_io_types": { 00:06:34.894 "read": true, 00:06:34.894 "write": true, 00:06:34.894 "unmap": true, 00:06:34.894 "flush": true, 00:06:34.894 "reset": true, 00:06:34.894 "nvme_admin": false, 00:06:34.894 "nvme_io": false, 00:06:34.894 "nvme_io_md": false, 00:06:34.894 "write_zeroes": true, 00:06:34.894 "zcopy": false, 00:06:34.894 "get_zone_info": false, 00:06:34.894 "zone_management": false, 00:06:34.894 "zone_append": false, 00:06:34.894 "compare": false, 00:06:34.894 "compare_and_write": false, 00:06:34.894 "abort": false, 00:06:34.894 "seek_hole": false, 00:06:34.894 "seek_data": false, 00:06:34.894 "copy": false, 00:06:34.894 "nvme_iov_md": false 00:06:34.894 }, 00:06:34.894 "memory_domains": [ 00:06:34.894 { 00:06:34.894 "dma_device_id": "system", 00:06:34.894 "dma_device_type": 1 00:06:34.894 }, 00:06:34.894 { 00:06:34.894 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:06:34.894 "dma_device_type": 2 00:06:34.894 }, 00:06:34.894 { 00:06:34.894 "dma_device_id": "system", 00:06:34.894 "dma_device_type": 1 00:06:34.894 }, 00:06:34.894 { 00:06:34.894 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:06:34.894 "dma_device_type": 2 00:06:34.894 } 00:06:34.894 ], 00:06:34.894 "driver_specific": { 00:06:34.894 "raid": { 00:06:34.894 "uuid": "cbb83ce8-e2f1-4ae9-9801-c10d5a166758", 00:06:34.894 "strip_size_kb": 64, 00:06:34.894 "state": "online", 00:06:34.894 "raid_level": "raid0", 00:06:34.894 "superblock": true, 00:06:34.894 "num_base_bdevs": 2, 00:06:34.894 "num_base_bdevs_discovered": 2, 00:06:34.894 "num_base_bdevs_operational": 2, 00:06:34.894 "base_bdevs_list": [ 00:06:34.894 { 00:06:34.894 "name": "pt1", 00:06:34.894 "uuid": "00000000-0000-0000-0000-000000000001", 00:06:34.894 "is_configured": true, 00:06:34.894 "data_offset": 2048, 00:06:34.894 "data_size": 63488 00:06:34.894 }, 00:06:34.894 { 00:06:34.894 "name": "pt2", 00:06:34.894 "uuid": "00000000-0000-0000-0000-000000000002", 00:06:34.894 "is_configured": true, 00:06:34.894 "data_offset": 2048, 00:06:34.894 "data_size": 63488 00:06:34.894 } 00:06:34.894 ] 00:06:34.894 } 00:06:34.894 } 00:06:34.894 }' 00:06:34.894 23:40:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:06:34.894 23:40:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:06:34.894 pt2' 00:06:34.894 23:40:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:06:34.894 23:40:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:06:34.894 23:40:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:06:34.894 23:40:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:06:34.894 23:40:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:34.894 23:40:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:34.894 23:40:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:06:35.153 23:40:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:35.153 23:40:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:06:35.153 23:40:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:06:35.153 23:40:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:06:35.153 23:40:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:06:35.153 23:40:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:06:35.153 23:40:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:35.153 23:40:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:35.153 23:40:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:35.153 23:40:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:06:35.153 23:40:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:06:35.153 23:40:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:06:35.153 23:40:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:35.153 23:40:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:35.153 23:40:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:06:35.153 [2024-11-26 23:40:23.119403] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:06:35.153 23:40:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:35.153 23:40:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' cbb83ce8-e2f1-4ae9-9801-c10d5a166758 '!=' cbb83ce8-e2f1-4ae9-9801-c10d5a166758 ']' 00:06:35.153 23:40:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy raid0 00:06:35.154 23:40:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:06:35.154 23:40:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # return 1 00:06:35.154 23:40:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 72316 00:06:35.154 23:40:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # '[' -z 72316 ']' 00:06:35.154 23:40:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@958 -- # kill -0 72316 00:06:35.154 23:40:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # uname 00:06:35.154 23:40:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:06:35.154 23:40:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 72316 00:06:35.154 killing process with pid 72316 00:06:35.154 23:40:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:06:35.154 23:40:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:06:35.154 23:40:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 72316' 00:06:35.154 23:40:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@973 -- # kill 72316 00:06:35.154 23:40:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@978 -- # wait 72316 00:06:35.154 [2024-11-26 23:40:23.182215] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:06:35.154 [2024-11-26 23:40:23.182314] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:06:35.154 [2024-11-26 23:40:23.182396] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:06:35.154 [2024-11-26 23:40:23.182430] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:06:35.154 [2024-11-26 23:40:23.204612] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:06:35.413 ************************************ 00:06:35.413 END TEST raid_superblock_test 00:06:35.413 ************************************ 00:06:35.413 23:40:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:06:35.413 00:06:35.413 real 0m3.099s 00:06:35.413 user 0m4.775s 00:06:35.413 sys 0m0.643s 00:06:35.413 23:40:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:35.413 23:40:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:35.413 23:40:23 bdev_raid -- bdev/bdev_raid.sh@971 -- # run_test raid_read_error_test raid_io_error_test raid0 2 read 00:06:35.413 23:40:23 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:06:35.413 23:40:23 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:35.413 23:40:23 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:06:35.413 ************************************ 00:06:35.413 START TEST raid_read_error_test 00:06:35.413 ************************************ 00:06:35.413 23:40:23 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test raid0 2 read 00:06:35.413 23:40:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid0 00:06:35.413 23:40:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=2 00:06:35.413 23:40:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=read 00:06:35.413 23:40:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:06:35.413 23:40:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:06:35.413 23:40:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:06:35.413 23:40:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:06:35.413 23:40:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:06:35.413 23:40:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:06:35.413 23:40:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:06:35.413 23:40:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:06:35.413 23:40:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:06:35.413 23:40:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:06:35.413 23:40:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:06:35.413 23:40:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:06:35.413 23:40:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:06:35.413 23:40:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:06:35.413 23:40:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:06:35.413 23:40:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid0 '!=' raid1 ']' 00:06:35.413 23:40:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:06:35.413 23:40:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:06:35.413 23:40:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:06:35.413 23:40:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.8STrIf83U2 00:06:35.413 23:40:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=72511 00:06:35.413 23:40:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:06:35.413 23:40:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 72511 00:06:35.413 23:40:23 bdev_raid.raid_read_error_test -- common/autotest_common.sh@835 -- # '[' -z 72511 ']' 00:06:35.414 23:40:23 bdev_raid.raid_read_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:35.414 23:40:23 bdev_raid.raid_read_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:35.414 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:35.414 23:40:23 bdev_raid.raid_read_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:35.414 23:40:23 bdev_raid.raid_read_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:35.414 23:40:23 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:06:35.673 [2024-11-26 23:40:23.577476] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:06:35.673 [2024-11-26 23:40:23.577622] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid72511 ] 00:06:35.673 [2024-11-26 23:40:23.725609] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:35.673 [2024-11-26 23:40:23.750524] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:35.673 [2024-11-26 23:40:23.792152] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:06:35.673 [2024-11-26 23:40:23.792189] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:06:36.610 23:40:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:36.610 23:40:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@868 -- # return 0 00:06:36.610 23:40:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:06:36.610 23:40:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:06:36.610 23:40:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:36.610 23:40:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:06:36.610 BaseBdev1_malloc 00:06:36.610 23:40:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:36.610 23:40:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:06:36.610 23:40:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:36.611 23:40:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:06:36.611 true 00:06:36.611 23:40:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:36.611 23:40:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:06:36.611 23:40:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:36.611 23:40:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:06:36.611 [2024-11-26 23:40:24.414991] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:06:36.611 [2024-11-26 23:40:24.415054] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:06:36.611 [2024-11-26 23:40:24.415082] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006980 00:06:36.611 [2024-11-26 23:40:24.415094] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:06:36.611 [2024-11-26 23:40:24.417316] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:06:36.611 [2024-11-26 23:40:24.417363] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:06:36.611 BaseBdev1 00:06:36.611 23:40:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:36.611 23:40:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:06:36.611 23:40:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:06:36.611 23:40:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:36.611 23:40:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:06:36.611 BaseBdev2_malloc 00:06:36.611 23:40:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:36.611 23:40:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:06:36.611 23:40:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:36.611 23:40:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:06:36.611 true 00:06:36.611 23:40:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:36.611 23:40:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:06:36.611 23:40:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:36.611 23:40:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:06:36.611 [2024-11-26 23:40:24.443374] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:06:36.611 [2024-11-26 23:40:24.443425] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:06:36.611 [2024-11-26 23:40:24.443447] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:06:36.611 [2024-11-26 23:40:24.443468] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:06:36.611 [2024-11-26 23:40:24.445535] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:06:36.611 [2024-11-26 23:40:24.445571] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:06:36.611 BaseBdev2 00:06:36.611 23:40:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:36.611 23:40:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 -s 00:06:36.611 23:40:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:36.611 23:40:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:06:36.611 [2024-11-26 23:40:24.451414] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:06:36.611 [2024-11-26 23:40:24.453190] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:06:36.611 [2024-11-26 23:40:24.453400] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:06:36.611 [2024-11-26 23:40:24.453414] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:06:36.611 [2024-11-26 23:40:24.453668] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002870 00:06:36.611 [2024-11-26 23:40:24.453818] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:06:36.611 [2024-11-26 23:40:24.453840] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:06:36.611 [2024-11-26 23:40:24.453968] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:06:36.611 23:40:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:36.611 23:40:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 2 00:06:36.611 23:40:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:06:36.611 23:40:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:06:36.611 23:40:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:06:36.611 23:40:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:06:36.611 23:40:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:06:36.611 23:40:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:06:36.611 23:40:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:06:36.611 23:40:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:06:36.611 23:40:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:06:36.611 23:40:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:06:36.611 23:40:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:06:36.611 23:40:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:36.611 23:40:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:06:36.611 23:40:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:36.611 23:40:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:06:36.611 "name": "raid_bdev1", 00:06:36.611 "uuid": "d6b9ee9c-514f-4f64-922e-5e7ae46835d2", 00:06:36.611 "strip_size_kb": 64, 00:06:36.611 "state": "online", 00:06:36.611 "raid_level": "raid0", 00:06:36.611 "superblock": true, 00:06:36.611 "num_base_bdevs": 2, 00:06:36.611 "num_base_bdevs_discovered": 2, 00:06:36.611 "num_base_bdevs_operational": 2, 00:06:36.611 "base_bdevs_list": [ 00:06:36.611 { 00:06:36.611 "name": "BaseBdev1", 00:06:36.611 "uuid": "6717935e-da0e-5ccb-9788-d9c1e212d93a", 00:06:36.611 "is_configured": true, 00:06:36.611 "data_offset": 2048, 00:06:36.611 "data_size": 63488 00:06:36.611 }, 00:06:36.611 { 00:06:36.611 "name": "BaseBdev2", 00:06:36.611 "uuid": "b2056930-6d80-5faf-833e-157e7e8c73f0", 00:06:36.611 "is_configured": true, 00:06:36.611 "data_offset": 2048, 00:06:36.611 "data_size": 63488 00:06:36.611 } 00:06:36.611 ] 00:06:36.611 }' 00:06:36.611 23:40:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:06:36.611 23:40:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:06:36.871 23:40:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:06:36.871 23:40:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:06:36.871 [2024-11-26 23:40:24.978848] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002a10 00:06:37.813 23:40:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:06:37.813 23:40:25 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:37.813 23:40:25 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:06:37.813 23:40:25 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:37.813 23:40:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:06:37.813 23:40:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid0 = \r\a\i\d\1 ]] 00:06:37.813 23:40:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=2 00:06:37.813 23:40:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 2 00:06:37.814 23:40:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:06:37.814 23:40:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:06:37.814 23:40:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:06:37.814 23:40:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:06:37.814 23:40:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:06:37.814 23:40:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:06:37.814 23:40:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:06:37.814 23:40:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:06:37.814 23:40:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:06:37.814 23:40:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:06:37.814 23:40:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:06:37.814 23:40:25 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:37.814 23:40:25 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:06:37.814 23:40:25 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:38.072 23:40:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:06:38.072 "name": "raid_bdev1", 00:06:38.072 "uuid": "d6b9ee9c-514f-4f64-922e-5e7ae46835d2", 00:06:38.072 "strip_size_kb": 64, 00:06:38.072 "state": "online", 00:06:38.072 "raid_level": "raid0", 00:06:38.072 "superblock": true, 00:06:38.072 "num_base_bdevs": 2, 00:06:38.072 "num_base_bdevs_discovered": 2, 00:06:38.072 "num_base_bdevs_operational": 2, 00:06:38.072 "base_bdevs_list": [ 00:06:38.072 { 00:06:38.072 "name": "BaseBdev1", 00:06:38.072 "uuid": "6717935e-da0e-5ccb-9788-d9c1e212d93a", 00:06:38.072 "is_configured": true, 00:06:38.072 "data_offset": 2048, 00:06:38.072 "data_size": 63488 00:06:38.072 }, 00:06:38.072 { 00:06:38.072 "name": "BaseBdev2", 00:06:38.072 "uuid": "b2056930-6d80-5faf-833e-157e7e8c73f0", 00:06:38.072 "is_configured": true, 00:06:38.072 "data_offset": 2048, 00:06:38.072 "data_size": 63488 00:06:38.072 } 00:06:38.072 ] 00:06:38.072 }' 00:06:38.072 23:40:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:06:38.072 23:40:25 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:06:38.332 23:40:26 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:06:38.332 23:40:26 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:38.332 23:40:26 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:06:38.332 [2024-11-26 23:40:26.342390] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:06:38.332 [2024-11-26 23:40:26.342419] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:06:38.332 [2024-11-26 23:40:26.344998] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:06:38.332 [2024-11-26 23:40:26.345043] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:06:38.332 [2024-11-26 23:40:26.345077] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:06:38.332 [2024-11-26 23:40:26.345086] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:06:38.332 { 00:06:38.332 "results": [ 00:06:38.332 { 00:06:38.332 "job": "raid_bdev1", 00:06:38.332 "core_mask": "0x1", 00:06:38.332 "workload": "randrw", 00:06:38.332 "percentage": 50, 00:06:38.332 "status": "finished", 00:06:38.332 "queue_depth": 1, 00:06:38.332 "io_size": 131072, 00:06:38.332 "runtime": 1.3645, 00:06:38.332 "iops": 17081.71491388787, 00:06:38.332 "mibps": 2135.214364235984, 00:06:38.332 "io_failed": 1, 00:06:38.332 "io_timeout": 0, 00:06:38.332 "avg_latency_us": 80.4675176726721, 00:06:38.332 "min_latency_us": 25.041048034934498, 00:06:38.332 "max_latency_us": 1409.4532751091704 00:06:38.332 } 00:06:38.332 ], 00:06:38.332 "core_count": 1 00:06:38.332 } 00:06:38.332 23:40:26 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:38.332 23:40:26 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 72511 00:06:38.332 23:40:26 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # '[' -z 72511 ']' 00:06:38.332 23:40:26 bdev_raid.raid_read_error_test -- common/autotest_common.sh@958 -- # kill -0 72511 00:06:38.332 23:40:26 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # uname 00:06:38.332 23:40:26 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:06:38.332 23:40:26 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 72511 00:06:38.332 23:40:26 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:06:38.332 23:40:26 bdev_raid.raid_read_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:06:38.332 killing process with pid 72511 00:06:38.332 23:40:26 bdev_raid.raid_read_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 72511' 00:06:38.332 23:40:26 bdev_raid.raid_read_error_test -- common/autotest_common.sh@973 -- # kill 72511 00:06:38.332 23:40:26 bdev_raid.raid_read_error_test -- common/autotest_common.sh@978 -- # wait 72511 00:06:38.332 [2024-11-26 23:40:26.370951] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:06:38.332 [2024-11-26 23:40:26.385847] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:06:38.593 23:40:26 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:06:38.593 23:40:26 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.8STrIf83U2 00:06:38.593 23:40:26 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:06:38.593 23:40:26 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.73 00:06:38.593 23:40:26 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid0 00:06:38.593 23:40:26 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:06:38.593 23:40:26 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:06:38.593 23:40:26 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.73 != \0\.\0\0 ]] 00:06:38.593 00:06:38.593 real 0m3.114s 00:06:38.593 user 0m3.991s 00:06:38.593 sys 0m0.435s 00:06:38.593 23:40:26 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:38.593 23:40:26 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:06:38.593 ************************************ 00:06:38.593 END TEST raid_read_error_test 00:06:38.593 ************************************ 00:06:38.593 23:40:26 bdev_raid -- bdev/bdev_raid.sh@972 -- # run_test raid_write_error_test raid_io_error_test raid0 2 write 00:06:38.593 23:40:26 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:06:38.593 23:40:26 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:38.593 23:40:26 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:06:38.593 ************************************ 00:06:38.593 START TEST raid_write_error_test 00:06:38.593 ************************************ 00:06:38.593 23:40:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test raid0 2 write 00:06:38.593 23:40:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid0 00:06:38.593 23:40:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=2 00:06:38.593 23:40:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=write 00:06:38.593 23:40:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:06:38.593 23:40:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:06:38.593 23:40:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:06:38.593 23:40:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:06:38.593 23:40:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:06:38.593 23:40:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:06:38.593 23:40:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:06:38.593 23:40:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:06:38.593 23:40:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:06:38.593 23:40:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:06:38.593 23:40:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:06:38.593 23:40:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:06:38.593 23:40:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:06:38.593 23:40:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:06:38.593 23:40:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:06:38.593 23:40:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid0 '!=' raid1 ']' 00:06:38.593 23:40:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:06:38.593 23:40:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:06:38.593 23:40:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:06:38.593 23:40:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.76EpCFj9of 00:06:38.593 23:40:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=72640 00:06:38.593 23:40:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:06:38.593 23:40:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 72640 00:06:38.593 23:40:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@835 -- # '[' -z 72640 ']' 00:06:38.593 23:40:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:38.593 23:40:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:38.593 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:38.593 23:40:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:38.593 23:40:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:38.593 23:40:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:06:38.852 [2024-11-26 23:40:26.752496] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:06:38.852 [2024-11-26 23:40:26.752647] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid72640 ] 00:06:38.852 [2024-11-26 23:40:26.906725] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:38.852 [2024-11-26 23:40:26.931148] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:38.852 [2024-11-26 23:40:26.972358] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:06:38.852 [2024-11-26 23:40:26.972400] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:06:39.790 23:40:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:39.790 23:40:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@868 -- # return 0 00:06:39.790 23:40:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:06:39.790 23:40:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:06:39.790 23:40:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:39.790 23:40:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:06:39.790 BaseBdev1_malloc 00:06:39.790 23:40:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:39.790 23:40:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:06:39.790 23:40:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:39.790 23:40:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:06:39.790 true 00:06:39.790 23:40:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:39.790 23:40:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:06:39.790 23:40:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:39.790 23:40:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:06:39.790 [2024-11-26 23:40:27.599016] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:06:39.790 [2024-11-26 23:40:27.599071] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:06:39.790 [2024-11-26 23:40:27.599110] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006980 00:06:39.790 [2024-11-26 23:40:27.599122] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:06:39.790 [2024-11-26 23:40:27.601204] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:06:39.790 [2024-11-26 23:40:27.601240] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:06:39.790 BaseBdev1 00:06:39.790 23:40:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:39.790 23:40:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:06:39.790 23:40:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:06:39.790 23:40:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:39.790 23:40:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:06:39.790 BaseBdev2_malloc 00:06:39.790 23:40:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:39.790 23:40:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:06:39.790 23:40:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:39.790 23:40:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:06:39.790 true 00:06:39.790 23:40:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:39.790 23:40:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:06:39.790 23:40:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:39.790 23:40:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:06:39.790 [2024-11-26 23:40:27.627294] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:06:39.790 [2024-11-26 23:40:27.627360] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:06:39.790 [2024-11-26 23:40:27.627385] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:06:39.790 [2024-11-26 23:40:27.627408] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:06:39.790 [2024-11-26 23:40:27.629423] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:06:39.790 [2024-11-26 23:40:27.629459] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:06:39.790 BaseBdev2 00:06:39.790 23:40:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:39.790 23:40:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 -s 00:06:39.790 23:40:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:39.790 23:40:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:06:39.790 [2024-11-26 23:40:27.635325] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:06:39.790 [2024-11-26 23:40:27.637112] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:06:39.790 [2024-11-26 23:40:27.637278] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:06:39.790 [2024-11-26 23:40:27.637290] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:06:39.790 [2024-11-26 23:40:27.637546] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002870 00:06:39.790 [2024-11-26 23:40:27.637707] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:06:39.790 [2024-11-26 23:40:27.637723] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:06:39.790 [2024-11-26 23:40:27.637845] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:06:39.790 23:40:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:39.790 23:40:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 2 00:06:39.790 23:40:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:06:39.790 23:40:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:06:39.790 23:40:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:06:39.790 23:40:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:06:39.790 23:40:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:06:39.790 23:40:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:06:39.790 23:40:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:06:39.790 23:40:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:06:39.791 23:40:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:06:39.791 23:40:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:06:39.791 23:40:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:06:39.791 23:40:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:39.791 23:40:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:06:39.791 23:40:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:39.791 23:40:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:06:39.791 "name": "raid_bdev1", 00:06:39.791 "uuid": "8221f163-d487-45b2-82d7-a6703a0b8b68", 00:06:39.791 "strip_size_kb": 64, 00:06:39.791 "state": "online", 00:06:39.791 "raid_level": "raid0", 00:06:39.791 "superblock": true, 00:06:39.791 "num_base_bdevs": 2, 00:06:39.791 "num_base_bdevs_discovered": 2, 00:06:39.791 "num_base_bdevs_operational": 2, 00:06:39.791 "base_bdevs_list": [ 00:06:39.791 { 00:06:39.791 "name": "BaseBdev1", 00:06:39.791 "uuid": "8cc53c0a-3baf-5b13-b0a2-67a5d8cd8ee3", 00:06:39.791 "is_configured": true, 00:06:39.791 "data_offset": 2048, 00:06:39.791 "data_size": 63488 00:06:39.791 }, 00:06:39.791 { 00:06:39.791 "name": "BaseBdev2", 00:06:39.791 "uuid": "eaedc8d0-4da4-56b3-a830-1049cc6958f1", 00:06:39.791 "is_configured": true, 00:06:39.791 "data_offset": 2048, 00:06:39.791 "data_size": 63488 00:06:39.791 } 00:06:39.791 ] 00:06:39.791 }' 00:06:39.791 23:40:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:06:39.791 23:40:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:06:40.049 23:40:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:06:40.049 23:40:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:06:40.050 [2024-11-26 23:40:28.146815] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002a10 00:06:40.994 23:40:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:06:40.994 23:40:29 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:40.994 23:40:29 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:06:40.994 23:40:29 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:40.994 23:40:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:06:40.994 23:40:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid0 = \r\a\i\d\1 ]] 00:06:40.994 23:40:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=2 00:06:40.994 23:40:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 2 00:06:40.994 23:40:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:06:40.994 23:40:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:06:40.994 23:40:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:06:40.994 23:40:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:06:40.994 23:40:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:06:40.994 23:40:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:06:40.994 23:40:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:06:40.994 23:40:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:06:40.994 23:40:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:06:40.994 23:40:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:06:40.994 23:40:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:06:40.994 23:40:29 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:40.994 23:40:29 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:06:40.994 23:40:29 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:40.994 23:40:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:06:40.994 "name": "raid_bdev1", 00:06:40.994 "uuid": "8221f163-d487-45b2-82d7-a6703a0b8b68", 00:06:40.994 "strip_size_kb": 64, 00:06:40.994 "state": "online", 00:06:40.994 "raid_level": "raid0", 00:06:40.994 "superblock": true, 00:06:40.994 "num_base_bdevs": 2, 00:06:40.994 "num_base_bdevs_discovered": 2, 00:06:40.994 "num_base_bdevs_operational": 2, 00:06:40.994 "base_bdevs_list": [ 00:06:40.994 { 00:06:40.994 "name": "BaseBdev1", 00:06:40.994 "uuid": "8cc53c0a-3baf-5b13-b0a2-67a5d8cd8ee3", 00:06:40.994 "is_configured": true, 00:06:40.994 "data_offset": 2048, 00:06:40.994 "data_size": 63488 00:06:40.994 }, 00:06:40.994 { 00:06:40.994 "name": "BaseBdev2", 00:06:40.994 "uuid": "eaedc8d0-4da4-56b3-a830-1049cc6958f1", 00:06:40.994 "is_configured": true, 00:06:40.994 "data_offset": 2048, 00:06:40.994 "data_size": 63488 00:06:40.994 } 00:06:40.994 ] 00:06:40.994 }' 00:06:40.994 23:40:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:06:40.994 23:40:29 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:06:41.563 23:40:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:06:41.563 23:40:29 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:41.563 23:40:29 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:06:41.563 [2024-11-26 23:40:29.494322] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:06:41.563 [2024-11-26 23:40:29.494375] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:06:41.563 [2024-11-26 23:40:29.496930] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:06:41.563 [2024-11-26 23:40:29.496970] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:06:41.563 [2024-11-26 23:40:29.497003] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:06:41.563 [2024-11-26 23:40:29.497011] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:06:41.563 { 00:06:41.563 "results": [ 00:06:41.563 { 00:06:41.563 "job": "raid_bdev1", 00:06:41.563 "core_mask": "0x1", 00:06:41.563 "workload": "randrw", 00:06:41.563 "percentage": 50, 00:06:41.563 "status": "finished", 00:06:41.563 "queue_depth": 1, 00:06:41.563 "io_size": 131072, 00:06:41.563 "runtime": 1.348415, 00:06:41.563 "iops": 17298.82862471865, 00:06:41.563 "mibps": 2162.3535780898314, 00:06:41.563 "io_failed": 1, 00:06:41.563 "io_timeout": 0, 00:06:41.563 "avg_latency_us": 79.59162085728947, 00:06:41.563 "min_latency_us": 25.041048034934498, 00:06:41.563 "max_latency_us": 1387.989519650655 00:06:41.563 } 00:06:41.563 ], 00:06:41.563 "core_count": 1 00:06:41.563 } 00:06:41.563 23:40:29 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:41.563 23:40:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 72640 00:06:41.563 23:40:29 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # '[' -z 72640 ']' 00:06:41.563 23:40:29 bdev_raid.raid_write_error_test -- common/autotest_common.sh@958 -- # kill -0 72640 00:06:41.563 23:40:29 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # uname 00:06:41.563 23:40:29 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:06:41.563 23:40:29 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 72640 00:06:41.563 killing process with pid 72640 00:06:41.563 23:40:29 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:06:41.563 23:40:29 bdev_raid.raid_write_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:06:41.563 23:40:29 bdev_raid.raid_write_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 72640' 00:06:41.563 23:40:29 bdev_raid.raid_write_error_test -- common/autotest_common.sh@973 -- # kill 72640 00:06:41.563 23:40:29 bdev_raid.raid_write_error_test -- common/autotest_common.sh@978 -- # wait 72640 00:06:41.563 [2024-11-26 23:40:29.541319] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:06:41.563 [2024-11-26 23:40:29.556758] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:06:41.823 23:40:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.76EpCFj9of 00:06:41.823 23:40:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:06:41.823 23:40:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:06:41.823 23:40:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.74 00:06:41.823 23:40:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid0 00:06:41.823 23:40:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:06:41.823 23:40:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:06:41.823 23:40:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.74 != \0\.\0\0 ]] 00:06:41.823 00:06:41.823 real 0m3.103s 00:06:41.823 user 0m3.957s 00:06:41.823 sys 0m0.468s 00:06:41.823 23:40:29 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:41.823 23:40:29 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:06:41.823 ************************************ 00:06:41.823 END TEST raid_write_error_test 00:06:41.823 ************************************ 00:06:41.823 23:40:29 bdev_raid -- bdev/bdev_raid.sh@967 -- # for level in raid0 concat raid1 00:06:41.823 23:40:29 bdev_raid -- bdev/bdev_raid.sh@968 -- # run_test raid_state_function_test raid_state_function_test concat 2 false 00:06:41.823 23:40:29 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:06:41.823 23:40:29 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:41.823 23:40:29 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:06:41.823 ************************************ 00:06:41.823 START TEST raid_state_function_test 00:06:41.823 ************************************ 00:06:41.823 23:40:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1129 -- # raid_state_function_test concat 2 false 00:06:41.823 23:40:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=concat 00:06:41.823 23:40:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=2 00:06:41.823 23:40:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:06:41.823 23:40:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:06:41.823 23:40:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:06:41.823 23:40:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:06:41.823 23:40:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:06:41.823 23:40:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:06:41.823 23:40:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:06:41.823 23:40:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:06:41.823 23:40:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:06:41.823 23:40:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:06:41.823 23:40:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:06:41.823 23:40:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:06:41.823 23:40:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:06:41.823 23:40:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:06:41.823 23:40:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:06:41.823 23:40:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:06:41.823 23:40:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' concat '!=' raid1 ']' 00:06:41.823 23:40:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:06:41.823 23:40:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:06:41.823 23:40:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:06:41.823 23:40:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:06:41.823 23:40:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=72767 00:06:41.823 23:40:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:06:41.823 23:40:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 72767' 00:06:41.823 Process raid pid: 72767 00:06:41.823 23:40:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 72767 00:06:41.823 23:40:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@835 -- # '[' -z 72767 ']' 00:06:41.823 23:40:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:41.823 23:40:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:41.823 23:40:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:41.823 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:41.823 23:40:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:41.823 23:40:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:06:41.823 [2024-11-26 23:40:29.915567] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:06:41.823 [2024-11-26 23:40:29.915789] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:06:42.083 [2024-11-26 23:40:30.071579] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:42.083 [2024-11-26 23:40:30.096040] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:42.083 [2024-11-26 23:40:30.137380] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:06:42.083 [2024-11-26 23:40:30.137489] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:06:42.651 23:40:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:42.651 23:40:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@868 -- # return 0 00:06:42.651 23:40:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:06:42.651 23:40:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:42.651 23:40:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:06:42.651 [2024-11-26 23:40:30.739249] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:06:42.651 [2024-11-26 23:40:30.739376] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:06:42.651 [2024-11-26 23:40:30.739425] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:06:42.651 [2024-11-26 23:40:30.739458] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:06:42.651 23:40:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:42.651 23:40:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 2 00:06:42.651 23:40:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:06:42.651 23:40:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:06:42.651 23:40:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:06:42.651 23:40:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:06:42.651 23:40:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:06:42.651 23:40:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:06:42.651 23:40:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:06:42.651 23:40:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:06:42.651 23:40:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:06:42.652 23:40:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:06:42.652 23:40:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:06:42.652 23:40:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:42.652 23:40:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:06:42.652 23:40:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:42.652 23:40:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:06:42.652 "name": "Existed_Raid", 00:06:42.652 "uuid": "00000000-0000-0000-0000-000000000000", 00:06:42.652 "strip_size_kb": 64, 00:06:42.652 "state": "configuring", 00:06:42.652 "raid_level": "concat", 00:06:42.652 "superblock": false, 00:06:42.652 "num_base_bdevs": 2, 00:06:42.652 "num_base_bdevs_discovered": 0, 00:06:42.652 "num_base_bdevs_operational": 2, 00:06:42.652 "base_bdevs_list": [ 00:06:42.652 { 00:06:42.652 "name": "BaseBdev1", 00:06:42.652 "uuid": "00000000-0000-0000-0000-000000000000", 00:06:42.652 "is_configured": false, 00:06:42.652 "data_offset": 0, 00:06:42.652 "data_size": 0 00:06:42.652 }, 00:06:42.652 { 00:06:42.652 "name": "BaseBdev2", 00:06:42.652 "uuid": "00000000-0000-0000-0000-000000000000", 00:06:42.652 "is_configured": false, 00:06:42.652 "data_offset": 0, 00:06:42.652 "data_size": 0 00:06:42.652 } 00:06:42.652 ] 00:06:42.652 }' 00:06:42.652 23:40:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:06:42.652 23:40:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:06:43.220 23:40:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:06:43.220 23:40:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:43.220 23:40:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:06:43.220 [2024-11-26 23:40:31.106529] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:06:43.220 [2024-11-26 23:40:31.106570] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:06:43.220 23:40:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:43.220 23:40:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:06:43.220 23:40:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:43.220 23:40:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:06:43.220 [2024-11-26 23:40:31.114522] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:06:43.220 [2024-11-26 23:40:31.114602] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:06:43.220 [2024-11-26 23:40:31.114619] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:06:43.220 [2024-11-26 23:40:31.114647] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:06:43.220 23:40:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:43.220 23:40:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:06:43.220 23:40:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:43.220 23:40:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:06:43.220 BaseBdev1 00:06:43.220 [2024-11-26 23:40:31.131162] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:06:43.220 23:40:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:43.220 23:40:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:06:43.220 23:40:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:06:43.220 23:40:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:06:43.220 23:40:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:06:43.220 23:40:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:06:43.220 23:40:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:06:43.220 23:40:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:06:43.220 23:40:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:43.220 23:40:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:06:43.220 23:40:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:43.220 23:40:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:06:43.220 23:40:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:43.220 23:40:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:06:43.220 [ 00:06:43.220 { 00:06:43.220 "name": "BaseBdev1", 00:06:43.220 "aliases": [ 00:06:43.220 "cb7a847b-f0a8-46c9-8340-f71ea656a0e0" 00:06:43.220 ], 00:06:43.220 "product_name": "Malloc disk", 00:06:43.220 "block_size": 512, 00:06:43.220 "num_blocks": 65536, 00:06:43.220 "uuid": "cb7a847b-f0a8-46c9-8340-f71ea656a0e0", 00:06:43.220 "assigned_rate_limits": { 00:06:43.220 "rw_ios_per_sec": 0, 00:06:43.220 "rw_mbytes_per_sec": 0, 00:06:43.220 "r_mbytes_per_sec": 0, 00:06:43.220 "w_mbytes_per_sec": 0 00:06:43.220 }, 00:06:43.220 "claimed": true, 00:06:43.220 "claim_type": "exclusive_write", 00:06:43.220 "zoned": false, 00:06:43.220 "supported_io_types": { 00:06:43.220 "read": true, 00:06:43.220 "write": true, 00:06:43.220 "unmap": true, 00:06:43.220 "flush": true, 00:06:43.220 "reset": true, 00:06:43.220 "nvme_admin": false, 00:06:43.220 "nvme_io": false, 00:06:43.220 "nvme_io_md": false, 00:06:43.220 "write_zeroes": true, 00:06:43.220 "zcopy": true, 00:06:43.220 "get_zone_info": false, 00:06:43.220 "zone_management": false, 00:06:43.220 "zone_append": false, 00:06:43.220 "compare": false, 00:06:43.220 "compare_and_write": false, 00:06:43.220 "abort": true, 00:06:43.220 "seek_hole": false, 00:06:43.220 "seek_data": false, 00:06:43.220 "copy": true, 00:06:43.220 "nvme_iov_md": false 00:06:43.220 }, 00:06:43.220 "memory_domains": [ 00:06:43.220 { 00:06:43.220 "dma_device_id": "system", 00:06:43.220 "dma_device_type": 1 00:06:43.220 }, 00:06:43.220 { 00:06:43.220 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:06:43.220 "dma_device_type": 2 00:06:43.220 } 00:06:43.220 ], 00:06:43.220 "driver_specific": {} 00:06:43.220 } 00:06:43.220 ] 00:06:43.220 23:40:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:43.220 23:40:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:06:43.220 23:40:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 2 00:06:43.220 23:40:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:06:43.220 23:40:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:06:43.220 23:40:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:06:43.220 23:40:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:06:43.220 23:40:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:06:43.220 23:40:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:06:43.220 23:40:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:06:43.220 23:40:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:06:43.221 23:40:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:06:43.221 23:40:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:06:43.221 23:40:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:06:43.221 23:40:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:43.221 23:40:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:06:43.221 23:40:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:43.221 23:40:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:06:43.221 "name": "Existed_Raid", 00:06:43.221 "uuid": "00000000-0000-0000-0000-000000000000", 00:06:43.221 "strip_size_kb": 64, 00:06:43.221 "state": "configuring", 00:06:43.221 "raid_level": "concat", 00:06:43.221 "superblock": false, 00:06:43.221 "num_base_bdevs": 2, 00:06:43.221 "num_base_bdevs_discovered": 1, 00:06:43.221 "num_base_bdevs_operational": 2, 00:06:43.221 "base_bdevs_list": [ 00:06:43.221 { 00:06:43.221 "name": "BaseBdev1", 00:06:43.221 "uuid": "cb7a847b-f0a8-46c9-8340-f71ea656a0e0", 00:06:43.221 "is_configured": true, 00:06:43.221 "data_offset": 0, 00:06:43.221 "data_size": 65536 00:06:43.221 }, 00:06:43.221 { 00:06:43.221 "name": "BaseBdev2", 00:06:43.221 "uuid": "00000000-0000-0000-0000-000000000000", 00:06:43.221 "is_configured": false, 00:06:43.221 "data_offset": 0, 00:06:43.221 "data_size": 0 00:06:43.221 } 00:06:43.221 ] 00:06:43.221 }' 00:06:43.221 23:40:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:06:43.221 23:40:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:06:43.489 23:40:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:06:43.489 23:40:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:43.489 23:40:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:06:43.489 [2024-11-26 23:40:31.566467] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:06:43.489 [2024-11-26 23:40:31.566563] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:06:43.489 23:40:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:43.489 23:40:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:06:43.489 23:40:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:43.489 23:40:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:06:43.489 [2024-11-26 23:40:31.578459] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:06:43.489 [2024-11-26 23:40:31.580301] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:06:43.489 [2024-11-26 23:40:31.580387] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:06:43.489 23:40:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:43.489 23:40:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:06:43.490 23:40:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:06:43.490 23:40:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 2 00:06:43.490 23:40:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:06:43.490 23:40:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:06:43.490 23:40:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:06:43.490 23:40:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:06:43.490 23:40:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:06:43.490 23:40:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:06:43.490 23:40:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:06:43.490 23:40:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:06:43.490 23:40:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:06:43.490 23:40:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:06:43.490 23:40:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:06:43.490 23:40:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:43.490 23:40:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:06:43.490 23:40:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:43.758 23:40:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:06:43.758 "name": "Existed_Raid", 00:06:43.758 "uuid": "00000000-0000-0000-0000-000000000000", 00:06:43.758 "strip_size_kb": 64, 00:06:43.758 "state": "configuring", 00:06:43.758 "raid_level": "concat", 00:06:43.758 "superblock": false, 00:06:43.758 "num_base_bdevs": 2, 00:06:43.758 "num_base_bdevs_discovered": 1, 00:06:43.758 "num_base_bdevs_operational": 2, 00:06:43.758 "base_bdevs_list": [ 00:06:43.758 { 00:06:43.758 "name": "BaseBdev1", 00:06:43.758 "uuid": "cb7a847b-f0a8-46c9-8340-f71ea656a0e0", 00:06:43.758 "is_configured": true, 00:06:43.758 "data_offset": 0, 00:06:43.758 "data_size": 65536 00:06:43.758 }, 00:06:43.758 { 00:06:43.758 "name": "BaseBdev2", 00:06:43.758 "uuid": "00000000-0000-0000-0000-000000000000", 00:06:43.758 "is_configured": false, 00:06:43.758 "data_offset": 0, 00:06:43.758 "data_size": 0 00:06:43.758 } 00:06:43.758 ] 00:06:43.758 }' 00:06:43.758 23:40:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:06:43.758 23:40:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:06:44.030 23:40:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:06:44.030 23:40:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:44.030 23:40:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:06:44.030 [2024-11-26 23:40:32.004474] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:06:44.030 [2024-11-26 23:40:32.004517] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:06:44.030 [2024-11-26 23:40:32.004525] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 131072, blocklen 512 00:06:44.030 [2024-11-26 23:40:32.004801] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:06:44.030 [2024-11-26 23:40:32.004943] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:06:44.030 [2024-11-26 23:40:32.004957] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:06:44.030 [2024-11-26 23:40:32.005166] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:06:44.030 BaseBdev2 00:06:44.030 23:40:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:44.030 23:40:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:06:44.030 23:40:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:06:44.030 23:40:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:06:44.030 23:40:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:06:44.030 23:40:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:06:44.030 23:40:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:06:44.030 23:40:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:06:44.030 23:40:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:44.030 23:40:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:06:44.030 23:40:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:44.030 23:40:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:06:44.030 23:40:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:44.030 23:40:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:06:44.030 [ 00:06:44.030 { 00:06:44.030 "name": "BaseBdev2", 00:06:44.030 "aliases": [ 00:06:44.030 "7cafcaf8-9e48-40f0-bed8-72338c24f555" 00:06:44.030 ], 00:06:44.030 "product_name": "Malloc disk", 00:06:44.030 "block_size": 512, 00:06:44.030 "num_blocks": 65536, 00:06:44.030 "uuid": "7cafcaf8-9e48-40f0-bed8-72338c24f555", 00:06:44.030 "assigned_rate_limits": { 00:06:44.030 "rw_ios_per_sec": 0, 00:06:44.030 "rw_mbytes_per_sec": 0, 00:06:44.030 "r_mbytes_per_sec": 0, 00:06:44.030 "w_mbytes_per_sec": 0 00:06:44.030 }, 00:06:44.030 "claimed": true, 00:06:44.030 "claim_type": "exclusive_write", 00:06:44.030 "zoned": false, 00:06:44.030 "supported_io_types": { 00:06:44.030 "read": true, 00:06:44.030 "write": true, 00:06:44.030 "unmap": true, 00:06:44.030 "flush": true, 00:06:44.030 "reset": true, 00:06:44.030 "nvme_admin": false, 00:06:44.030 "nvme_io": false, 00:06:44.030 "nvme_io_md": false, 00:06:44.030 "write_zeroes": true, 00:06:44.030 "zcopy": true, 00:06:44.030 "get_zone_info": false, 00:06:44.030 "zone_management": false, 00:06:44.030 "zone_append": false, 00:06:44.030 "compare": false, 00:06:44.030 "compare_and_write": false, 00:06:44.030 "abort": true, 00:06:44.030 "seek_hole": false, 00:06:44.030 "seek_data": false, 00:06:44.030 "copy": true, 00:06:44.030 "nvme_iov_md": false 00:06:44.030 }, 00:06:44.030 "memory_domains": [ 00:06:44.030 { 00:06:44.030 "dma_device_id": "system", 00:06:44.030 "dma_device_type": 1 00:06:44.030 }, 00:06:44.030 { 00:06:44.030 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:06:44.030 "dma_device_type": 2 00:06:44.030 } 00:06:44.030 ], 00:06:44.030 "driver_specific": {} 00:06:44.030 } 00:06:44.030 ] 00:06:44.030 23:40:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:44.030 23:40:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:06:44.030 23:40:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:06:44.030 23:40:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:06:44.030 23:40:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online concat 64 2 00:06:44.030 23:40:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:06:44.030 23:40:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:06:44.030 23:40:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:06:44.030 23:40:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:06:44.030 23:40:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:06:44.030 23:40:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:06:44.030 23:40:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:06:44.030 23:40:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:06:44.030 23:40:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:06:44.030 23:40:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:06:44.030 23:40:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:06:44.030 23:40:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:44.030 23:40:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:06:44.030 23:40:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:44.030 23:40:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:06:44.030 "name": "Existed_Raid", 00:06:44.030 "uuid": "fb8c783c-25d0-464f-93ad-523e6a91f0e1", 00:06:44.030 "strip_size_kb": 64, 00:06:44.030 "state": "online", 00:06:44.030 "raid_level": "concat", 00:06:44.030 "superblock": false, 00:06:44.030 "num_base_bdevs": 2, 00:06:44.030 "num_base_bdevs_discovered": 2, 00:06:44.030 "num_base_bdevs_operational": 2, 00:06:44.030 "base_bdevs_list": [ 00:06:44.030 { 00:06:44.030 "name": "BaseBdev1", 00:06:44.030 "uuid": "cb7a847b-f0a8-46c9-8340-f71ea656a0e0", 00:06:44.030 "is_configured": true, 00:06:44.030 "data_offset": 0, 00:06:44.030 "data_size": 65536 00:06:44.030 }, 00:06:44.030 { 00:06:44.030 "name": "BaseBdev2", 00:06:44.030 "uuid": "7cafcaf8-9e48-40f0-bed8-72338c24f555", 00:06:44.030 "is_configured": true, 00:06:44.030 "data_offset": 0, 00:06:44.030 "data_size": 65536 00:06:44.030 } 00:06:44.030 ] 00:06:44.030 }' 00:06:44.030 23:40:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:06:44.030 23:40:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:06:44.600 23:40:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:06:44.600 23:40:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:06:44.600 23:40:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:06:44.600 23:40:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:06:44.600 23:40:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:06:44.600 23:40:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:06:44.600 23:40:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:06:44.600 23:40:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:06:44.600 23:40:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:44.600 23:40:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:06:44.600 [2024-11-26 23:40:32.487968] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:06:44.600 23:40:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:44.600 23:40:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:06:44.600 "name": "Existed_Raid", 00:06:44.600 "aliases": [ 00:06:44.600 "fb8c783c-25d0-464f-93ad-523e6a91f0e1" 00:06:44.600 ], 00:06:44.600 "product_name": "Raid Volume", 00:06:44.600 "block_size": 512, 00:06:44.600 "num_blocks": 131072, 00:06:44.600 "uuid": "fb8c783c-25d0-464f-93ad-523e6a91f0e1", 00:06:44.600 "assigned_rate_limits": { 00:06:44.600 "rw_ios_per_sec": 0, 00:06:44.600 "rw_mbytes_per_sec": 0, 00:06:44.600 "r_mbytes_per_sec": 0, 00:06:44.600 "w_mbytes_per_sec": 0 00:06:44.600 }, 00:06:44.600 "claimed": false, 00:06:44.600 "zoned": false, 00:06:44.600 "supported_io_types": { 00:06:44.600 "read": true, 00:06:44.600 "write": true, 00:06:44.600 "unmap": true, 00:06:44.600 "flush": true, 00:06:44.600 "reset": true, 00:06:44.600 "nvme_admin": false, 00:06:44.600 "nvme_io": false, 00:06:44.600 "nvme_io_md": false, 00:06:44.600 "write_zeroes": true, 00:06:44.600 "zcopy": false, 00:06:44.600 "get_zone_info": false, 00:06:44.600 "zone_management": false, 00:06:44.600 "zone_append": false, 00:06:44.600 "compare": false, 00:06:44.600 "compare_and_write": false, 00:06:44.600 "abort": false, 00:06:44.600 "seek_hole": false, 00:06:44.600 "seek_data": false, 00:06:44.600 "copy": false, 00:06:44.600 "nvme_iov_md": false 00:06:44.600 }, 00:06:44.600 "memory_domains": [ 00:06:44.600 { 00:06:44.600 "dma_device_id": "system", 00:06:44.600 "dma_device_type": 1 00:06:44.600 }, 00:06:44.600 { 00:06:44.600 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:06:44.600 "dma_device_type": 2 00:06:44.600 }, 00:06:44.600 { 00:06:44.600 "dma_device_id": "system", 00:06:44.600 "dma_device_type": 1 00:06:44.600 }, 00:06:44.600 { 00:06:44.600 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:06:44.600 "dma_device_type": 2 00:06:44.600 } 00:06:44.600 ], 00:06:44.600 "driver_specific": { 00:06:44.600 "raid": { 00:06:44.600 "uuid": "fb8c783c-25d0-464f-93ad-523e6a91f0e1", 00:06:44.600 "strip_size_kb": 64, 00:06:44.600 "state": "online", 00:06:44.600 "raid_level": "concat", 00:06:44.600 "superblock": false, 00:06:44.600 "num_base_bdevs": 2, 00:06:44.600 "num_base_bdevs_discovered": 2, 00:06:44.600 "num_base_bdevs_operational": 2, 00:06:44.600 "base_bdevs_list": [ 00:06:44.600 { 00:06:44.600 "name": "BaseBdev1", 00:06:44.600 "uuid": "cb7a847b-f0a8-46c9-8340-f71ea656a0e0", 00:06:44.600 "is_configured": true, 00:06:44.600 "data_offset": 0, 00:06:44.600 "data_size": 65536 00:06:44.600 }, 00:06:44.600 { 00:06:44.600 "name": "BaseBdev2", 00:06:44.600 "uuid": "7cafcaf8-9e48-40f0-bed8-72338c24f555", 00:06:44.600 "is_configured": true, 00:06:44.600 "data_offset": 0, 00:06:44.600 "data_size": 65536 00:06:44.600 } 00:06:44.600 ] 00:06:44.600 } 00:06:44.600 } 00:06:44.600 }' 00:06:44.600 23:40:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:06:44.600 23:40:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:06:44.600 BaseBdev2' 00:06:44.600 23:40:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:06:44.600 23:40:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:06:44.600 23:40:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:06:44.600 23:40:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:06:44.600 23:40:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:06:44.600 23:40:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:44.600 23:40:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:06:44.600 23:40:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:44.600 23:40:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:06:44.600 23:40:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:06:44.600 23:40:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:06:44.600 23:40:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:06:44.600 23:40:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:06:44.601 23:40:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:44.601 23:40:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:06:44.601 23:40:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:44.601 23:40:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:06:44.601 23:40:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:06:44.601 23:40:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:06:44.601 23:40:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:44.601 23:40:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:06:44.601 [2024-11-26 23:40:32.723362] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:06:44.601 [2024-11-26 23:40:32.723431] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:06:44.601 [2024-11-26 23:40:32.723500] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:06:44.860 23:40:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:44.860 23:40:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:06:44.860 23:40:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy concat 00:06:44.860 23:40:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:06:44.860 23:40:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # return 1 00:06:44.860 23:40:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:06:44.860 23:40:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline concat 64 1 00:06:44.860 23:40:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:06:44.860 23:40:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:06:44.860 23:40:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:06:44.860 23:40:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:06:44.860 23:40:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:06:44.860 23:40:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:06:44.860 23:40:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:06:44.860 23:40:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:06:44.860 23:40:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:06:44.860 23:40:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:06:44.860 23:40:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:44.860 23:40:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:06:44.860 23:40:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:06:44.860 23:40:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:44.860 23:40:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:06:44.860 "name": "Existed_Raid", 00:06:44.860 "uuid": "fb8c783c-25d0-464f-93ad-523e6a91f0e1", 00:06:44.860 "strip_size_kb": 64, 00:06:44.860 "state": "offline", 00:06:44.860 "raid_level": "concat", 00:06:44.860 "superblock": false, 00:06:44.860 "num_base_bdevs": 2, 00:06:44.860 "num_base_bdevs_discovered": 1, 00:06:44.860 "num_base_bdevs_operational": 1, 00:06:44.860 "base_bdevs_list": [ 00:06:44.860 { 00:06:44.860 "name": null, 00:06:44.860 "uuid": "00000000-0000-0000-0000-000000000000", 00:06:44.860 "is_configured": false, 00:06:44.860 "data_offset": 0, 00:06:44.860 "data_size": 65536 00:06:44.860 }, 00:06:44.860 { 00:06:44.860 "name": "BaseBdev2", 00:06:44.860 "uuid": "7cafcaf8-9e48-40f0-bed8-72338c24f555", 00:06:44.860 "is_configured": true, 00:06:44.860 "data_offset": 0, 00:06:44.860 "data_size": 65536 00:06:44.860 } 00:06:44.860 ] 00:06:44.860 }' 00:06:44.860 23:40:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:06:44.860 23:40:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:06:45.119 23:40:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:06:45.119 23:40:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:06:45.119 23:40:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:06:45.119 23:40:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:06:45.119 23:40:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:45.119 23:40:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:06:45.119 23:40:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:45.119 23:40:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:06:45.119 23:40:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:06:45.119 23:40:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:06:45.119 23:40:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:45.119 23:40:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:06:45.120 [2024-11-26 23:40:33.213776] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:06:45.120 [2024-11-26 23:40:33.213878] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:06:45.120 23:40:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:45.120 23:40:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:06:45.120 23:40:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:06:45.120 23:40:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:06:45.120 23:40:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:06:45.120 23:40:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:45.120 23:40:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:06:45.120 23:40:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:45.378 23:40:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:06:45.378 23:40:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:06:45.378 23:40:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 2 -gt 2 ']' 00:06:45.378 23:40:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 72767 00:06:45.378 23:40:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # '[' -z 72767 ']' 00:06:45.378 23:40:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@958 -- # kill -0 72767 00:06:45.378 23:40:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # uname 00:06:45.378 23:40:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:06:45.378 23:40:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 72767 00:06:45.378 killing process with pid 72767 00:06:45.378 23:40:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:06:45.378 23:40:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:06:45.378 23:40:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 72767' 00:06:45.378 23:40:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@973 -- # kill 72767 00:06:45.378 [2024-11-26 23:40:33.305936] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:06:45.378 23:40:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@978 -- # wait 72767 00:06:45.378 [2024-11-26 23:40:33.306879] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:06:45.637 23:40:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:06:45.637 00:06:45.637 real 0m3.687s 00:06:45.637 user 0m5.867s 00:06:45.637 sys 0m0.687s 00:06:45.637 23:40:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:45.637 23:40:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:06:45.637 ************************************ 00:06:45.637 END TEST raid_state_function_test 00:06:45.637 ************************************ 00:06:45.637 23:40:33 bdev_raid -- bdev/bdev_raid.sh@969 -- # run_test raid_state_function_test_sb raid_state_function_test concat 2 true 00:06:45.637 23:40:33 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:06:45.637 23:40:33 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:45.638 23:40:33 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:06:45.638 ************************************ 00:06:45.638 START TEST raid_state_function_test_sb 00:06:45.638 ************************************ 00:06:45.638 23:40:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1129 -- # raid_state_function_test concat 2 true 00:06:45.638 23:40:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=concat 00:06:45.638 23:40:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=2 00:06:45.638 23:40:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:06:45.638 23:40:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:06:45.638 23:40:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:06:45.638 23:40:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:06:45.638 23:40:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:06:45.638 23:40:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:06:45.638 23:40:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:06:45.638 23:40:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:06:45.638 23:40:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:06:45.638 23:40:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:06:45.638 23:40:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:06:45.638 23:40:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:06:45.638 23:40:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:06:45.638 23:40:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:06:45.638 23:40:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:06:45.638 23:40:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:06:45.638 23:40:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' concat '!=' raid1 ']' 00:06:45.638 23:40:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:06:45.638 23:40:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:06:45.638 23:40:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:06:45.638 23:40:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:06:45.638 23:40:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=73009 00:06:45.638 23:40:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:06:45.638 23:40:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 73009' 00:06:45.638 Process raid pid: 73009 00:06:45.638 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:45.638 23:40:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 73009 00:06:45.638 23:40:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@835 -- # '[' -z 73009 ']' 00:06:45.638 23:40:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:45.638 23:40:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:45.638 23:40:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:45.638 23:40:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:45.638 23:40:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:06:45.638 [2024-11-26 23:40:33.673283] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:06:45.638 [2024-11-26 23:40:33.673564] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:06:45.897 [2024-11-26 23:40:33.830892] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:45.897 [2024-11-26 23:40:33.855001] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:45.897 [2024-11-26 23:40:33.896464] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:06:45.897 [2024-11-26 23:40:33.896578] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:06:46.465 23:40:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:46.465 23:40:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@868 -- # return 0 00:06:46.465 23:40:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:06:46.465 23:40:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:46.465 23:40:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:06:46.465 [2024-11-26 23:40:34.494455] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:06:46.465 [2024-11-26 23:40:34.494551] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:06:46.465 [2024-11-26 23:40:34.494579] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:06:46.465 [2024-11-26 23:40:34.494603] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:06:46.465 23:40:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:46.465 23:40:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 2 00:06:46.465 23:40:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:06:46.465 23:40:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:06:46.465 23:40:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:06:46.465 23:40:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:06:46.465 23:40:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:06:46.465 23:40:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:06:46.465 23:40:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:06:46.465 23:40:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:06:46.465 23:40:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:06:46.465 23:40:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:06:46.465 23:40:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:06:46.465 23:40:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:46.465 23:40:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:06:46.465 23:40:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:46.465 23:40:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:06:46.465 "name": "Existed_Raid", 00:06:46.465 "uuid": "5ad17b95-ed8d-464e-99e2-5b0b1ae12bff", 00:06:46.465 "strip_size_kb": 64, 00:06:46.465 "state": "configuring", 00:06:46.465 "raid_level": "concat", 00:06:46.465 "superblock": true, 00:06:46.465 "num_base_bdevs": 2, 00:06:46.465 "num_base_bdevs_discovered": 0, 00:06:46.465 "num_base_bdevs_operational": 2, 00:06:46.465 "base_bdevs_list": [ 00:06:46.465 { 00:06:46.465 "name": "BaseBdev1", 00:06:46.465 "uuid": "00000000-0000-0000-0000-000000000000", 00:06:46.465 "is_configured": false, 00:06:46.465 "data_offset": 0, 00:06:46.465 "data_size": 0 00:06:46.465 }, 00:06:46.465 { 00:06:46.465 "name": "BaseBdev2", 00:06:46.465 "uuid": "00000000-0000-0000-0000-000000000000", 00:06:46.465 "is_configured": false, 00:06:46.465 "data_offset": 0, 00:06:46.465 "data_size": 0 00:06:46.465 } 00:06:46.465 ] 00:06:46.465 }' 00:06:46.465 23:40:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:06:46.465 23:40:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:06:47.038 23:40:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:06:47.038 23:40:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:47.038 23:40:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:06:47.038 [2024-11-26 23:40:34.937589] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:06:47.038 [2024-11-26 23:40:34.937692] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:06:47.038 23:40:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:47.038 23:40:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:06:47.038 23:40:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:47.038 23:40:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:06:47.038 [2024-11-26 23:40:34.949581] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:06:47.038 [2024-11-26 23:40:34.949682] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:06:47.038 [2024-11-26 23:40:34.949708] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:06:47.038 [2024-11-26 23:40:34.949745] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:06:47.038 23:40:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:47.038 23:40:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:06:47.038 23:40:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:47.038 23:40:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:06:47.038 [2024-11-26 23:40:34.970378] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:06:47.038 BaseBdev1 00:06:47.038 23:40:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:47.038 23:40:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:06:47.038 23:40:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:06:47.038 23:40:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:06:47.038 23:40:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:06:47.038 23:40:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:06:47.038 23:40:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:06:47.038 23:40:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:06:47.038 23:40:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:47.038 23:40:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:06:47.038 23:40:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:47.038 23:40:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:06:47.038 23:40:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:47.038 23:40:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:06:47.038 [ 00:06:47.038 { 00:06:47.038 "name": "BaseBdev1", 00:06:47.038 "aliases": [ 00:06:47.038 "b683746e-0860-43fd-a037-8ba34efb7311" 00:06:47.038 ], 00:06:47.038 "product_name": "Malloc disk", 00:06:47.038 "block_size": 512, 00:06:47.038 "num_blocks": 65536, 00:06:47.038 "uuid": "b683746e-0860-43fd-a037-8ba34efb7311", 00:06:47.038 "assigned_rate_limits": { 00:06:47.038 "rw_ios_per_sec": 0, 00:06:47.038 "rw_mbytes_per_sec": 0, 00:06:47.038 "r_mbytes_per_sec": 0, 00:06:47.038 "w_mbytes_per_sec": 0 00:06:47.038 }, 00:06:47.038 "claimed": true, 00:06:47.038 "claim_type": "exclusive_write", 00:06:47.038 "zoned": false, 00:06:47.038 "supported_io_types": { 00:06:47.038 "read": true, 00:06:47.038 "write": true, 00:06:47.038 "unmap": true, 00:06:47.038 "flush": true, 00:06:47.038 "reset": true, 00:06:47.038 "nvme_admin": false, 00:06:47.038 "nvme_io": false, 00:06:47.038 "nvme_io_md": false, 00:06:47.038 "write_zeroes": true, 00:06:47.038 "zcopy": true, 00:06:47.038 "get_zone_info": false, 00:06:47.038 "zone_management": false, 00:06:47.038 "zone_append": false, 00:06:47.038 "compare": false, 00:06:47.038 "compare_and_write": false, 00:06:47.038 "abort": true, 00:06:47.038 "seek_hole": false, 00:06:47.038 "seek_data": false, 00:06:47.038 "copy": true, 00:06:47.038 "nvme_iov_md": false 00:06:47.038 }, 00:06:47.038 "memory_domains": [ 00:06:47.038 { 00:06:47.038 "dma_device_id": "system", 00:06:47.038 "dma_device_type": 1 00:06:47.038 }, 00:06:47.038 { 00:06:47.038 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:06:47.038 "dma_device_type": 2 00:06:47.038 } 00:06:47.038 ], 00:06:47.038 "driver_specific": {} 00:06:47.038 } 00:06:47.038 ] 00:06:47.038 23:40:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:47.038 23:40:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:06:47.038 23:40:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 2 00:06:47.038 23:40:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:06:47.038 23:40:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:06:47.038 23:40:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:06:47.038 23:40:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:06:47.038 23:40:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:06:47.038 23:40:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:06:47.038 23:40:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:06:47.038 23:40:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:06:47.038 23:40:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:06:47.038 23:40:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:06:47.038 23:40:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:06:47.038 23:40:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:47.038 23:40:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:06:47.038 23:40:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:47.038 23:40:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:06:47.038 "name": "Existed_Raid", 00:06:47.038 "uuid": "c0b08171-e9ac-40b2-b4d6-cc7c37840473", 00:06:47.038 "strip_size_kb": 64, 00:06:47.038 "state": "configuring", 00:06:47.038 "raid_level": "concat", 00:06:47.038 "superblock": true, 00:06:47.038 "num_base_bdevs": 2, 00:06:47.038 "num_base_bdevs_discovered": 1, 00:06:47.038 "num_base_bdevs_operational": 2, 00:06:47.038 "base_bdevs_list": [ 00:06:47.038 { 00:06:47.038 "name": "BaseBdev1", 00:06:47.038 "uuid": "b683746e-0860-43fd-a037-8ba34efb7311", 00:06:47.038 "is_configured": true, 00:06:47.038 "data_offset": 2048, 00:06:47.038 "data_size": 63488 00:06:47.038 }, 00:06:47.038 { 00:06:47.038 "name": "BaseBdev2", 00:06:47.038 "uuid": "00000000-0000-0000-0000-000000000000", 00:06:47.038 "is_configured": false, 00:06:47.038 "data_offset": 0, 00:06:47.038 "data_size": 0 00:06:47.038 } 00:06:47.038 ] 00:06:47.038 }' 00:06:47.038 23:40:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:06:47.038 23:40:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:06:47.604 23:40:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:06:47.604 23:40:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:47.604 23:40:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:06:47.604 [2024-11-26 23:40:35.457567] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:06:47.605 [2024-11-26 23:40:35.457664] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:06:47.605 23:40:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:47.605 23:40:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:06:47.605 23:40:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:47.605 23:40:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:06:47.605 [2024-11-26 23:40:35.465587] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:06:47.605 [2024-11-26 23:40:35.467479] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:06:47.605 [2024-11-26 23:40:35.467554] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:06:47.605 23:40:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:47.605 23:40:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:06:47.605 23:40:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:06:47.605 23:40:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 2 00:06:47.605 23:40:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:06:47.605 23:40:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:06:47.605 23:40:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:06:47.605 23:40:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:06:47.605 23:40:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:06:47.605 23:40:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:06:47.605 23:40:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:06:47.605 23:40:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:06:47.605 23:40:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:06:47.605 23:40:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:06:47.605 23:40:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:06:47.605 23:40:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:47.605 23:40:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:06:47.605 23:40:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:47.605 23:40:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:06:47.605 "name": "Existed_Raid", 00:06:47.605 "uuid": "3667e3cf-74cb-4f00-8072-2ebc556017a2", 00:06:47.605 "strip_size_kb": 64, 00:06:47.605 "state": "configuring", 00:06:47.605 "raid_level": "concat", 00:06:47.605 "superblock": true, 00:06:47.605 "num_base_bdevs": 2, 00:06:47.605 "num_base_bdevs_discovered": 1, 00:06:47.605 "num_base_bdevs_operational": 2, 00:06:47.605 "base_bdevs_list": [ 00:06:47.605 { 00:06:47.605 "name": "BaseBdev1", 00:06:47.605 "uuid": "b683746e-0860-43fd-a037-8ba34efb7311", 00:06:47.605 "is_configured": true, 00:06:47.605 "data_offset": 2048, 00:06:47.605 "data_size": 63488 00:06:47.605 }, 00:06:47.605 { 00:06:47.605 "name": "BaseBdev2", 00:06:47.605 "uuid": "00000000-0000-0000-0000-000000000000", 00:06:47.605 "is_configured": false, 00:06:47.605 "data_offset": 0, 00:06:47.605 "data_size": 0 00:06:47.605 } 00:06:47.605 ] 00:06:47.605 }' 00:06:47.605 23:40:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:06:47.605 23:40:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:06:47.865 23:40:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:06:47.865 23:40:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:47.865 23:40:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:06:47.865 [2024-11-26 23:40:35.895927] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:06:47.865 [2024-11-26 23:40:35.896114] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:06:47.865 [2024-11-26 23:40:35.896131] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:06:47.865 [2024-11-26 23:40:35.896409] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:06:47.865 BaseBdev2 00:06:47.865 [2024-11-26 23:40:35.896557] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:06:47.865 [2024-11-26 23:40:35.896572] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:06:47.865 [2024-11-26 23:40:35.896682] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:06:47.865 23:40:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:47.865 23:40:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:06:47.865 23:40:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:06:47.865 23:40:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:06:47.865 23:40:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:06:47.865 23:40:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:06:47.865 23:40:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:06:47.865 23:40:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:06:47.865 23:40:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:47.865 23:40:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:06:47.865 23:40:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:47.865 23:40:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:06:47.865 23:40:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:47.865 23:40:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:06:47.865 [ 00:06:47.865 { 00:06:47.865 "name": "BaseBdev2", 00:06:47.865 "aliases": [ 00:06:47.865 "b69abfa4-b472-4413-8da7-69ddda0ea9e1" 00:06:47.865 ], 00:06:47.865 "product_name": "Malloc disk", 00:06:47.865 "block_size": 512, 00:06:47.865 "num_blocks": 65536, 00:06:47.865 "uuid": "b69abfa4-b472-4413-8da7-69ddda0ea9e1", 00:06:47.865 "assigned_rate_limits": { 00:06:47.865 "rw_ios_per_sec": 0, 00:06:47.865 "rw_mbytes_per_sec": 0, 00:06:47.865 "r_mbytes_per_sec": 0, 00:06:47.865 "w_mbytes_per_sec": 0 00:06:47.865 }, 00:06:47.865 "claimed": true, 00:06:47.865 "claim_type": "exclusive_write", 00:06:47.865 "zoned": false, 00:06:47.865 "supported_io_types": { 00:06:47.865 "read": true, 00:06:47.865 "write": true, 00:06:47.865 "unmap": true, 00:06:47.865 "flush": true, 00:06:47.865 "reset": true, 00:06:47.865 "nvme_admin": false, 00:06:47.865 "nvme_io": false, 00:06:47.865 "nvme_io_md": false, 00:06:47.865 "write_zeroes": true, 00:06:47.865 "zcopy": true, 00:06:47.865 "get_zone_info": false, 00:06:47.865 "zone_management": false, 00:06:47.865 "zone_append": false, 00:06:47.865 "compare": false, 00:06:47.865 "compare_and_write": false, 00:06:47.865 "abort": true, 00:06:47.865 "seek_hole": false, 00:06:47.865 "seek_data": false, 00:06:47.865 "copy": true, 00:06:47.865 "nvme_iov_md": false 00:06:47.865 }, 00:06:47.865 "memory_domains": [ 00:06:47.865 { 00:06:47.865 "dma_device_id": "system", 00:06:47.865 "dma_device_type": 1 00:06:47.865 }, 00:06:47.865 { 00:06:47.865 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:06:47.865 "dma_device_type": 2 00:06:47.865 } 00:06:47.865 ], 00:06:47.865 "driver_specific": {} 00:06:47.865 } 00:06:47.865 ] 00:06:47.865 23:40:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:47.865 23:40:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:06:47.865 23:40:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:06:47.865 23:40:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:06:47.865 23:40:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online concat 64 2 00:06:47.865 23:40:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:06:47.865 23:40:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:06:47.865 23:40:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:06:47.865 23:40:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:06:47.865 23:40:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:06:47.865 23:40:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:06:47.865 23:40:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:06:47.865 23:40:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:06:47.865 23:40:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:06:47.865 23:40:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:06:47.865 23:40:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:06:47.865 23:40:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:47.865 23:40:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:06:47.865 23:40:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:47.865 23:40:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:06:47.865 "name": "Existed_Raid", 00:06:47.865 "uuid": "3667e3cf-74cb-4f00-8072-2ebc556017a2", 00:06:47.865 "strip_size_kb": 64, 00:06:47.865 "state": "online", 00:06:47.865 "raid_level": "concat", 00:06:47.865 "superblock": true, 00:06:47.865 "num_base_bdevs": 2, 00:06:47.865 "num_base_bdevs_discovered": 2, 00:06:47.865 "num_base_bdevs_operational": 2, 00:06:47.865 "base_bdevs_list": [ 00:06:47.865 { 00:06:47.865 "name": "BaseBdev1", 00:06:47.865 "uuid": "b683746e-0860-43fd-a037-8ba34efb7311", 00:06:47.865 "is_configured": true, 00:06:47.865 "data_offset": 2048, 00:06:47.865 "data_size": 63488 00:06:47.865 }, 00:06:47.865 { 00:06:47.865 "name": "BaseBdev2", 00:06:47.865 "uuid": "b69abfa4-b472-4413-8da7-69ddda0ea9e1", 00:06:47.865 "is_configured": true, 00:06:47.865 "data_offset": 2048, 00:06:47.865 "data_size": 63488 00:06:47.865 } 00:06:47.865 ] 00:06:47.865 }' 00:06:47.865 23:40:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:06:47.866 23:40:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:06:48.435 23:40:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:06:48.435 23:40:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:06:48.435 23:40:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:06:48.435 23:40:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:06:48.435 23:40:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:06:48.435 23:40:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:06:48.435 23:40:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:06:48.435 23:40:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:48.435 23:40:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:06:48.435 23:40:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:06:48.435 [2024-11-26 23:40:36.375471] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:06:48.435 23:40:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:48.435 23:40:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:06:48.435 "name": "Existed_Raid", 00:06:48.435 "aliases": [ 00:06:48.435 "3667e3cf-74cb-4f00-8072-2ebc556017a2" 00:06:48.435 ], 00:06:48.435 "product_name": "Raid Volume", 00:06:48.435 "block_size": 512, 00:06:48.435 "num_blocks": 126976, 00:06:48.435 "uuid": "3667e3cf-74cb-4f00-8072-2ebc556017a2", 00:06:48.435 "assigned_rate_limits": { 00:06:48.435 "rw_ios_per_sec": 0, 00:06:48.435 "rw_mbytes_per_sec": 0, 00:06:48.435 "r_mbytes_per_sec": 0, 00:06:48.435 "w_mbytes_per_sec": 0 00:06:48.435 }, 00:06:48.435 "claimed": false, 00:06:48.435 "zoned": false, 00:06:48.435 "supported_io_types": { 00:06:48.435 "read": true, 00:06:48.435 "write": true, 00:06:48.435 "unmap": true, 00:06:48.435 "flush": true, 00:06:48.435 "reset": true, 00:06:48.435 "nvme_admin": false, 00:06:48.435 "nvme_io": false, 00:06:48.435 "nvme_io_md": false, 00:06:48.435 "write_zeroes": true, 00:06:48.435 "zcopy": false, 00:06:48.435 "get_zone_info": false, 00:06:48.435 "zone_management": false, 00:06:48.435 "zone_append": false, 00:06:48.435 "compare": false, 00:06:48.435 "compare_and_write": false, 00:06:48.435 "abort": false, 00:06:48.435 "seek_hole": false, 00:06:48.435 "seek_data": false, 00:06:48.435 "copy": false, 00:06:48.435 "nvme_iov_md": false 00:06:48.435 }, 00:06:48.435 "memory_domains": [ 00:06:48.435 { 00:06:48.435 "dma_device_id": "system", 00:06:48.435 "dma_device_type": 1 00:06:48.435 }, 00:06:48.435 { 00:06:48.435 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:06:48.435 "dma_device_type": 2 00:06:48.435 }, 00:06:48.435 { 00:06:48.435 "dma_device_id": "system", 00:06:48.435 "dma_device_type": 1 00:06:48.435 }, 00:06:48.435 { 00:06:48.435 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:06:48.435 "dma_device_type": 2 00:06:48.435 } 00:06:48.435 ], 00:06:48.435 "driver_specific": { 00:06:48.435 "raid": { 00:06:48.435 "uuid": "3667e3cf-74cb-4f00-8072-2ebc556017a2", 00:06:48.435 "strip_size_kb": 64, 00:06:48.435 "state": "online", 00:06:48.435 "raid_level": "concat", 00:06:48.435 "superblock": true, 00:06:48.435 "num_base_bdevs": 2, 00:06:48.435 "num_base_bdevs_discovered": 2, 00:06:48.435 "num_base_bdevs_operational": 2, 00:06:48.435 "base_bdevs_list": [ 00:06:48.435 { 00:06:48.435 "name": "BaseBdev1", 00:06:48.435 "uuid": "b683746e-0860-43fd-a037-8ba34efb7311", 00:06:48.435 "is_configured": true, 00:06:48.435 "data_offset": 2048, 00:06:48.435 "data_size": 63488 00:06:48.435 }, 00:06:48.435 { 00:06:48.435 "name": "BaseBdev2", 00:06:48.435 "uuid": "b69abfa4-b472-4413-8da7-69ddda0ea9e1", 00:06:48.435 "is_configured": true, 00:06:48.435 "data_offset": 2048, 00:06:48.435 "data_size": 63488 00:06:48.435 } 00:06:48.435 ] 00:06:48.435 } 00:06:48.435 } 00:06:48.435 }' 00:06:48.435 23:40:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:06:48.435 23:40:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:06:48.435 BaseBdev2' 00:06:48.435 23:40:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:06:48.435 23:40:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:06:48.435 23:40:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:06:48.435 23:40:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:06:48.435 23:40:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:48.436 23:40:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:06:48.436 23:40:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:06:48.436 23:40:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:48.436 23:40:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:06:48.436 23:40:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:06:48.436 23:40:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:06:48.695 23:40:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:06:48.695 23:40:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:06:48.695 23:40:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:48.695 23:40:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:06:48.695 23:40:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:48.695 23:40:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:06:48.695 23:40:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:06:48.695 23:40:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:06:48.695 23:40:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:48.695 23:40:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:06:48.695 [2024-11-26 23:40:36.622831] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:06:48.695 [2024-11-26 23:40:36.622865] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:06:48.695 [2024-11-26 23:40:36.622942] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:06:48.695 23:40:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:48.695 23:40:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:06:48.695 23:40:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy concat 00:06:48.695 23:40:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:06:48.695 23:40:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # return 1 00:06:48.695 23:40:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:06:48.695 23:40:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline concat 64 1 00:06:48.695 23:40:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:06:48.695 23:40:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:06:48.695 23:40:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:06:48.695 23:40:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:06:48.695 23:40:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:06:48.695 23:40:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:06:48.695 23:40:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:06:48.695 23:40:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:06:48.695 23:40:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:06:48.695 23:40:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:06:48.695 23:40:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:48.695 23:40:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:06:48.695 23:40:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:06:48.695 23:40:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:48.695 23:40:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:06:48.695 "name": "Existed_Raid", 00:06:48.695 "uuid": "3667e3cf-74cb-4f00-8072-2ebc556017a2", 00:06:48.695 "strip_size_kb": 64, 00:06:48.695 "state": "offline", 00:06:48.695 "raid_level": "concat", 00:06:48.695 "superblock": true, 00:06:48.695 "num_base_bdevs": 2, 00:06:48.695 "num_base_bdevs_discovered": 1, 00:06:48.695 "num_base_bdevs_operational": 1, 00:06:48.695 "base_bdevs_list": [ 00:06:48.695 { 00:06:48.695 "name": null, 00:06:48.695 "uuid": "00000000-0000-0000-0000-000000000000", 00:06:48.695 "is_configured": false, 00:06:48.695 "data_offset": 0, 00:06:48.695 "data_size": 63488 00:06:48.695 }, 00:06:48.695 { 00:06:48.695 "name": "BaseBdev2", 00:06:48.695 "uuid": "b69abfa4-b472-4413-8da7-69ddda0ea9e1", 00:06:48.695 "is_configured": true, 00:06:48.695 "data_offset": 2048, 00:06:48.695 "data_size": 63488 00:06:48.695 } 00:06:48.695 ] 00:06:48.695 }' 00:06:48.695 23:40:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:06:48.695 23:40:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:06:48.954 23:40:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:06:48.954 23:40:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:06:48.954 23:40:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:06:48.954 23:40:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:06:48.954 23:40:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:48.954 23:40:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:06:48.954 23:40:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:48.954 23:40:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:06:48.954 23:40:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:06:48.954 23:40:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:06:48.954 23:40:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:48.954 23:40:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:06:48.954 [2024-11-26 23:40:37.073326] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:06:48.954 [2024-11-26 23:40:37.073432] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:06:49.212 23:40:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:49.212 23:40:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:06:49.212 23:40:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:06:49.212 23:40:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:06:49.212 23:40:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:49.212 23:40:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:06:49.212 23:40:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:06:49.212 23:40:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:49.212 23:40:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:06:49.212 23:40:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:06:49.212 23:40:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 2 -gt 2 ']' 00:06:49.212 23:40:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 73009 00:06:49.212 23:40:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # '[' -z 73009 ']' 00:06:49.212 23:40:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@958 -- # kill -0 73009 00:06:49.212 23:40:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # uname 00:06:49.212 23:40:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:06:49.212 23:40:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 73009 00:06:49.212 killing process with pid 73009 00:06:49.212 23:40:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:06:49.212 23:40:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:06:49.212 23:40:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@972 -- # echo 'killing process with pid 73009' 00:06:49.212 23:40:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@973 -- # kill 73009 00:06:49.212 [2024-11-26 23:40:37.166062] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:06:49.212 23:40:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@978 -- # wait 73009 00:06:49.212 [2024-11-26 23:40:37.167039] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:06:49.470 23:40:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:06:49.470 00:06:49.470 real 0m3.794s 00:06:49.470 user 0m6.023s 00:06:49.470 sys 0m0.707s 00:06:49.470 23:40:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:49.470 23:40:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:06:49.470 ************************************ 00:06:49.470 END TEST raid_state_function_test_sb 00:06:49.470 ************************************ 00:06:49.470 23:40:37 bdev_raid -- bdev/bdev_raid.sh@970 -- # run_test raid_superblock_test raid_superblock_test concat 2 00:06:49.470 23:40:37 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:06:49.470 23:40:37 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:49.470 23:40:37 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:06:49.470 ************************************ 00:06:49.470 START TEST raid_superblock_test 00:06:49.470 ************************************ 00:06:49.470 23:40:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1129 -- # raid_superblock_test concat 2 00:06:49.470 23:40:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=concat 00:06:49.470 23:40:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=2 00:06:49.470 23:40:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:06:49.470 23:40:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:06:49.470 23:40:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:06:49.470 23:40:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:06:49.470 23:40:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:06:49.470 23:40:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:06:49.470 23:40:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:06:49.470 23:40:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:06:49.470 23:40:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:06:49.470 23:40:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:06:49.470 23:40:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:06:49.470 23:40:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' concat '!=' raid1 ']' 00:06:49.470 23:40:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@405 -- # strip_size=64 00:06:49.471 23:40:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@406 -- # strip_size_create_arg='-z 64' 00:06:49.471 23:40:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=73239 00:06:49.471 23:40:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:06:49.471 23:40:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 73239 00:06:49.471 23:40:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@835 -- # '[' -z 73239 ']' 00:06:49.471 23:40:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:49.471 23:40:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:49.471 23:40:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:49.471 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:49.471 23:40:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:49.471 23:40:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:49.471 [2024-11-26 23:40:37.530511] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:06:49.471 [2024-11-26 23:40:37.530736] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid73239 ] 00:06:49.730 [2024-11-26 23:40:37.663056] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:49.730 [2024-11-26 23:40:37.688964] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:49.730 [2024-11-26 23:40:37.730848] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:06:49.730 [2024-11-26 23:40:37.730959] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:06:50.298 23:40:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:50.298 23:40:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@868 -- # return 0 00:06:50.298 23:40:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:06:50.298 23:40:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:06:50.298 23:40:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:06:50.298 23:40:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:06:50.298 23:40:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:06:50.298 23:40:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:06:50.298 23:40:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:06:50.298 23:40:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:06:50.298 23:40:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:06:50.298 23:40:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:50.298 23:40:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:50.298 malloc1 00:06:50.298 23:40:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:50.298 23:40:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:06:50.298 23:40:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:50.298 23:40:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:50.298 [2024-11-26 23:40:38.377380] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:06:50.298 [2024-11-26 23:40:38.377437] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:06:50.298 [2024-11-26 23:40:38.377473] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:06:50.298 [2024-11-26 23:40:38.377486] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:06:50.299 [2024-11-26 23:40:38.379590] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:06:50.299 [2024-11-26 23:40:38.379626] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:06:50.299 pt1 00:06:50.299 23:40:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:50.299 23:40:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:06:50.299 23:40:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:06:50.299 23:40:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:06:50.299 23:40:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:06:50.299 23:40:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:06:50.299 23:40:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:06:50.299 23:40:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:06:50.299 23:40:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:06:50.299 23:40:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:06:50.299 23:40:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:50.299 23:40:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:50.299 malloc2 00:06:50.299 23:40:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:50.299 23:40:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:06:50.299 23:40:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:50.299 23:40:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:50.299 [2024-11-26 23:40:38.405612] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:06:50.299 [2024-11-26 23:40:38.405724] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:06:50.299 [2024-11-26 23:40:38.405759] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:06:50.299 [2024-11-26 23:40:38.405791] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:06:50.299 [2024-11-26 23:40:38.407880] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:06:50.299 [2024-11-26 23:40:38.407964] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:06:50.299 pt2 00:06:50.299 23:40:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:50.299 23:40:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:06:50.299 23:40:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:06:50.299 23:40:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''pt1 pt2'\''' -n raid_bdev1 -s 00:06:50.299 23:40:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:50.299 23:40:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:50.299 [2024-11-26 23:40:38.417630] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:06:50.299 [2024-11-26 23:40:38.419431] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:06:50.299 [2024-11-26 23:40:38.419611] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:06:50.299 [2024-11-26 23:40:38.419659] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:06:50.299 [2024-11-26 23:40:38.419953] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:06:50.299 [2024-11-26 23:40:38.420132] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:06:50.299 [2024-11-26 23:40:38.420175] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:06:50.299 [2024-11-26 23:40:38.420328] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:06:50.299 23:40:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:50.299 23:40:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online concat 64 2 00:06:50.299 23:40:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:06:50.299 23:40:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:06:50.299 23:40:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:06:50.299 23:40:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:06:50.299 23:40:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:06:50.299 23:40:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:06:50.299 23:40:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:06:50.299 23:40:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:06:50.299 23:40:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:06:50.559 23:40:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:06:50.559 23:40:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:06:50.559 23:40:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:50.559 23:40:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:50.559 23:40:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:50.559 23:40:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:06:50.559 "name": "raid_bdev1", 00:06:50.559 "uuid": "cc5248e7-fb7d-4751-8f8e-6a0d7269301c", 00:06:50.559 "strip_size_kb": 64, 00:06:50.559 "state": "online", 00:06:50.559 "raid_level": "concat", 00:06:50.559 "superblock": true, 00:06:50.559 "num_base_bdevs": 2, 00:06:50.559 "num_base_bdevs_discovered": 2, 00:06:50.559 "num_base_bdevs_operational": 2, 00:06:50.559 "base_bdevs_list": [ 00:06:50.559 { 00:06:50.559 "name": "pt1", 00:06:50.559 "uuid": "00000000-0000-0000-0000-000000000001", 00:06:50.559 "is_configured": true, 00:06:50.559 "data_offset": 2048, 00:06:50.559 "data_size": 63488 00:06:50.559 }, 00:06:50.559 { 00:06:50.559 "name": "pt2", 00:06:50.559 "uuid": "00000000-0000-0000-0000-000000000002", 00:06:50.559 "is_configured": true, 00:06:50.559 "data_offset": 2048, 00:06:50.559 "data_size": 63488 00:06:50.559 } 00:06:50.559 ] 00:06:50.559 }' 00:06:50.559 23:40:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:06:50.559 23:40:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:50.819 23:40:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:06:50.820 23:40:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:06:50.820 23:40:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:06:50.820 23:40:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:06:50.820 23:40:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:06:50.820 23:40:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:06:50.820 23:40:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:06:50.820 23:40:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:06:50.820 23:40:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:50.820 23:40:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:50.820 [2024-11-26 23:40:38.813297] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:06:50.820 23:40:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:50.820 23:40:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:06:50.820 "name": "raid_bdev1", 00:06:50.820 "aliases": [ 00:06:50.820 "cc5248e7-fb7d-4751-8f8e-6a0d7269301c" 00:06:50.820 ], 00:06:50.820 "product_name": "Raid Volume", 00:06:50.820 "block_size": 512, 00:06:50.820 "num_blocks": 126976, 00:06:50.820 "uuid": "cc5248e7-fb7d-4751-8f8e-6a0d7269301c", 00:06:50.820 "assigned_rate_limits": { 00:06:50.820 "rw_ios_per_sec": 0, 00:06:50.820 "rw_mbytes_per_sec": 0, 00:06:50.820 "r_mbytes_per_sec": 0, 00:06:50.820 "w_mbytes_per_sec": 0 00:06:50.820 }, 00:06:50.820 "claimed": false, 00:06:50.820 "zoned": false, 00:06:50.820 "supported_io_types": { 00:06:50.820 "read": true, 00:06:50.820 "write": true, 00:06:50.820 "unmap": true, 00:06:50.820 "flush": true, 00:06:50.820 "reset": true, 00:06:50.820 "nvme_admin": false, 00:06:50.820 "nvme_io": false, 00:06:50.820 "nvme_io_md": false, 00:06:50.820 "write_zeroes": true, 00:06:50.820 "zcopy": false, 00:06:50.820 "get_zone_info": false, 00:06:50.820 "zone_management": false, 00:06:50.820 "zone_append": false, 00:06:50.820 "compare": false, 00:06:50.820 "compare_and_write": false, 00:06:50.820 "abort": false, 00:06:50.820 "seek_hole": false, 00:06:50.820 "seek_data": false, 00:06:50.820 "copy": false, 00:06:50.820 "nvme_iov_md": false 00:06:50.820 }, 00:06:50.820 "memory_domains": [ 00:06:50.820 { 00:06:50.820 "dma_device_id": "system", 00:06:50.820 "dma_device_type": 1 00:06:50.820 }, 00:06:50.820 { 00:06:50.820 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:06:50.820 "dma_device_type": 2 00:06:50.820 }, 00:06:50.820 { 00:06:50.820 "dma_device_id": "system", 00:06:50.820 "dma_device_type": 1 00:06:50.820 }, 00:06:50.820 { 00:06:50.820 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:06:50.820 "dma_device_type": 2 00:06:50.820 } 00:06:50.820 ], 00:06:50.820 "driver_specific": { 00:06:50.820 "raid": { 00:06:50.820 "uuid": "cc5248e7-fb7d-4751-8f8e-6a0d7269301c", 00:06:50.820 "strip_size_kb": 64, 00:06:50.820 "state": "online", 00:06:50.820 "raid_level": "concat", 00:06:50.820 "superblock": true, 00:06:50.820 "num_base_bdevs": 2, 00:06:50.820 "num_base_bdevs_discovered": 2, 00:06:50.820 "num_base_bdevs_operational": 2, 00:06:50.820 "base_bdevs_list": [ 00:06:50.820 { 00:06:50.820 "name": "pt1", 00:06:50.820 "uuid": "00000000-0000-0000-0000-000000000001", 00:06:50.820 "is_configured": true, 00:06:50.820 "data_offset": 2048, 00:06:50.820 "data_size": 63488 00:06:50.820 }, 00:06:50.820 { 00:06:50.820 "name": "pt2", 00:06:50.820 "uuid": "00000000-0000-0000-0000-000000000002", 00:06:50.820 "is_configured": true, 00:06:50.820 "data_offset": 2048, 00:06:50.820 "data_size": 63488 00:06:50.820 } 00:06:50.820 ] 00:06:50.820 } 00:06:50.820 } 00:06:50.820 }' 00:06:50.820 23:40:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:06:50.820 23:40:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:06:50.820 pt2' 00:06:50.820 23:40:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:06:50.820 23:40:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:06:50.820 23:40:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:06:50.820 23:40:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:06:51.082 23:40:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:51.082 23:40:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:51.082 23:40:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:06:51.082 23:40:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:51.082 23:40:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:06:51.083 23:40:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:06:51.083 23:40:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:06:51.083 23:40:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:06:51.083 23:40:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:51.083 23:40:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:06:51.083 23:40:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:51.083 23:40:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:51.083 23:40:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:06:51.083 23:40:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:06:51.083 23:40:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:06:51.083 23:40:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:51.083 23:40:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:51.083 23:40:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:06:51.083 [2024-11-26 23:40:39.056746] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:06:51.083 23:40:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:51.083 23:40:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=cc5248e7-fb7d-4751-8f8e-6a0d7269301c 00:06:51.083 23:40:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z cc5248e7-fb7d-4751-8f8e-6a0d7269301c ']' 00:06:51.083 23:40:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:06:51.083 23:40:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:51.083 23:40:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:51.083 [2024-11-26 23:40:39.104457] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:06:51.083 [2024-11-26 23:40:39.104483] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:06:51.083 [2024-11-26 23:40:39.104548] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:06:51.083 [2024-11-26 23:40:39.104596] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:06:51.083 [2024-11-26 23:40:39.104614] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:06:51.083 23:40:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:51.083 23:40:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:06:51.083 23:40:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:06:51.083 23:40:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:51.083 23:40:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:51.083 23:40:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:51.083 23:40:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:06:51.083 23:40:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:06:51.083 23:40:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:06:51.083 23:40:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:06:51.083 23:40:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:51.083 23:40:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:51.083 23:40:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:51.083 23:40:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:06:51.083 23:40:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:06:51.083 23:40:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:51.083 23:40:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:51.083 23:40:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:51.083 23:40:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:06:51.083 23:40:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:06:51.083 23:40:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:51.083 23:40:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:51.083 23:40:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:51.342 23:40:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:06:51.342 23:40:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:06:51.342 23:40:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@652 -- # local es=0 00:06:51.343 23:40:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:06:51.343 23:40:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:06:51.343 23:40:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:06:51.343 23:40:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:06:51.343 23:40:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:06:51.343 23:40:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:06:51.343 23:40:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:51.343 23:40:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:51.343 [2024-11-26 23:40:39.240303] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:06:51.343 [2024-11-26 23:40:39.242146] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:06:51.343 [2024-11-26 23:40:39.242230] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:06:51.343 [2024-11-26 23:40:39.242273] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:06:51.343 [2024-11-26 23:40:39.242288] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:06:51.343 [2024-11-26 23:40:39.242296] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state configuring 00:06:51.343 request: 00:06:51.343 { 00:06:51.343 "name": "raid_bdev1", 00:06:51.343 "raid_level": "concat", 00:06:51.343 "base_bdevs": [ 00:06:51.343 "malloc1", 00:06:51.343 "malloc2" 00:06:51.343 ], 00:06:51.343 "strip_size_kb": 64, 00:06:51.343 "superblock": false, 00:06:51.343 "method": "bdev_raid_create", 00:06:51.343 "req_id": 1 00:06:51.343 } 00:06:51.343 Got JSON-RPC error response 00:06:51.343 response: 00:06:51.343 { 00:06:51.343 "code": -17, 00:06:51.343 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:06:51.343 } 00:06:51.343 23:40:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:06:51.343 23:40:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # es=1 00:06:51.343 23:40:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:06:51.343 23:40:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:06:51.343 23:40:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:06:51.343 23:40:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:06:51.343 23:40:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:06:51.343 23:40:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:51.343 23:40:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:51.343 23:40:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:51.343 23:40:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:06:51.343 23:40:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:06:51.343 23:40:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:06:51.343 23:40:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:51.343 23:40:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:51.343 [2024-11-26 23:40:39.304180] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:06:51.343 [2024-11-26 23:40:39.304236] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:06:51.343 [2024-11-26 23:40:39.304260] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:06:51.343 [2024-11-26 23:40:39.304268] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:06:51.343 [2024-11-26 23:40:39.306415] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:06:51.343 [2024-11-26 23:40:39.306493] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:06:51.343 [2024-11-26 23:40:39.306569] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:06:51.343 [2024-11-26 23:40:39.306601] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:06:51.343 pt1 00:06:51.343 23:40:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:51.343 23:40:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring concat 64 2 00:06:51.343 23:40:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:06:51.343 23:40:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:06:51.343 23:40:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:06:51.343 23:40:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:06:51.343 23:40:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:06:51.343 23:40:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:06:51.343 23:40:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:06:51.343 23:40:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:06:51.343 23:40:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:06:51.343 23:40:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:06:51.343 23:40:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:06:51.343 23:40:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:51.343 23:40:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:51.343 23:40:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:51.343 23:40:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:06:51.343 "name": "raid_bdev1", 00:06:51.343 "uuid": "cc5248e7-fb7d-4751-8f8e-6a0d7269301c", 00:06:51.343 "strip_size_kb": 64, 00:06:51.343 "state": "configuring", 00:06:51.343 "raid_level": "concat", 00:06:51.343 "superblock": true, 00:06:51.343 "num_base_bdevs": 2, 00:06:51.343 "num_base_bdevs_discovered": 1, 00:06:51.343 "num_base_bdevs_operational": 2, 00:06:51.343 "base_bdevs_list": [ 00:06:51.343 { 00:06:51.343 "name": "pt1", 00:06:51.343 "uuid": "00000000-0000-0000-0000-000000000001", 00:06:51.343 "is_configured": true, 00:06:51.343 "data_offset": 2048, 00:06:51.343 "data_size": 63488 00:06:51.343 }, 00:06:51.343 { 00:06:51.343 "name": null, 00:06:51.343 "uuid": "00000000-0000-0000-0000-000000000002", 00:06:51.343 "is_configured": false, 00:06:51.343 "data_offset": 2048, 00:06:51.343 "data_size": 63488 00:06:51.343 } 00:06:51.343 ] 00:06:51.343 }' 00:06:51.343 23:40:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:06:51.343 23:40:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:51.602 23:40:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 2 -gt 2 ']' 00:06:51.602 23:40:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:06:51.602 23:40:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:06:51.602 23:40:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:06:51.602 23:40:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:51.602 23:40:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:51.862 [2024-11-26 23:40:39.735452] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:06:51.862 [2024-11-26 23:40:39.735563] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:06:51.862 [2024-11-26 23:40:39.735602] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:06:51.862 [2024-11-26 23:40:39.735630] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:06:51.862 [2024-11-26 23:40:39.736062] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:06:51.862 [2024-11-26 23:40:39.736123] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:06:51.862 [2024-11-26 23:40:39.736236] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:06:51.862 [2024-11-26 23:40:39.736284] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:06:51.862 [2024-11-26 23:40:39.736408] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:06:51.862 [2024-11-26 23:40:39.736445] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:06:51.862 [2024-11-26 23:40:39.736759] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:06:51.862 [2024-11-26 23:40:39.736918] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:06:51.862 [2024-11-26 23:40:39.736965] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:06:51.862 [2024-11-26 23:40:39.737127] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:06:51.862 pt2 00:06:51.862 23:40:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:51.862 23:40:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:06:51.862 23:40:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:06:51.862 23:40:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online concat 64 2 00:06:51.862 23:40:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:06:51.862 23:40:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:06:51.862 23:40:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:06:51.862 23:40:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:06:51.862 23:40:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:06:51.862 23:40:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:06:51.862 23:40:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:06:51.862 23:40:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:06:51.862 23:40:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:06:51.862 23:40:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:06:51.862 23:40:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:06:51.862 23:40:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:51.862 23:40:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:51.862 23:40:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:51.862 23:40:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:06:51.862 "name": "raid_bdev1", 00:06:51.862 "uuid": "cc5248e7-fb7d-4751-8f8e-6a0d7269301c", 00:06:51.862 "strip_size_kb": 64, 00:06:51.862 "state": "online", 00:06:51.862 "raid_level": "concat", 00:06:51.862 "superblock": true, 00:06:51.862 "num_base_bdevs": 2, 00:06:51.862 "num_base_bdevs_discovered": 2, 00:06:51.862 "num_base_bdevs_operational": 2, 00:06:51.862 "base_bdevs_list": [ 00:06:51.862 { 00:06:51.862 "name": "pt1", 00:06:51.862 "uuid": "00000000-0000-0000-0000-000000000001", 00:06:51.862 "is_configured": true, 00:06:51.862 "data_offset": 2048, 00:06:51.862 "data_size": 63488 00:06:51.862 }, 00:06:51.862 { 00:06:51.862 "name": "pt2", 00:06:51.862 "uuid": "00000000-0000-0000-0000-000000000002", 00:06:51.862 "is_configured": true, 00:06:51.862 "data_offset": 2048, 00:06:51.862 "data_size": 63488 00:06:51.862 } 00:06:51.862 ] 00:06:51.862 }' 00:06:51.862 23:40:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:06:51.862 23:40:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:52.123 23:40:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:06:52.123 23:40:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:06:52.123 23:40:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:06:52.123 23:40:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:06:52.123 23:40:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:06:52.123 23:40:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:06:52.123 23:40:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:06:52.123 23:40:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:52.123 23:40:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:06:52.123 23:40:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:52.123 [2024-11-26 23:40:40.162975] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:06:52.123 23:40:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:52.123 23:40:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:06:52.123 "name": "raid_bdev1", 00:06:52.123 "aliases": [ 00:06:52.123 "cc5248e7-fb7d-4751-8f8e-6a0d7269301c" 00:06:52.123 ], 00:06:52.123 "product_name": "Raid Volume", 00:06:52.123 "block_size": 512, 00:06:52.123 "num_blocks": 126976, 00:06:52.123 "uuid": "cc5248e7-fb7d-4751-8f8e-6a0d7269301c", 00:06:52.123 "assigned_rate_limits": { 00:06:52.123 "rw_ios_per_sec": 0, 00:06:52.123 "rw_mbytes_per_sec": 0, 00:06:52.123 "r_mbytes_per_sec": 0, 00:06:52.123 "w_mbytes_per_sec": 0 00:06:52.123 }, 00:06:52.123 "claimed": false, 00:06:52.123 "zoned": false, 00:06:52.123 "supported_io_types": { 00:06:52.123 "read": true, 00:06:52.123 "write": true, 00:06:52.123 "unmap": true, 00:06:52.123 "flush": true, 00:06:52.123 "reset": true, 00:06:52.123 "nvme_admin": false, 00:06:52.123 "nvme_io": false, 00:06:52.123 "nvme_io_md": false, 00:06:52.123 "write_zeroes": true, 00:06:52.123 "zcopy": false, 00:06:52.123 "get_zone_info": false, 00:06:52.123 "zone_management": false, 00:06:52.123 "zone_append": false, 00:06:52.123 "compare": false, 00:06:52.123 "compare_and_write": false, 00:06:52.123 "abort": false, 00:06:52.123 "seek_hole": false, 00:06:52.123 "seek_data": false, 00:06:52.123 "copy": false, 00:06:52.123 "nvme_iov_md": false 00:06:52.123 }, 00:06:52.123 "memory_domains": [ 00:06:52.123 { 00:06:52.123 "dma_device_id": "system", 00:06:52.123 "dma_device_type": 1 00:06:52.123 }, 00:06:52.123 { 00:06:52.123 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:06:52.123 "dma_device_type": 2 00:06:52.123 }, 00:06:52.123 { 00:06:52.123 "dma_device_id": "system", 00:06:52.123 "dma_device_type": 1 00:06:52.123 }, 00:06:52.123 { 00:06:52.123 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:06:52.123 "dma_device_type": 2 00:06:52.123 } 00:06:52.123 ], 00:06:52.123 "driver_specific": { 00:06:52.123 "raid": { 00:06:52.123 "uuid": "cc5248e7-fb7d-4751-8f8e-6a0d7269301c", 00:06:52.123 "strip_size_kb": 64, 00:06:52.123 "state": "online", 00:06:52.123 "raid_level": "concat", 00:06:52.123 "superblock": true, 00:06:52.123 "num_base_bdevs": 2, 00:06:52.123 "num_base_bdevs_discovered": 2, 00:06:52.123 "num_base_bdevs_operational": 2, 00:06:52.123 "base_bdevs_list": [ 00:06:52.123 { 00:06:52.123 "name": "pt1", 00:06:52.123 "uuid": "00000000-0000-0000-0000-000000000001", 00:06:52.123 "is_configured": true, 00:06:52.123 "data_offset": 2048, 00:06:52.123 "data_size": 63488 00:06:52.123 }, 00:06:52.123 { 00:06:52.123 "name": "pt2", 00:06:52.123 "uuid": "00000000-0000-0000-0000-000000000002", 00:06:52.123 "is_configured": true, 00:06:52.123 "data_offset": 2048, 00:06:52.123 "data_size": 63488 00:06:52.123 } 00:06:52.123 ] 00:06:52.123 } 00:06:52.123 } 00:06:52.123 }' 00:06:52.123 23:40:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:06:52.123 23:40:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:06:52.123 pt2' 00:06:52.382 23:40:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:06:52.382 23:40:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:06:52.382 23:40:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:06:52.382 23:40:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:06:52.382 23:40:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:06:52.382 23:40:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:52.382 23:40:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:52.382 23:40:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:52.382 23:40:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:06:52.382 23:40:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:06:52.382 23:40:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:06:52.382 23:40:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:06:52.382 23:40:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:06:52.382 23:40:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:52.382 23:40:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:52.382 23:40:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:52.382 23:40:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:06:52.382 23:40:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:06:52.382 23:40:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:06:52.382 23:40:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:06:52.382 23:40:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:52.382 23:40:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:52.382 [2024-11-26 23:40:40.414533] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:06:52.382 23:40:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:52.382 23:40:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' cc5248e7-fb7d-4751-8f8e-6a0d7269301c '!=' cc5248e7-fb7d-4751-8f8e-6a0d7269301c ']' 00:06:52.382 23:40:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy concat 00:06:52.382 23:40:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:06:52.382 23:40:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # return 1 00:06:52.382 23:40:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 73239 00:06:52.382 23:40:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # '[' -z 73239 ']' 00:06:52.382 23:40:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@958 -- # kill -0 73239 00:06:52.382 23:40:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # uname 00:06:52.382 23:40:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:06:52.382 23:40:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 73239 00:06:52.382 killing process with pid 73239 00:06:52.382 23:40:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:06:52.382 23:40:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:06:52.382 23:40:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 73239' 00:06:52.382 23:40:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@973 -- # kill 73239 00:06:52.382 [2024-11-26 23:40:40.486252] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:06:52.382 [2024-11-26 23:40:40.486352] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:06:52.382 [2024-11-26 23:40:40.486403] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:06:52.382 [2024-11-26 23:40:40.486413] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:06:52.382 23:40:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@978 -- # wait 73239 00:06:52.382 [2024-11-26 23:40:40.508627] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:06:52.641 23:40:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:06:52.641 00:06:52.641 real 0m3.275s 00:06:52.641 user 0m5.100s 00:06:52.641 sys 0m0.662s 00:06:52.641 23:40:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:52.641 23:40:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:52.641 ************************************ 00:06:52.641 END TEST raid_superblock_test 00:06:52.641 ************************************ 00:06:52.900 23:40:40 bdev_raid -- bdev/bdev_raid.sh@971 -- # run_test raid_read_error_test raid_io_error_test concat 2 read 00:06:52.900 23:40:40 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:06:52.900 23:40:40 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:52.900 23:40:40 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:06:52.900 ************************************ 00:06:52.900 START TEST raid_read_error_test 00:06:52.900 ************************************ 00:06:52.900 23:40:40 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test concat 2 read 00:06:52.900 23:40:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=concat 00:06:52.900 23:40:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=2 00:06:52.900 23:40:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=read 00:06:52.900 23:40:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:06:52.900 23:40:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:06:52.900 23:40:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:06:52.900 23:40:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:06:52.900 23:40:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:06:52.900 23:40:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:06:52.900 23:40:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:06:52.900 23:40:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:06:52.900 23:40:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:06:52.900 23:40:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:06:52.900 23:40:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:06:52.900 23:40:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:06:52.900 23:40:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:06:52.900 23:40:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:06:52.900 23:40:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:06:52.900 23:40:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # '[' concat '!=' raid1 ']' 00:06:52.900 23:40:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:06:52.900 23:40:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:06:52.900 23:40:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:06:52.900 23:40:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.mQmk5JIVfk 00:06:52.900 23:40:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=73440 00:06:52.900 23:40:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:06:52.900 23:40:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 73440 00:06:52.900 23:40:40 bdev_raid.raid_read_error_test -- common/autotest_common.sh@835 -- # '[' -z 73440 ']' 00:06:52.900 23:40:40 bdev_raid.raid_read_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:52.900 23:40:40 bdev_raid.raid_read_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:52.900 23:40:40 bdev_raid.raid_read_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:52.900 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:52.900 23:40:40 bdev_raid.raid_read_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:52.900 23:40:40 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:06:52.900 [2024-11-26 23:40:40.891010] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:06:52.900 [2024-11-26 23:40:40.891193] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid73440 ] 00:06:53.158 [2024-11-26 23:40:41.044302] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:53.158 [2024-11-26 23:40:41.068999] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:53.158 [2024-11-26 23:40:41.110208] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:06:53.158 [2024-11-26 23:40:41.110322] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:06:53.727 23:40:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:53.727 23:40:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@868 -- # return 0 00:06:53.727 23:40:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:06:53.727 23:40:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:06:53.727 23:40:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:53.727 23:40:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:06:53.727 BaseBdev1_malloc 00:06:53.727 23:40:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:53.727 23:40:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:06:53.727 23:40:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:53.727 23:40:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:06:53.727 true 00:06:53.727 23:40:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:53.727 23:40:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:06:53.727 23:40:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:53.727 23:40:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:06:53.727 [2024-11-26 23:40:41.748897] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:06:53.727 [2024-11-26 23:40:41.748953] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:06:53.727 [2024-11-26 23:40:41.748973] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006980 00:06:53.727 [2024-11-26 23:40:41.748982] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:06:53.727 [2024-11-26 23:40:41.751057] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:06:53.727 [2024-11-26 23:40:41.751093] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:06:53.727 BaseBdev1 00:06:53.727 23:40:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:53.727 23:40:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:06:53.727 23:40:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:06:53.727 23:40:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:53.727 23:40:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:06:53.727 BaseBdev2_malloc 00:06:53.728 23:40:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:53.728 23:40:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:06:53.728 23:40:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:53.728 23:40:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:06:53.728 true 00:06:53.728 23:40:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:53.728 23:40:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:06:53.728 23:40:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:53.728 23:40:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:06:53.728 [2024-11-26 23:40:41.789322] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:06:53.728 [2024-11-26 23:40:41.789432] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:06:53.728 [2024-11-26 23:40:41.789455] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:06:53.728 [2024-11-26 23:40:41.789473] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:06:53.728 [2024-11-26 23:40:41.791589] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:06:53.728 [2024-11-26 23:40:41.791623] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:06:53.728 BaseBdev2 00:06:53.728 23:40:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:53.728 23:40:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 -s 00:06:53.728 23:40:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:53.728 23:40:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:06:53.728 [2024-11-26 23:40:41.801353] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:06:53.728 [2024-11-26 23:40:41.803200] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:06:53.728 [2024-11-26 23:40:41.803446] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:06:53.728 [2024-11-26 23:40:41.803492] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:06:53.728 [2024-11-26 23:40:41.803746] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002870 00:06:53.728 [2024-11-26 23:40:41.803927] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:06:53.728 [2024-11-26 23:40:41.803970] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:06:53.728 [2024-11-26 23:40:41.804127] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:06:53.728 23:40:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:53.728 23:40:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online concat 64 2 00:06:53.728 23:40:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:06:53.728 23:40:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:06:53.728 23:40:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:06:53.728 23:40:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:06:53.728 23:40:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:06:53.728 23:40:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:06:53.728 23:40:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:06:53.728 23:40:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:06:53.728 23:40:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:06:53.728 23:40:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:06:53.728 23:40:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:06:53.728 23:40:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:53.728 23:40:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:06:53.728 23:40:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:53.988 23:40:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:06:53.988 "name": "raid_bdev1", 00:06:53.988 "uuid": "a64c1031-ef60-4ed3-83e2-f880b5ddf49a", 00:06:53.988 "strip_size_kb": 64, 00:06:53.988 "state": "online", 00:06:53.988 "raid_level": "concat", 00:06:53.988 "superblock": true, 00:06:53.988 "num_base_bdevs": 2, 00:06:53.988 "num_base_bdevs_discovered": 2, 00:06:53.988 "num_base_bdevs_operational": 2, 00:06:53.988 "base_bdevs_list": [ 00:06:53.988 { 00:06:53.988 "name": "BaseBdev1", 00:06:53.988 "uuid": "4aca95f5-2ac8-573f-a293-4517d09cfd04", 00:06:53.988 "is_configured": true, 00:06:53.988 "data_offset": 2048, 00:06:53.988 "data_size": 63488 00:06:53.988 }, 00:06:53.988 { 00:06:53.988 "name": "BaseBdev2", 00:06:53.988 "uuid": "06748dc5-3ef7-5b4c-86e7-1b26c0a39181", 00:06:53.988 "is_configured": true, 00:06:53.988 "data_offset": 2048, 00:06:53.988 "data_size": 63488 00:06:53.988 } 00:06:53.988 ] 00:06:53.988 }' 00:06:53.988 23:40:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:06:53.988 23:40:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:06:54.246 23:40:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:06:54.246 23:40:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:06:54.246 [2024-11-26 23:40:42.352754] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002a10 00:06:55.181 23:40:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:06:55.181 23:40:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:55.181 23:40:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:06:55.181 23:40:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:55.181 23:40:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:06:55.181 23:40:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ concat = \r\a\i\d\1 ]] 00:06:55.181 23:40:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=2 00:06:55.181 23:40:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online concat 64 2 00:06:55.181 23:40:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:06:55.181 23:40:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:06:55.181 23:40:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:06:55.181 23:40:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:06:55.181 23:40:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:06:55.181 23:40:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:06:55.181 23:40:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:06:55.181 23:40:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:06:55.181 23:40:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:06:55.181 23:40:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:06:55.181 23:40:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:55.181 23:40:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:06:55.181 23:40:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:06:55.441 23:40:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:55.441 23:40:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:06:55.441 "name": "raid_bdev1", 00:06:55.441 "uuid": "a64c1031-ef60-4ed3-83e2-f880b5ddf49a", 00:06:55.441 "strip_size_kb": 64, 00:06:55.441 "state": "online", 00:06:55.441 "raid_level": "concat", 00:06:55.441 "superblock": true, 00:06:55.441 "num_base_bdevs": 2, 00:06:55.441 "num_base_bdevs_discovered": 2, 00:06:55.441 "num_base_bdevs_operational": 2, 00:06:55.441 "base_bdevs_list": [ 00:06:55.441 { 00:06:55.441 "name": "BaseBdev1", 00:06:55.441 "uuid": "4aca95f5-2ac8-573f-a293-4517d09cfd04", 00:06:55.441 "is_configured": true, 00:06:55.441 "data_offset": 2048, 00:06:55.441 "data_size": 63488 00:06:55.441 }, 00:06:55.441 { 00:06:55.441 "name": "BaseBdev2", 00:06:55.441 "uuid": "06748dc5-3ef7-5b4c-86e7-1b26c0a39181", 00:06:55.441 "is_configured": true, 00:06:55.441 "data_offset": 2048, 00:06:55.441 "data_size": 63488 00:06:55.441 } 00:06:55.441 ] 00:06:55.441 }' 00:06:55.441 23:40:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:06:55.441 23:40:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:06:55.699 23:40:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:06:55.699 23:40:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:55.699 23:40:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:06:55.699 [2024-11-26 23:40:43.680146] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:06:55.699 [2024-11-26 23:40:43.680226] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:06:55.699 [2024-11-26 23:40:43.682787] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:06:55.699 [2024-11-26 23:40:43.682874] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:06:55.699 [2024-11-26 23:40:43.682949] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:06:55.699 [2024-11-26 23:40:43.682999] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:06:55.699 { 00:06:55.699 "results": [ 00:06:55.699 { 00:06:55.699 "job": "raid_bdev1", 00:06:55.699 "core_mask": "0x1", 00:06:55.699 "workload": "randrw", 00:06:55.699 "percentage": 50, 00:06:55.699 "status": "finished", 00:06:55.699 "queue_depth": 1, 00:06:55.699 "io_size": 131072, 00:06:55.699 "runtime": 1.32836, 00:06:55.699 "iops": 17167.78584118763, 00:06:55.699 "mibps": 2145.9732301484537, 00:06:55.699 "io_failed": 1, 00:06:55.699 "io_timeout": 0, 00:06:55.699 "avg_latency_us": 80.17999767930526, 00:06:55.699 "min_latency_us": 24.817467248908297, 00:06:55.699 "max_latency_us": 1366.5257641921398 00:06:55.699 } 00:06:55.699 ], 00:06:55.699 "core_count": 1 00:06:55.699 } 00:06:55.699 23:40:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:55.699 23:40:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 73440 00:06:55.699 23:40:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # '[' -z 73440 ']' 00:06:55.700 23:40:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@958 -- # kill -0 73440 00:06:55.700 23:40:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # uname 00:06:55.700 23:40:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:06:55.700 23:40:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 73440 00:06:55.700 23:40:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:06:55.700 23:40:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:06:55.700 23:40:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 73440' 00:06:55.700 killing process with pid 73440 00:06:55.700 23:40:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@973 -- # kill 73440 00:06:55.700 [2024-11-26 23:40:43.712184] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:06:55.700 23:40:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@978 -- # wait 73440 00:06:55.700 [2024-11-26 23:40:43.726829] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:06:55.970 23:40:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.mQmk5JIVfk 00:06:55.970 23:40:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:06:55.970 23:40:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:06:55.970 23:40:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.75 00:06:55.970 23:40:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy concat 00:06:55.970 23:40:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:06:55.970 23:40:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:06:55.970 23:40:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.75 != \0\.\0\0 ]] 00:06:55.970 00:06:55.970 real 0m3.143s 00:06:55.970 user 0m3.992s 00:06:55.970 sys 0m0.467s 00:06:55.970 23:40:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:55.970 ************************************ 00:06:55.970 END TEST raid_read_error_test 00:06:55.970 ************************************ 00:06:55.970 23:40:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:06:55.970 23:40:43 bdev_raid -- bdev/bdev_raid.sh@972 -- # run_test raid_write_error_test raid_io_error_test concat 2 write 00:06:55.970 23:40:43 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:06:55.970 23:40:43 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:55.970 23:40:43 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:06:55.970 ************************************ 00:06:55.970 START TEST raid_write_error_test 00:06:55.970 ************************************ 00:06:55.970 23:40:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test concat 2 write 00:06:55.970 23:40:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=concat 00:06:55.970 23:40:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=2 00:06:55.970 23:40:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=write 00:06:55.970 23:40:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:06:55.970 23:40:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:06:55.970 23:40:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:06:55.970 23:40:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:06:55.970 23:40:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:06:55.970 23:40:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:06:55.970 23:40:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:06:55.970 23:40:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:06:55.970 23:40:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:06:55.970 23:40:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:06:55.970 23:40:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:06:55.970 23:40:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:06:55.970 23:40:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:06:55.970 23:40:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:06:55.970 23:40:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:06:55.970 23:40:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # '[' concat '!=' raid1 ']' 00:06:55.970 23:40:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:06:55.970 23:40:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:06:55.970 23:40:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:06:55.970 23:40:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.0GMjatVRJ0 00:06:55.970 23:40:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=73569 00:06:55.970 23:40:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:06:55.970 23:40:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 73569 00:06:55.970 23:40:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@835 -- # '[' -z 73569 ']' 00:06:55.970 23:40:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:55.970 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:55.970 23:40:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:55.970 23:40:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:55.970 23:40:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:55.970 23:40:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:06:55.970 [2024-11-26 23:40:44.098568] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:06:56.229 [2024-11-26 23:40:44.098768] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid73569 ] 00:06:56.229 [2024-11-26 23:40:44.251121] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:56.229 [2024-11-26 23:40:44.275290] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:56.229 [2024-11-26 23:40:44.316522] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:06:56.229 [2024-11-26 23:40:44.316555] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:06:56.808 23:40:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:56.808 23:40:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@868 -- # return 0 00:06:56.808 23:40:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:06:56.808 23:40:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:06:56.808 23:40:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:56.808 23:40:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:06:56.808 BaseBdev1_malloc 00:06:56.808 23:40:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:56.808 23:40:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:06:56.808 23:40:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:56.808 23:40:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:06:57.068 true 00:06:57.068 23:40:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:57.068 23:40:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:06:57.068 23:40:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:57.068 23:40:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:06:57.068 [2024-11-26 23:40:44.943241] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:06:57.068 [2024-11-26 23:40:44.943409] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:06:57.068 [2024-11-26 23:40:44.943434] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006980 00:06:57.068 [2024-11-26 23:40:44.943444] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:06:57.068 [2024-11-26 23:40:44.945519] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:06:57.068 [2024-11-26 23:40:44.945551] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:06:57.068 BaseBdev1 00:06:57.068 23:40:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:57.068 23:40:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:06:57.068 23:40:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:06:57.069 23:40:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:57.069 23:40:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:06:57.069 BaseBdev2_malloc 00:06:57.069 23:40:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:57.069 23:40:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:06:57.069 23:40:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:57.069 23:40:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:06:57.069 true 00:06:57.069 23:40:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:57.069 23:40:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:06:57.069 23:40:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:57.069 23:40:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:06:57.069 [2024-11-26 23:40:44.979896] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:06:57.069 [2024-11-26 23:40:44.979942] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:06:57.069 [2024-11-26 23:40:44.979958] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:06:57.069 [2024-11-26 23:40:44.979974] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:06:57.069 [2024-11-26 23:40:44.981964] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:06:57.069 [2024-11-26 23:40:44.981999] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:06:57.069 BaseBdev2 00:06:57.069 23:40:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:57.069 23:40:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 -s 00:06:57.069 23:40:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:57.069 23:40:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:06:57.069 [2024-11-26 23:40:44.991928] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:06:57.069 [2024-11-26 23:40:44.993693] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:06:57.069 [2024-11-26 23:40:44.993855] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:06:57.069 [2024-11-26 23:40:44.993867] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:06:57.069 [2024-11-26 23:40:44.994142] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002870 00:06:57.069 [2024-11-26 23:40:44.994290] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:06:57.069 [2024-11-26 23:40:44.994307] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:06:57.069 [2024-11-26 23:40:44.994426] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:06:57.069 23:40:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:57.069 23:40:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online concat 64 2 00:06:57.069 23:40:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:06:57.069 23:40:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:06:57.069 23:40:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:06:57.069 23:40:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:06:57.069 23:40:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:06:57.069 23:40:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:06:57.069 23:40:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:06:57.069 23:40:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:06:57.069 23:40:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:06:57.069 23:40:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:06:57.069 23:40:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:06:57.069 23:40:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:57.069 23:40:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:06:57.069 23:40:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:57.069 23:40:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:06:57.069 "name": "raid_bdev1", 00:06:57.069 "uuid": "cf991239-b215-4a5e-88cc-f744a6259abf", 00:06:57.069 "strip_size_kb": 64, 00:06:57.069 "state": "online", 00:06:57.069 "raid_level": "concat", 00:06:57.069 "superblock": true, 00:06:57.069 "num_base_bdevs": 2, 00:06:57.069 "num_base_bdevs_discovered": 2, 00:06:57.069 "num_base_bdevs_operational": 2, 00:06:57.069 "base_bdevs_list": [ 00:06:57.069 { 00:06:57.069 "name": "BaseBdev1", 00:06:57.069 "uuid": "ac79541b-85f7-5ef9-b83e-bfec33012eca", 00:06:57.069 "is_configured": true, 00:06:57.069 "data_offset": 2048, 00:06:57.069 "data_size": 63488 00:06:57.069 }, 00:06:57.069 { 00:06:57.069 "name": "BaseBdev2", 00:06:57.069 "uuid": "6be7cb8e-e305-5c0c-aff3-ce3f087e44f0", 00:06:57.069 "is_configured": true, 00:06:57.069 "data_offset": 2048, 00:06:57.069 "data_size": 63488 00:06:57.069 } 00:06:57.069 ] 00:06:57.069 }' 00:06:57.069 23:40:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:06:57.069 23:40:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:06:57.329 23:40:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:06:57.329 23:40:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:06:57.587 [2024-11-26 23:40:45.535427] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002a10 00:06:58.527 23:40:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:06:58.527 23:40:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:58.527 23:40:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:06:58.527 23:40:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:58.527 23:40:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:06:58.527 23:40:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ concat = \r\a\i\d\1 ]] 00:06:58.527 23:40:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=2 00:06:58.527 23:40:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online concat 64 2 00:06:58.527 23:40:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:06:58.527 23:40:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:06:58.527 23:40:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:06:58.527 23:40:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:06:58.527 23:40:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:06:58.527 23:40:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:06:58.527 23:40:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:06:58.527 23:40:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:06:58.527 23:40:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:06:58.527 23:40:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:06:58.527 23:40:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:58.527 23:40:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:06:58.527 23:40:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:06:58.527 23:40:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:58.527 23:40:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:06:58.527 "name": "raid_bdev1", 00:06:58.527 "uuid": "cf991239-b215-4a5e-88cc-f744a6259abf", 00:06:58.527 "strip_size_kb": 64, 00:06:58.527 "state": "online", 00:06:58.527 "raid_level": "concat", 00:06:58.527 "superblock": true, 00:06:58.527 "num_base_bdevs": 2, 00:06:58.527 "num_base_bdevs_discovered": 2, 00:06:58.527 "num_base_bdevs_operational": 2, 00:06:58.527 "base_bdevs_list": [ 00:06:58.527 { 00:06:58.527 "name": "BaseBdev1", 00:06:58.527 "uuid": "ac79541b-85f7-5ef9-b83e-bfec33012eca", 00:06:58.527 "is_configured": true, 00:06:58.527 "data_offset": 2048, 00:06:58.527 "data_size": 63488 00:06:58.527 }, 00:06:58.527 { 00:06:58.527 "name": "BaseBdev2", 00:06:58.527 "uuid": "6be7cb8e-e305-5c0c-aff3-ce3f087e44f0", 00:06:58.527 "is_configured": true, 00:06:58.527 "data_offset": 2048, 00:06:58.527 "data_size": 63488 00:06:58.527 } 00:06:58.527 ] 00:06:58.527 }' 00:06:58.527 23:40:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:06:58.527 23:40:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:06:58.787 23:40:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:06:58.787 23:40:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:58.787 23:40:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:06:58.787 [2024-11-26 23:40:46.914801] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:06:58.787 [2024-11-26 23:40:46.914897] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:06:59.047 [2024-11-26 23:40:46.917491] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:06:59.047 [2024-11-26 23:40:46.917569] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:06:59.047 [2024-11-26 23:40:46.917620] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:06:59.047 [2024-11-26 23:40:46.917660] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:06:59.047 { 00:06:59.047 "results": [ 00:06:59.047 { 00:06:59.047 "job": "raid_bdev1", 00:06:59.047 "core_mask": "0x1", 00:06:59.047 "workload": "randrw", 00:06:59.047 "percentage": 50, 00:06:59.047 "status": "finished", 00:06:59.047 "queue_depth": 1, 00:06:59.047 "io_size": 131072, 00:06:59.047 "runtime": 1.380527, 00:06:59.047 "iops": 17194.158462674037, 00:06:59.047 "mibps": 2149.2698078342546, 00:06:59.047 "io_failed": 1, 00:06:59.047 "io_timeout": 0, 00:06:59.047 "avg_latency_us": 80.02356673157956, 00:06:59.047 "min_latency_us": 25.041048034934498, 00:06:59.047 "max_latency_us": 1373.6803493449781 00:06:59.047 } 00:06:59.047 ], 00:06:59.047 "core_count": 1 00:06:59.047 } 00:06:59.047 23:40:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:59.047 23:40:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 73569 00:06:59.047 23:40:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # '[' -z 73569 ']' 00:06:59.047 23:40:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@958 -- # kill -0 73569 00:06:59.047 23:40:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # uname 00:06:59.047 23:40:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:06:59.047 23:40:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 73569 00:06:59.047 23:40:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:06:59.047 killing process with pid 73569 00:06:59.047 23:40:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:06:59.047 23:40:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 73569' 00:06:59.047 23:40:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@973 -- # kill 73569 00:06:59.047 [2024-11-26 23:40:46.951787] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:06:59.047 23:40:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@978 -- # wait 73569 00:06:59.047 [2024-11-26 23:40:46.966474] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:06:59.047 23:40:47 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:06:59.047 23:40:47 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:06:59.047 23:40:47 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.0GMjatVRJ0 00:06:59.047 23:40:47 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.72 00:06:59.047 23:40:47 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy concat 00:06:59.047 23:40:47 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:06:59.047 23:40:47 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:06:59.047 23:40:47 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.72 != \0\.\0\0 ]] 00:06:59.047 00:06:59.047 real 0m3.173s 00:06:59.047 user 0m4.081s 00:06:59.047 sys 0m0.467s 00:06:59.322 ************************************ 00:06:59.322 END TEST raid_write_error_test 00:06:59.322 ************************************ 00:06:59.322 23:40:47 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:59.322 23:40:47 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:06:59.322 23:40:47 bdev_raid -- bdev/bdev_raid.sh@967 -- # for level in raid0 concat raid1 00:06:59.322 23:40:47 bdev_raid -- bdev/bdev_raid.sh@968 -- # run_test raid_state_function_test raid_state_function_test raid1 2 false 00:06:59.322 23:40:47 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:06:59.322 23:40:47 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:59.322 23:40:47 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:06:59.322 ************************************ 00:06:59.322 START TEST raid_state_function_test 00:06:59.322 ************************************ 00:06:59.322 23:40:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1129 -- # raid_state_function_test raid1 2 false 00:06:59.322 23:40:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=raid1 00:06:59.322 23:40:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=2 00:06:59.322 23:40:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:06:59.322 23:40:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:06:59.322 23:40:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:06:59.322 23:40:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:06:59.322 23:40:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:06:59.322 23:40:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:06:59.322 23:40:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:06:59.322 23:40:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:06:59.322 23:40:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:06:59.322 23:40:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:06:59.322 23:40:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:06:59.322 23:40:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:06:59.322 23:40:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:06:59.322 23:40:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:06:59.322 23:40:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:06:59.322 23:40:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:06:59.322 23:40:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' raid1 '!=' raid1 ']' 00:06:59.322 23:40:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@219 -- # strip_size=0 00:06:59.322 23:40:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:06:59.322 23:40:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:06:59.322 Process raid pid: 73700 00:06:59.322 23:40:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=73700 00:06:59.322 23:40:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:06:59.322 23:40:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 73700' 00:06:59.322 23:40:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 73700 00:06:59.322 23:40:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@835 -- # '[' -z 73700 ']' 00:06:59.322 23:40:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:59.322 23:40:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:59.323 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:59.323 23:40:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:59.323 23:40:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:59.323 23:40:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:06:59.323 [2024-11-26 23:40:47.330817] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:06:59.323 [2024-11-26 23:40:47.330952] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:06:59.585 [2024-11-26 23:40:47.469607] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:59.586 [2024-11-26 23:40:47.493819] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:59.586 [2024-11-26 23:40:47.534960] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:06:59.586 [2024-11-26 23:40:47.534996] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:00.154 23:40:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:00.154 23:40:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@868 -- # return 0 00:07:00.154 23:40:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:00.154 23:40:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:00.154 23:40:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:00.154 [2024-11-26 23:40:48.172983] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:07:00.154 [2024-11-26 23:40:48.173046] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:07:00.154 [2024-11-26 23:40:48.173065] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:00.154 [2024-11-26 23:40:48.173076] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:00.154 23:40:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:00.154 23:40:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:07:00.154 23:40:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:00.154 23:40:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:00.154 23:40:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:07:00.154 23:40:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:07:00.154 23:40:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:00.154 23:40:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:00.154 23:40:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:00.154 23:40:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:00.154 23:40:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:00.154 23:40:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:00.154 23:40:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:00.154 23:40:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:00.154 23:40:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:00.154 23:40:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:00.154 23:40:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:00.154 "name": "Existed_Raid", 00:07:00.154 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:00.154 "strip_size_kb": 0, 00:07:00.154 "state": "configuring", 00:07:00.154 "raid_level": "raid1", 00:07:00.154 "superblock": false, 00:07:00.154 "num_base_bdevs": 2, 00:07:00.154 "num_base_bdevs_discovered": 0, 00:07:00.154 "num_base_bdevs_operational": 2, 00:07:00.154 "base_bdevs_list": [ 00:07:00.154 { 00:07:00.154 "name": "BaseBdev1", 00:07:00.154 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:00.154 "is_configured": false, 00:07:00.154 "data_offset": 0, 00:07:00.154 "data_size": 0 00:07:00.154 }, 00:07:00.154 { 00:07:00.154 "name": "BaseBdev2", 00:07:00.154 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:00.154 "is_configured": false, 00:07:00.154 "data_offset": 0, 00:07:00.154 "data_size": 0 00:07:00.154 } 00:07:00.154 ] 00:07:00.154 }' 00:07:00.154 23:40:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:00.154 23:40:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:00.745 23:40:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:07:00.745 23:40:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:00.745 23:40:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:00.745 [2024-11-26 23:40:48.580278] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:07:00.746 [2024-11-26 23:40:48.580448] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:07:00.746 23:40:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:00.746 23:40:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:00.746 23:40:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:00.746 23:40:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:00.746 [2024-11-26 23:40:48.592259] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:07:00.746 [2024-11-26 23:40:48.592424] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:07:00.746 [2024-11-26 23:40:48.592457] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:00.746 [2024-11-26 23:40:48.592498] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:00.746 23:40:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:00.746 23:40:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:07:00.746 23:40:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:00.746 23:40:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:00.746 [2024-11-26 23:40:48.619537] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:00.746 BaseBdev1 00:07:00.746 23:40:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:00.746 23:40:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:07:00.746 23:40:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:07:00.746 23:40:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:07:00.747 23:40:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:07:00.747 23:40:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:07:00.747 23:40:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:07:00.747 23:40:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:07:00.747 23:40:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:00.747 23:40:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:00.747 23:40:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:00.747 23:40:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:07:00.747 23:40:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:00.747 23:40:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:00.747 [ 00:07:00.747 { 00:07:00.747 "name": "BaseBdev1", 00:07:00.747 "aliases": [ 00:07:00.747 "ea6a3176-7157-42a8-8f8d-8e26a8c63188" 00:07:00.747 ], 00:07:00.747 "product_name": "Malloc disk", 00:07:00.747 "block_size": 512, 00:07:00.747 "num_blocks": 65536, 00:07:00.747 "uuid": "ea6a3176-7157-42a8-8f8d-8e26a8c63188", 00:07:00.747 "assigned_rate_limits": { 00:07:00.747 "rw_ios_per_sec": 0, 00:07:00.747 "rw_mbytes_per_sec": 0, 00:07:00.747 "r_mbytes_per_sec": 0, 00:07:00.747 "w_mbytes_per_sec": 0 00:07:00.747 }, 00:07:00.747 "claimed": true, 00:07:00.747 "claim_type": "exclusive_write", 00:07:00.747 "zoned": false, 00:07:00.747 "supported_io_types": { 00:07:00.747 "read": true, 00:07:00.747 "write": true, 00:07:00.747 "unmap": true, 00:07:00.747 "flush": true, 00:07:00.747 "reset": true, 00:07:00.747 "nvme_admin": false, 00:07:00.747 "nvme_io": false, 00:07:00.747 "nvme_io_md": false, 00:07:00.747 "write_zeroes": true, 00:07:00.747 "zcopy": true, 00:07:00.747 "get_zone_info": false, 00:07:00.747 "zone_management": false, 00:07:00.747 "zone_append": false, 00:07:00.747 "compare": false, 00:07:00.747 "compare_and_write": false, 00:07:00.747 "abort": true, 00:07:00.747 "seek_hole": false, 00:07:00.747 "seek_data": false, 00:07:00.747 "copy": true, 00:07:00.747 "nvme_iov_md": false 00:07:00.747 }, 00:07:00.747 "memory_domains": [ 00:07:00.747 { 00:07:00.747 "dma_device_id": "system", 00:07:00.747 "dma_device_type": 1 00:07:00.747 }, 00:07:00.747 { 00:07:00.747 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:00.747 "dma_device_type": 2 00:07:00.747 } 00:07:00.747 ], 00:07:00.747 "driver_specific": {} 00:07:00.747 } 00:07:00.747 ] 00:07:00.747 23:40:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:00.747 23:40:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:07:00.747 23:40:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:07:00.747 23:40:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:00.747 23:40:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:00.747 23:40:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:07:00.747 23:40:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:07:00.747 23:40:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:00.747 23:40:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:00.747 23:40:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:00.747 23:40:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:00.747 23:40:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:00.747 23:40:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:00.747 23:40:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:00.747 23:40:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:00.747 23:40:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:00.747 23:40:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:00.747 23:40:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:00.747 "name": "Existed_Raid", 00:07:00.747 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:00.747 "strip_size_kb": 0, 00:07:00.747 "state": "configuring", 00:07:00.747 "raid_level": "raid1", 00:07:00.747 "superblock": false, 00:07:00.747 "num_base_bdevs": 2, 00:07:00.747 "num_base_bdevs_discovered": 1, 00:07:00.747 "num_base_bdevs_operational": 2, 00:07:00.747 "base_bdevs_list": [ 00:07:00.747 { 00:07:00.747 "name": "BaseBdev1", 00:07:00.747 "uuid": "ea6a3176-7157-42a8-8f8d-8e26a8c63188", 00:07:00.747 "is_configured": true, 00:07:00.747 "data_offset": 0, 00:07:00.747 "data_size": 65536 00:07:00.747 }, 00:07:00.747 { 00:07:00.747 "name": "BaseBdev2", 00:07:00.747 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:00.747 "is_configured": false, 00:07:00.747 "data_offset": 0, 00:07:00.747 "data_size": 0 00:07:00.747 } 00:07:00.747 ] 00:07:00.747 }' 00:07:00.747 23:40:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:00.747 23:40:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:01.006 23:40:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:07:01.006 23:40:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:01.006 23:40:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:01.006 [2024-11-26 23:40:49.114774] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:07:01.006 [2024-11-26 23:40:49.114855] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:07:01.006 23:40:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:01.006 23:40:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:01.006 23:40:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:01.006 23:40:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:01.006 [2024-11-26 23:40:49.126757] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:01.006 [2024-11-26 23:40:49.129018] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:01.006 [2024-11-26 23:40:49.129141] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:01.006 23:40:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:01.006 23:40:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:07:01.007 23:40:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:07:01.007 23:40:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:07:01.007 23:40:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:01.007 23:40:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:01.007 23:40:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:07:01.007 23:40:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:07:01.007 23:40:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:01.007 23:40:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:01.007 23:40:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:01.007 23:40:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:01.007 23:40:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:01.289 23:40:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:01.289 23:40:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:01.289 23:40:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:01.289 23:40:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:01.289 23:40:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:01.289 23:40:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:01.289 "name": "Existed_Raid", 00:07:01.289 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:01.289 "strip_size_kb": 0, 00:07:01.289 "state": "configuring", 00:07:01.289 "raid_level": "raid1", 00:07:01.289 "superblock": false, 00:07:01.289 "num_base_bdevs": 2, 00:07:01.289 "num_base_bdevs_discovered": 1, 00:07:01.289 "num_base_bdevs_operational": 2, 00:07:01.289 "base_bdevs_list": [ 00:07:01.289 { 00:07:01.289 "name": "BaseBdev1", 00:07:01.289 "uuid": "ea6a3176-7157-42a8-8f8d-8e26a8c63188", 00:07:01.289 "is_configured": true, 00:07:01.289 "data_offset": 0, 00:07:01.289 "data_size": 65536 00:07:01.289 }, 00:07:01.289 { 00:07:01.289 "name": "BaseBdev2", 00:07:01.289 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:01.289 "is_configured": false, 00:07:01.289 "data_offset": 0, 00:07:01.289 "data_size": 0 00:07:01.289 } 00:07:01.289 ] 00:07:01.289 }' 00:07:01.289 23:40:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:01.289 23:40:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:01.558 23:40:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:07:01.558 23:40:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:01.558 23:40:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:01.558 [2024-11-26 23:40:49.583080] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:07:01.558 [2024-11-26 23:40:49.583237] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:07:01.558 [2024-11-26 23:40:49.583267] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:07:01.558 [2024-11-26 23:40:49.583658] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:07:01.558 [2024-11-26 23:40:49.583881] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:07:01.558 [2024-11-26 23:40:49.583931] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:07:01.558 [2024-11-26 23:40:49.584239] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:01.558 BaseBdev2 00:07:01.558 23:40:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:01.558 23:40:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:07:01.558 23:40:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:07:01.558 23:40:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:07:01.558 23:40:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:07:01.558 23:40:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:07:01.558 23:40:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:07:01.559 23:40:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:07:01.559 23:40:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:01.559 23:40:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:01.559 23:40:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:01.559 23:40:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:07:01.559 23:40:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:01.559 23:40:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:01.559 [ 00:07:01.559 { 00:07:01.559 "name": "BaseBdev2", 00:07:01.559 "aliases": [ 00:07:01.559 "54a462ba-6d31-4770-887f-57c2647e7ae2" 00:07:01.559 ], 00:07:01.559 "product_name": "Malloc disk", 00:07:01.559 "block_size": 512, 00:07:01.559 "num_blocks": 65536, 00:07:01.559 "uuid": "54a462ba-6d31-4770-887f-57c2647e7ae2", 00:07:01.559 "assigned_rate_limits": { 00:07:01.559 "rw_ios_per_sec": 0, 00:07:01.559 "rw_mbytes_per_sec": 0, 00:07:01.559 "r_mbytes_per_sec": 0, 00:07:01.559 "w_mbytes_per_sec": 0 00:07:01.559 }, 00:07:01.559 "claimed": true, 00:07:01.559 "claim_type": "exclusive_write", 00:07:01.559 "zoned": false, 00:07:01.559 "supported_io_types": { 00:07:01.559 "read": true, 00:07:01.559 "write": true, 00:07:01.559 "unmap": true, 00:07:01.559 "flush": true, 00:07:01.559 "reset": true, 00:07:01.559 "nvme_admin": false, 00:07:01.559 "nvme_io": false, 00:07:01.559 "nvme_io_md": false, 00:07:01.559 "write_zeroes": true, 00:07:01.559 "zcopy": true, 00:07:01.559 "get_zone_info": false, 00:07:01.559 "zone_management": false, 00:07:01.559 "zone_append": false, 00:07:01.559 "compare": false, 00:07:01.559 "compare_and_write": false, 00:07:01.559 "abort": true, 00:07:01.559 "seek_hole": false, 00:07:01.559 "seek_data": false, 00:07:01.559 "copy": true, 00:07:01.559 "nvme_iov_md": false 00:07:01.559 }, 00:07:01.559 "memory_domains": [ 00:07:01.559 { 00:07:01.559 "dma_device_id": "system", 00:07:01.559 "dma_device_type": 1 00:07:01.559 }, 00:07:01.559 { 00:07:01.559 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:01.559 "dma_device_type": 2 00:07:01.559 } 00:07:01.559 ], 00:07:01.559 "driver_specific": {} 00:07:01.559 } 00:07:01.559 ] 00:07:01.559 23:40:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:01.559 23:40:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:07:01.559 23:40:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:07:01.559 23:40:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:07:01.559 23:40:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid1 0 2 00:07:01.559 23:40:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:01.559 23:40:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:01.559 23:40:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:07:01.559 23:40:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:07:01.559 23:40:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:01.559 23:40:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:01.559 23:40:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:01.559 23:40:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:01.559 23:40:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:01.559 23:40:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:01.559 23:40:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:01.559 23:40:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:01.559 23:40:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:01.559 23:40:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:01.559 23:40:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:01.559 "name": "Existed_Raid", 00:07:01.559 "uuid": "619b6e11-3aa6-423d-abd3-b8da00ffe866", 00:07:01.559 "strip_size_kb": 0, 00:07:01.559 "state": "online", 00:07:01.559 "raid_level": "raid1", 00:07:01.559 "superblock": false, 00:07:01.559 "num_base_bdevs": 2, 00:07:01.559 "num_base_bdevs_discovered": 2, 00:07:01.559 "num_base_bdevs_operational": 2, 00:07:01.559 "base_bdevs_list": [ 00:07:01.559 { 00:07:01.559 "name": "BaseBdev1", 00:07:01.559 "uuid": "ea6a3176-7157-42a8-8f8d-8e26a8c63188", 00:07:01.559 "is_configured": true, 00:07:01.559 "data_offset": 0, 00:07:01.559 "data_size": 65536 00:07:01.559 }, 00:07:01.559 { 00:07:01.559 "name": "BaseBdev2", 00:07:01.559 "uuid": "54a462ba-6d31-4770-887f-57c2647e7ae2", 00:07:01.559 "is_configured": true, 00:07:01.559 "data_offset": 0, 00:07:01.559 "data_size": 65536 00:07:01.559 } 00:07:01.559 ] 00:07:01.559 }' 00:07:01.559 23:40:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:01.559 23:40:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:02.129 23:40:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:07:02.129 23:40:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:07:02.129 23:40:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:07:02.129 23:40:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:07:02.129 23:40:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:07:02.129 23:40:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:07:02.129 23:40:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:07:02.129 23:40:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:07:02.129 23:40:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:02.129 23:40:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:02.129 [2024-11-26 23:40:50.038730] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:02.129 23:40:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:02.129 23:40:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:07:02.129 "name": "Existed_Raid", 00:07:02.129 "aliases": [ 00:07:02.129 "619b6e11-3aa6-423d-abd3-b8da00ffe866" 00:07:02.129 ], 00:07:02.129 "product_name": "Raid Volume", 00:07:02.129 "block_size": 512, 00:07:02.129 "num_blocks": 65536, 00:07:02.129 "uuid": "619b6e11-3aa6-423d-abd3-b8da00ffe866", 00:07:02.129 "assigned_rate_limits": { 00:07:02.129 "rw_ios_per_sec": 0, 00:07:02.129 "rw_mbytes_per_sec": 0, 00:07:02.129 "r_mbytes_per_sec": 0, 00:07:02.129 "w_mbytes_per_sec": 0 00:07:02.129 }, 00:07:02.129 "claimed": false, 00:07:02.129 "zoned": false, 00:07:02.129 "supported_io_types": { 00:07:02.129 "read": true, 00:07:02.129 "write": true, 00:07:02.129 "unmap": false, 00:07:02.129 "flush": false, 00:07:02.129 "reset": true, 00:07:02.129 "nvme_admin": false, 00:07:02.129 "nvme_io": false, 00:07:02.129 "nvme_io_md": false, 00:07:02.129 "write_zeroes": true, 00:07:02.129 "zcopy": false, 00:07:02.129 "get_zone_info": false, 00:07:02.129 "zone_management": false, 00:07:02.129 "zone_append": false, 00:07:02.129 "compare": false, 00:07:02.129 "compare_and_write": false, 00:07:02.129 "abort": false, 00:07:02.129 "seek_hole": false, 00:07:02.129 "seek_data": false, 00:07:02.129 "copy": false, 00:07:02.129 "nvme_iov_md": false 00:07:02.129 }, 00:07:02.129 "memory_domains": [ 00:07:02.129 { 00:07:02.129 "dma_device_id": "system", 00:07:02.129 "dma_device_type": 1 00:07:02.129 }, 00:07:02.129 { 00:07:02.129 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:02.129 "dma_device_type": 2 00:07:02.129 }, 00:07:02.129 { 00:07:02.129 "dma_device_id": "system", 00:07:02.129 "dma_device_type": 1 00:07:02.129 }, 00:07:02.129 { 00:07:02.129 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:02.129 "dma_device_type": 2 00:07:02.129 } 00:07:02.129 ], 00:07:02.129 "driver_specific": { 00:07:02.129 "raid": { 00:07:02.129 "uuid": "619b6e11-3aa6-423d-abd3-b8da00ffe866", 00:07:02.129 "strip_size_kb": 0, 00:07:02.129 "state": "online", 00:07:02.129 "raid_level": "raid1", 00:07:02.129 "superblock": false, 00:07:02.129 "num_base_bdevs": 2, 00:07:02.129 "num_base_bdevs_discovered": 2, 00:07:02.129 "num_base_bdevs_operational": 2, 00:07:02.129 "base_bdevs_list": [ 00:07:02.129 { 00:07:02.129 "name": "BaseBdev1", 00:07:02.129 "uuid": "ea6a3176-7157-42a8-8f8d-8e26a8c63188", 00:07:02.129 "is_configured": true, 00:07:02.129 "data_offset": 0, 00:07:02.129 "data_size": 65536 00:07:02.129 }, 00:07:02.129 { 00:07:02.129 "name": "BaseBdev2", 00:07:02.129 "uuid": "54a462ba-6d31-4770-887f-57c2647e7ae2", 00:07:02.129 "is_configured": true, 00:07:02.129 "data_offset": 0, 00:07:02.129 "data_size": 65536 00:07:02.129 } 00:07:02.129 ] 00:07:02.129 } 00:07:02.129 } 00:07:02.129 }' 00:07:02.129 23:40:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:07:02.129 23:40:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:07:02.129 BaseBdev2' 00:07:02.129 23:40:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:02.129 23:40:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:07:02.129 23:40:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:02.129 23:40:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:07:02.129 23:40:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:02.129 23:40:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:02.129 23:40:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:02.129 23:40:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:02.129 23:40:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:02.129 23:40:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:02.129 23:40:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:02.129 23:40:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:02.129 23:40:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:07:02.129 23:40:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:02.129 23:40:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:02.129 23:40:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:02.129 23:40:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:02.129 23:40:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:02.129 23:40:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:07:02.129 23:40:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:02.129 23:40:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:02.129 [2024-11-26 23:40:50.194337] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:07:02.129 23:40:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:02.129 23:40:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:07:02.129 23:40:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy raid1 00:07:02.129 23:40:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:07:02.130 23:40:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@199 -- # return 0 00:07:02.130 23:40:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:07:02.130 23:40:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid1 0 1 00:07:02.130 23:40:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:02.130 23:40:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:02.130 23:40:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:07:02.130 23:40:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:07:02.130 23:40:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:07:02.130 23:40:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:02.130 23:40:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:02.130 23:40:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:02.130 23:40:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:02.130 23:40:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:02.130 23:40:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:02.130 23:40:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:02.130 23:40:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:02.130 23:40:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:02.389 23:40:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:02.389 "name": "Existed_Raid", 00:07:02.389 "uuid": "619b6e11-3aa6-423d-abd3-b8da00ffe866", 00:07:02.389 "strip_size_kb": 0, 00:07:02.389 "state": "online", 00:07:02.389 "raid_level": "raid1", 00:07:02.389 "superblock": false, 00:07:02.389 "num_base_bdevs": 2, 00:07:02.389 "num_base_bdevs_discovered": 1, 00:07:02.389 "num_base_bdevs_operational": 1, 00:07:02.389 "base_bdevs_list": [ 00:07:02.389 { 00:07:02.389 "name": null, 00:07:02.389 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:02.389 "is_configured": false, 00:07:02.389 "data_offset": 0, 00:07:02.389 "data_size": 65536 00:07:02.389 }, 00:07:02.389 { 00:07:02.389 "name": "BaseBdev2", 00:07:02.389 "uuid": "54a462ba-6d31-4770-887f-57c2647e7ae2", 00:07:02.389 "is_configured": true, 00:07:02.389 "data_offset": 0, 00:07:02.389 "data_size": 65536 00:07:02.389 } 00:07:02.389 ] 00:07:02.389 }' 00:07:02.389 23:40:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:02.389 23:40:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:02.649 23:40:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:07:02.649 23:40:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:07:02.649 23:40:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:02.649 23:40:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:02.649 23:40:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:02.649 23:40:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:07:02.649 23:40:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:02.649 23:40:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:07:02.649 23:40:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:07:02.649 23:40:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:07:02.649 23:40:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:02.649 23:40:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:02.649 [2024-11-26 23:40:50.682063] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:07:02.649 [2024-11-26 23:40:50.682301] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:02.649 [2024-11-26 23:40:50.703477] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:02.649 [2024-11-26 23:40:50.703543] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:02.649 [2024-11-26 23:40:50.703556] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:07:02.649 23:40:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:02.649 23:40:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:07:02.649 23:40:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:07:02.649 23:40:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:02.649 23:40:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:02.649 23:40:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:02.649 23:40:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:07:02.649 23:40:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:02.649 23:40:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:07:02.649 23:40:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:07:02.649 23:40:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 2 -gt 2 ']' 00:07:02.649 23:40:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 73700 00:07:02.649 23:40:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # '[' -z 73700 ']' 00:07:02.649 23:40:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@958 -- # kill -0 73700 00:07:02.649 23:40:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # uname 00:07:02.649 23:40:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:07:02.649 23:40:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 73700 00:07:02.908 killing process with pid 73700 00:07:02.908 23:40:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:07:02.908 23:40:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:07:02.908 23:40:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 73700' 00:07:02.908 23:40:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@973 -- # kill 73700 00:07:02.908 [2024-11-26 23:40:50.786621] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:02.908 23:40:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@978 -- # wait 73700 00:07:02.908 [2024-11-26 23:40:50.788270] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:03.168 23:40:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:07:03.168 00:07:03.168 real 0m3.868s 00:07:03.168 user 0m6.015s 00:07:03.168 sys 0m0.708s 00:07:03.168 23:40:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:07:03.168 23:40:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:03.168 ************************************ 00:07:03.168 END TEST raid_state_function_test 00:07:03.168 ************************************ 00:07:03.168 23:40:51 bdev_raid -- bdev/bdev_raid.sh@969 -- # run_test raid_state_function_test_sb raid_state_function_test raid1 2 true 00:07:03.168 23:40:51 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:07:03.168 23:40:51 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:03.168 23:40:51 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:03.168 ************************************ 00:07:03.168 START TEST raid_state_function_test_sb 00:07:03.168 ************************************ 00:07:03.168 23:40:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1129 -- # raid_state_function_test raid1 2 true 00:07:03.168 23:40:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=raid1 00:07:03.168 23:40:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=2 00:07:03.168 23:40:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:07:03.168 23:40:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:07:03.168 23:40:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:07:03.168 23:40:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:03.168 23:40:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:07:03.168 23:40:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:07:03.168 23:40:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:03.168 23:40:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:07:03.168 23:40:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:07:03.168 23:40:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:03.168 23:40:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:07:03.168 23:40:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:07:03.168 23:40:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:07:03.168 23:40:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:07:03.168 23:40:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:07:03.168 23:40:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:07:03.168 23:40:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' raid1 '!=' raid1 ']' 00:07:03.168 23:40:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@219 -- # strip_size=0 00:07:03.168 23:40:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:07:03.168 23:40:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:07:03.168 23:40:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=73938 00:07:03.168 23:40:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:07:03.168 23:40:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 73938' 00:07:03.168 Process raid pid: 73938 00:07:03.168 23:40:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 73938 00:07:03.168 23:40:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@835 -- # '[' -z 73938 ']' 00:07:03.168 23:40:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:03.168 23:40:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:03.168 23:40:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:03.168 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:03.168 23:40:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:03.168 23:40:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:03.168 [2024-11-26 23:40:51.267539] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:07:03.168 [2024-11-26 23:40:51.267772] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:07:03.433 [2024-11-26 23:40:51.423522] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:03.433 [2024-11-26 23:40:51.467032] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:03.433 [2024-11-26 23:40:51.543472] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:03.433 [2024-11-26 23:40:51.543540] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:04.000 23:40:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:04.000 23:40:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@868 -- # return 0 00:07:04.000 23:40:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:04.001 23:40:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:04.001 23:40:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:04.001 [2024-11-26 23:40:52.106517] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:07:04.001 [2024-11-26 23:40:52.106578] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:07:04.001 [2024-11-26 23:40:52.106588] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:04.001 [2024-11-26 23:40:52.106597] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:04.001 23:40:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:04.001 23:40:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:07:04.001 23:40:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:04.001 23:40:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:04.001 23:40:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:07:04.001 23:40:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:07:04.001 23:40:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:04.001 23:40:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:04.001 23:40:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:04.001 23:40:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:04.001 23:40:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:04.001 23:40:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:04.001 23:40:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:04.001 23:40:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:04.001 23:40:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:04.260 23:40:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:04.260 23:40:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:04.260 "name": "Existed_Raid", 00:07:04.260 "uuid": "58a4086e-61c6-427d-b019-55fb9eb42990", 00:07:04.260 "strip_size_kb": 0, 00:07:04.260 "state": "configuring", 00:07:04.260 "raid_level": "raid1", 00:07:04.260 "superblock": true, 00:07:04.260 "num_base_bdevs": 2, 00:07:04.260 "num_base_bdevs_discovered": 0, 00:07:04.260 "num_base_bdevs_operational": 2, 00:07:04.260 "base_bdevs_list": [ 00:07:04.260 { 00:07:04.260 "name": "BaseBdev1", 00:07:04.260 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:04.260 "is_configured": false, 00:07:04.260 "data_offset": 0, 00:07:04.260 "data_size": 0 00:07:04.260 }, 00:07:04.260 { 00:07:04.260 "name": "BaseBdev2", 00:07:04.260 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:04.260 "is_configured": false, 00:07:04.260 "data_offset": 0, 00:07:04.260 "data_size": 0 00:07:04.260 } 00:07:04.260 ] 00:07:04.260 }' 00:07:04.260 23:40:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:04.260 23:40:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:04.521 23:40:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:07:04.521 23:40:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:04.521 23:40:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:04.521 [2024-11-26 23:40:52.489814] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:07:04.521 [2024-11-26 23:40:52.489935] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:07:04.521 23:40:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:04.521 23:40:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:04.521 23:40:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:04.521 23:40:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:04.521 [2024-11-26 23:40:52.497799] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:07:04.521 [2024-11-26 23:40:52.497896] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:07:04.521 [2024-11-26 23:40:52.497924] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:04.521 [2024-11-26 23:40:52.497962] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:04.521 23:40:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:04.521 23:40:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:07:04.521 23:40:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:04.521 23:40:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:04.521 [2024-11-26 23:40:52.515339] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:04.521 BaseBdev1 00:07:04.521 23:40:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:04.521 23:40:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:07:04.521 23:40:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:07:04.521 23:40:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:07:04.521 23:40:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:07:04.521 23:40:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:07:04.521 23:40:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:07:04.521 23:40:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:07:04.521 23:40:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:04.521 23:40:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:04.521 23:40:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:04.521 23:40:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:07:04.521 23:40:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:04.521 23:40:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:04.521 [ 00:07:04.521 { 00:07:04.521 "name": "BaseBdev1", 00:07:04.521 "aliases": [ 00:07:04.521 "f448046b-b955-47b0-b6fc-aa732385b31a" 00:07:04.521 ], 00:07:04.521 "product_name": "Malloc disk", 00:07:04.521 "block_size": 512, 00:07:04.521 "num_blocks": 65536, 00:07:04.521 "uuid": "f448046b-b955-47b0-b6fc-aa732385b31a", 00:07:04.521 "assigned_rate_limits": { 00:07:04.521 "rw_ios_per_sec": 0, 00:07:04.521 "rw_mbytes_per_sec": 0, 00:07:04.521 "r_mbytes_per_sec": 0, 00:07:04.521 "w_mbytes_per_sec": 0 00:07:04.521 }, 00:07:04.521 "claimed": true, 00:07:04.521 "claim_type": "exclusive_write", 00:07:04.521 "zoned": false, 00:07:04.521 "supported_io_types": { 00:07:04.521 "read": true, 00:07:04.521 "write": true, 00:07:04.521 "unmap": true, 00:07:04.521 "flush": true, 00:07:04.521 "reset": true, 00:07:04.521 "nvme_admin": false, 00:07:04.521 "nvme_io": false, 00:07:04.521 "nvme_io_md": false, 00:07:04.521 "write_zeroes": true, 00:07:04.521 "zcopy": true, 00:07:04.521 "get_zone_info": false, 00:07:04.521 "zone_management": false, 00:07:04.521 "zone_append": false, 00:07:04.521 "compare": false, 00:07:04.521 "compare_and_write": false, 00:07:04.521 "abort": true, 00:07:04.521 "seek_hole": false, 00:07:04.521 "seek_data": false, 00:07:04.521 "copy": true, 00:07:04.521 "nvme_iov_md": false 00:07:04.521 }, 00:07:04.521 "memory_domains": [ 00:07:04.521 { 00:07:04.521 "dma_device_id": "system", 00:07:04.521 "dma_device_type": 1 00:07:04.521 }, 00:07:04.521 { 00:07:04.521 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:04.521 "dma_device_type": 2 00:07:04.521 } 00:07:04.521 ], 00:07:04.521 "driver_specific": {} 00:07:04.521 } 00:07:04.521 ] 00:07:04.521 23:40:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:04.521 23:40:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:07:04.521 23:40:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:07:04.521 23:40:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:04.521 23:40:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:04.521 23:40:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:07:04.521 23:40:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:07:04.521 23:40:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:04.521 23:40:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:04.521 23:40:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:04.521 23:40:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:04.521 23:40:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:04.521 23:40:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:04.522 23:40:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:04.522 23:40:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:04.522 23:40:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:04.522 23:40:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:04.522 23:40:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:04.522 "name": "Existed_Raid", 00:07:04.522 "uuid": "f1b27a93-a7dd-449d-88fe-c64005af0101", 00:07:04.522 "strip_size_kb": 0, 00:07:04.522 "state": "configuring", 00:07:04.522 "raid_level": "raid1", 00:07:04.522 "superblock": true, 00:07:04.522 "num_base_bdevs": 2, 00:07:04.522 "num_base_bdevs_discovered": 1, 00:07:04.522 "num_base_bdevs_operational": 2, 00:07:04.522 "base_bdevs_list": [ 00:07:04.522 { 00:07:04.522 "name": "BaseBdev1", 00:07:04.522 "uuid": "f448046b-b955-47b0-b6fc-aa732385b31a", 00:07:04.522 "is_configured": true, 00:07:04.522 "data_offset": 2048, 00:07:04.522 "data_size": 63488 00:07:04.522 }, 00:07:04.522 { 00:07:04.522 "name": "BaseBdev2", 00:07:04.522 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:04.522 "is_configured": false, 00:07:04.522 "data_offset": 0, 00:07:04.522 "data_size": 0 00:07:04.522 } 00:07:04.522 ] 00:07:04.522 }' 00:07:04.522 23:40:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:04.522 23:40:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:05.090 23:40:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:07:05.090 23:40:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:05.090 23:40:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:05.090 [2024-11-26 23:40:53.006572] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:07:05.090 [2024-11-26 23:40:53.006637] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:07:05.090 23:40:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:05.090 23:40:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:05.090 23:40:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:05.090 23:40:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:05.090 [2024-11-26 23:40:53.018583] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:05.090 [2024-11-26 23:40:53.020517] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:05.090 [2024-11-26 23:40:53.020560] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:05.090 23:40:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:05.090 23:40:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:07:05.090 23:40:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:07:05.090 23:40:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:07:05.090 23:40:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:05.090 23:40:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:05.090 23:40:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:07:05.090 23:40:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:07:05.090 23:40:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:05.090 23:40:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:05.090 23:40:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:05.090 23:40:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:05.090 23:40:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:05.090 23:40:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:05.090 23:40:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:05.090 23:40:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:05.090 23:40:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:05.090 23:40:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:05.090 23:40:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:05.090 "name": "Existed_Raid", 00:07:05.090 "uuid": "ac5274bf-ce87-4808-be2e-e65d8784c858", 00:07:05.090 "strip_size_kb": 0, 00:07:05.090 "state": "configuring", 00:07:05.090 "raid_level": "raid1", 00:07:05.090 "superblock": true, 00:07:05.090 "num_base_bdevs": 2, 00:07:05.090 "num_base_bdevs_discovered": 1, 00:07:05.090 "num_base_bdevs_operational": 2, 00:07:05.090 "base_bdevs_list": [ 00:07:05.090 { 00:07:05.090 "name": "BaseBdev1", 00:07:05.090 "uuid": "f448046b-b955-47b0-b6fc-aa732385b31a", 00:07:05.090 "is_configured": true, 00:07:05.090 "data_offset": 2048, 00:07:05.090 "data_size": 63488 00:07:05.090 }, 00:07:05.090 { 00:07:05.090 "name": "BaseBdev2", 00:07:05.090 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:05.090 "is_configured": false, 00:07:05.090 "data_offset": 0, 00:07:05.090 "data_size": 0 00:07:05.090 } 00:07:05.090 ] 00:07:05.090 }' 00:07:05.090 23:40:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:05.090 23:40:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:05.349 23:40:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:07:05.349 23:40:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:05.349 23:40:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:05.349 [2024-11-26 23:40:53.456895] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:07:05.349 [2024-11-26 23:40:53.457189] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:07:05.349 [2024-11-26 23:40:53.457247] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:07:05.349 BaseBdev2 00:07:05.349 [2024-11-26 23:40:53.457567] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:07:05.349 [2024-11-26 23:40:53.457761] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:07:05.349 [2024-11-26 23:40:53.457813] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:07:05.349 [2024-11-26 23:40:53.457965] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:05.349 23:40:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:05.349 23:40:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:07:05.349 23:40:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:07:05.349 23:40:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:07:05.349 23:40:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:07:05.349 23:40:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:07:05.349 23:40:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:07:05.349 23:40:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:07:05.349 23:40:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:05.349 23:40:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:05.349 23:40:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:05.349 23:40:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:07:05.349 23:40:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:05.349 23:40:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:05.609 [ 00:07:05.609 { 00:07:05.609 "name": "BaseBdev2", 00:07:05.609 "aliases": [ 00:07:05.609 "aba04dff-b73c-40f7-b1eb-938a5a164e6f" 00:07:05.609 ], 00:07:05.609 "product_name": "Malloc disk", 00:07:05.609 "block_size": 512, 00:07:05.609 "num_blocks": 65536, 00:07:05.609 "uuid": "aba04dff-b73c-40f7-b1eb-938a5a164e6f", 00:07:05.609 "assigned_rate_limits": { 00:07:05.609 "rw_ios_per_sec": 0, 00:07:05.609 "rw_mbytes_per_sec": 0, 00:07:05.609 "r_mbytes_per_sec": 0, 00:07:05.609 "w_mbytes_per_sec": 0 00:07:05.609 }, 00:07:05.609 "claimed": true, 00:07:05.609 "claim_type": "exclusive_write", 00:07:05.609 "zoned": false, 00:07:05.609 "supported_io_types": { 00:07:05.609 "read": true, 00:07:05.609 "write": true, 00:07:05.609 "unmap": true, 00:07:05.609 "flush": true, 00:07:05.609 "reset": true, 00:07:05.609 "nvme_admin": false, 00:07:05.609 "nvme_io": false, 00:07:05.609 "nvme_io_md": false, 00:07:05.609 "write_zeroes": true, 00:07:05.609 "zcopy": true, 00:07:05.609 "get_zone_info": false, 00:07:05.609 "zone_management": false, 00:07:05.609 "zone_append": false, 00:07:05.609 "compare": false, 00:07:05.609 "compare_and_write": false, 00:07:05.609 "abort": true, 00:07:05.609 "seek_hole": false, 00:07:05.609 "seek_data": false, 00:07:05.609 "copy": true, 00:07:05.609 "nvme_iov_md": false 00:07:05.609 }, 00:07:05.609 "memory_domains": [ 00:07:05.609 { 00:07:05.609 "dma_device_id": "system", 00:07:05.609 "dma_device_type": 1 00:07:05.609 }, 00:07:05.609 { 00:07:05.609 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:05.609 "dma_device_type": 2 00:07:05.609 } 00:07:05.609 ], 00:07:05.609 "driver_specific": {} 00:07:05.609 } 00:07:05.609 ] 00:07:05.609 23:40:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:05.609 23:40:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:07:05.609 23:40:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:07:05.609 23:40:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:07:05.609 23:40:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid1 0 2 00:07:05.609 23:40:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:05.609 23:40:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:05.609 23:40:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:07:05.609 23:40:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:07:05.609 23:40:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:05.609 23:40:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:05.609 23:40:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:05.609 23:40:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:05.609 23:40:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:05.609 23:40:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:05.609 23:40:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:05.609 23:40:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:05.609 23:40:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:05.609 23:40:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:05.609 23:40:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:05.609 "name": "Existed_Raid", 00:07:05.609 "uuid": "ac5274bf-ce87-4808-be2e-e65d8784c858", 00:07:05.609 "strip_size_kb": 0, 00:07:05.609 "state": "online", 00:07:05.609 "raid_level": "raid1", 00:07:05.609 "superblock": true, 00:07:05.609 "num_base_bdevs": 2, 00:07:05.609 "num_base_bdevs_discovered": 2, 00:07:05.609 "num_base_bdevs_operational": 2, 00:07:05.609 "base_bdevs_list": [ 00:07:05.609 { 00:07:05.609 "name": "BaseBdev1", 00:07:05.609 "uuid": "f448046b-b955-47b0-b6fc-aa732385b31a", 00:07:05.609 "is_configured": true, 00:07:05.609 "data_offset": 2048, 00:07:05.609 "data_size": 63488 00:07:05.609 }, 00:07:05.609 { 00:07:05.609 "name": "BaseBdev2", 00:07:05.609 "uuid": "aba04dff-b73c-40f7-b1eb-938a5a164e6f", 00:07:05.609 "is_configured": true, 00:07:05.609 "data_offset": 2048, 00:07:05.609 "data_size": 63488 00:07:05.609 } 00:07:05.609 ] 00:07:05.609 }' 00:07:05.609 23:40:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:05.609 23:40:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:05.870 23:40:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:07:05.870 23:40:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:07:05.870 23:40:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:07:05.870 23:40:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:07:05.870 23:40:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:07:05.870 23:40:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:07:05.870 23:40:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:07:05.870 23:40:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:05.870 23:40:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:05.870 23:40:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:07:05.870 [2024-11-26 23:40:53.940436] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:05.870 23:40:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:05.870 23:40:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:07:05.870 "name": "Existed_Raid", 00:07:05.870 "aliases": [ 00:07:05.870 "ac5274bf-ce87-4808-be2e-e65d8784c858" 00:07:05.870 ], 00:07:05.870 "product_name": "Raid Volume", 00:07:05.870 "block_size": 512, 00:07:05.870 "num_blocks": 63488, 00:07:05.870 "uuid": "ac5274bf-ce87-4808-be2e-e65d8784c858", 00:07:05.870 "assigned_rate_limits": { 00:07:05.870 "rw_ios_per_sec": 0, 00:07:05.870 "rw_mbytes_per_sec": 0, 00:07:05.870 "r_mbytes_per_sec": 0, 00:07:05.870 "w_mbytes_per_sec": 0 00:07:05.870 }, 00:07:05.870 "claimed": false, 00:07:05.870 "zoned": false, 00:07:05.870 "supported_io_types": { 00:07:05.870 "read": true, 00:07:05.870 "write": true, 00:07:05.870 "unmap": false, 00:07:05.870 "flush": false, 00:07:05.870 "reset": true, 00:07:05.870 "nvme_admin": false, 00:07:05.870 "nvme_io": false, 00:07:05.870 "nvme_io_md": false, 00:07:05.870 "write_zeroes": true, 00:07:05.870 "zcopy": false, 00:07:05.870 "get_zone_info": false, 00:07:05.870 "zone_management": false, 00:07:05.870 "zone_append": false, 00:07:05.870 "compare": false, 00:07:05.870 "compare_and_write": false, 00:07:05.870 "abort": false, 00:07:05.870 "seek_hole": false, 00:07:05.870 "seek_data": false, 00:07:05.870 "copy": false, 00:07:05.870 "nvme_iov_md": false 00:07:05.870 }, 00:07:05.870 "memory_domains": [ 00:07:05.870 { 00:07:05.870 "dma_device_id": "system", 00:07:05.870 "dma_device_type": 1 00:07:05.870 }, 00:07:05.870 { 00:07:05.870 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:05.870 "dma_device_type": 2 00:07:05.870 }, 00:07:05.870 { 00:07:05.870 "dma_device_id": "system", 00:07:05.870 "dma_device_type": 1 00:07:05.870 }, 00:07:05.870 { 00:07:05.870 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:05.870 "dma_device_type": 2 00:07:05.870 } 00:07:05.870 ], 00:07:05.870 "driver_specific": { 00:07:05.870 "raid": { 00:07:05.870 "uuid": "ac5274bf-ce87-4808-be2e-e65d8784c858", 00:07:05.870 "strip_size_kb": 0, 00:07:05.870 "state": "online", 00:07:05.870 "raid_level": "raid1", 00:07:05.870 "superblock": true, 00:07:05.870 "num_base_bdevs": 2, 00:07:05.870 "num_base_bdevs_discovered": 2, 00:07:05.870 "num_base_bdevs_operational": 2, 00:07:05.870 "base_bdevs_list": [ 00:07:05.870 { 00:07:05.870 "name": "BaseBdev1", 00:07:05.870 "uuid": "f448046b-b955-47b0-b6fc-aa732385b31a", 00:07:05.870 "is_configured": true, 00:07:05.870 "data_offset": 2048, 00:07:05.870 "data_size": 63488 00:07:05.870 }, 00:07:05.870 { 00:07:05.870 "name": "BaseBdev2", 00:07:05.870 "uuid": "aba04dff-b73c-40f7-b1eb-938a5a164e6f", 00:07:05.870 "is_configured": true, 00:07:05.870 "data_offset": 2048, 00:07:05.870 "data_size": 63488 00:07:05.870 } 00:07:05.870 ] 00:07:05.870 } 00:07:05.870 } 00:07:05.870 }' 00:07:05.870 23:40:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:07:06.130 23:40:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:07:06.130 BaseBdev2' 00:07:06.130 23:40:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:06.130 23:40:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:07:06.130 23:40:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:06.130 23:40:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:06.130 23:40:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:07:06.130 23:40:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:06.130 23:40:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:06.130 23:40:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:06.130 23:40:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:06.130 23:40:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:06.130 23:40:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:06.130 23:40:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:07:06.130 23:40:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:06.130 23:40:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:06.130 23:40:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:06.130 23:40:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:06.130 23:40:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:06.130 23:40:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:06.130 23:40:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:07:06.130 23:40:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:06.130 23:40:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:06.130 [2024-11-26 23:40:54.183796] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:07:06.130 23:40:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:06.130 23:40:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:07:06.130 23:40:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy raid1 00:07:06.130 23:40:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:07:06.130 23:40:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@199 -- # return 0 00:07:06.130 23:40:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:07:06.130 23:40:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid1 0 1 00:07:06.130 23:40:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:06.130 23:40:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:06.130 23:40:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:07:06.130 23:40:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:07:06.130 23:40:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:07:06.130 23:40:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:06.130 23:40:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:06.130 23:40:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:06.130 23:40:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:06.130 23:40:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:06.130 23:40:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:06.130 23:40:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:06.130 23:40:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:06.130 23:40:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:06.130 23:40:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:06.130 "name": "Existed_Raid", 00:07:06.131 "uuid": "ac5274bf-ce87-4808-be2e-e65d8784c858", 00:07:06.131 "strip_size_kb": 0, 00:07:06.131 "state": "online", 00:07:06.131 "raid_level": "raid1", 00:07:06.131 "superblock": true, 00:07:06.131 "num_base_bdevs": 2, 00:07:06.131 "num_base_bdevs_discovered": 1, 00:07:06.131 "num_base_bdevs_operational": 1, 00:07:06.131 "base_bdevs_list": [ 00:07:06.131 { 00:07:06.131 "name": null, 00:07:06.131 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:06.131 "is_configured": false, 00:07:06.131 "data_offset": 0, 00:07:06.131 "data_size": 63488 00:07:06.131 }, 00:07:06.131 { 00:07:06.131 "name": "BaseBdev2", 00:07:06.131 "uuid": "aba04dff-b73c-40f7-b1eb-938a5a164e6f", 00:07:06.131 "is_configured": true, 00:07:06.131 "data_offset": 2048, 00:07:06.131 "data_size": 63488 00:07:06.131 } 00:07:06.131 ] 00:07:06.131 }' 00:07:06.131 23:40:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:06.131 23:40:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:06.699 23:40:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:07:06.699 23:40:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:07:06.699 23:40:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:06.699 23:40:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:06.699 23:40:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:06.699 23:40:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:07:06.699 23:40:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:06.699 23:40:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:07:06.699 23:40:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:07:06.699 23:40:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:07:06.699 23:40:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:06.699 23:40:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:06.699 [2024-11-26 23:40:54.614488] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:07:06.699 [2024-11-26 23:40:54.614656] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:06.699 [2024-11-26 23:40:54.626443] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:06.699 [2024-11-26 23:40:54.626572] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:06.699 [2024-11-26 23:40:54.626616] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:07:06.699 23:40:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:06.699 23:40:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:07:06.699 23:40:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:07:06.699 23:40:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:06.699 23:40:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:07:06.699 23:40:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:06.699 23:40:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:06.699 23:40:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:06.699 23:40:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:07:06.699 23:40:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:07:06.699 23:40:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 2 -gt 2 ']' 00:07:06.699 23:40:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 73938 00:07:06.699 23:40:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # '[' -z 73938 ']' 00:07:06.699 23:40:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@958 -- # kill -0 73938 00:07:06.699 23:40:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # uname 00:07:06.699 23:40:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:07:06.699 23:40:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 73938 00:07:06.699 killing process with pid 73938 00:07:06.699 23:40:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:07:06.699 23:40:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:07:06.699 23:40:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@972 -- # echo 'killing process with pid 73938' 00:07:06.699 23:40:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@973 -- # kill 73938 00:07:06.699 [2024-11-26 23:40:54.712262] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:06.699 23:40:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@978 -- # wait 73938 00:07:06.699 [2024-11-26 23:40:54.713291] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:06.958 23:40:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:07:06.958 00:07:06.958 real 0m3.743s 00:07:06.958 user 0m5.838s 00:07:06.958 sys 0m0.798s 00:07:06.958 23:40:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1130 -- # xtrace_disable 00:07:06.958 23:40:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:06.958 ************************************ 00:07:06.958 END TEST raid_state_function_test_sb 00:07:06.958 ************************************ 00:07:06.958 23:40:54 bdev_raid -- bdev/bdev_raid.sh@970 -- # run_test raid_superblock_test raid_superblock_test raid1 2 00:07:06.958 23:40:54 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:07:06.958 23:40:54 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:06.958 23:40:54 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:06.958 ************************************ 00:07:06.958 START TEST raid_superblock_test 00:07:06.958 ************************************ 00:07:06.958 23:40:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1129 -- # raid_superblock_test raid1 2 00:07:06.958 23:40:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=raid1 00:07:06.958 23:40:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=2 00:07:06.958 23:40:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:07:06.958 23:40:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:07:06.958 23:40:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:07:06.958 23:40:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:07:06.958 23:40:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:07:06.958 23:40:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:07:06.958 23:40:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:07:06.958 23:40:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:07:06.958 23:40:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:07:06.958 23:40:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:07:06.958 23:40:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:07:06.958 23:40:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' raid1 '!=' raid1 ']' 00:07:06.958 23:40:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@408 -- # strip_size=0 00:07:06.958 23:40:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=74179 00:07:06.958 23:40:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:07:06.958 23:40:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 74179 00:07:06.958 23:40:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@835 -- # '[' -z 74179 ']' 00:07:06.958 23:40:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:06.958 23:40:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:06.958 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:06.958 23:40:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:06.958 23:40:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:06.958 23:40:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:06.958 [2024-11-26 23:40:55.066215] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:07:06.958 [2024-11-26 23:40:55.066460] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid74179 ] 00:07:07.218 [2024-11-26 23:40:55.219786] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:07.218 [2024-11-26 23:40:55.247380] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:07.218 [2024-11-26 23:40:55.290980] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:07.218 [2024-11-26 23:40:55.291097] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:07.786 23:40:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:07.786 23:40:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@868 -- # return 0 00:07:07.786 23:40:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:07:07.786 23:40:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:07:07.786 23:40:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:07:07.786 23:40:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:07:07.786 23:40:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:07:07.786 23:40:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:07:07.786 23:40:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:07:07.786 23:40:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:07:07.786 23:40:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:07:07.786 23:40:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:07.786 23:40:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:07.786 malloc1 00:07:07.786 23:40:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:07.786 23:40:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:07:07.786 23:40:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:07.786 23:40:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:08.046 [2024-11-26 23:40:55.920064] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:07:08.046 [2024-11-26 23:40:55.920197] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:08.046 [2024-11-26 23:40:55.920237] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:07:08.047 [2024-11-26 23:40:55.920270] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:08.047 [2024-11-26 23:40:55.922443] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:08.047 [2024-11-26 23:40:55.922539] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:07:08.047 pt1 00:07:08.047 23:40:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:08.047 23:40:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:07:08.047 23:40:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:07:08.047 23:40:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:07:08.047 23:40:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:07:08.047 23:40:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:07:08.047 23:40:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:07:08.047 23:40:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:07:08.047 23:40:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:07:08.047 23:40:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:07:08.047 23:40:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:08.047 23:40:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:08.047 malloc2 00:07:08.047 23:40:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:08.047 23:40:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:07:08.047 23:40:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:08.047 23:40:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:08.047 [2024-11-26 23:40:55.948844] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:07:08.047 [2024-11-26 23:40:55.948953] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:08.047 [2024-11-26 23:40:55.948992] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:07:08.047 [2024-11-26 23:40:55.949023] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:08.047 [2024-11-26 23:40:55.951395] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:08.047 [2024-11-26 23:40:55.951473] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:07:08.047 pt2 00:07:08.047 23:40:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:08.047 23:40:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:07:08.047 23:40:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:07:08.047 23:40:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''pt1 pt2'\''' -n raid_bdev1 -s 00:07:08.047 23:40:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:08.047 23:40:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:08.047 [2024-11-26 23:40:55.960859] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:07:08.047 [2024-11-26 23:40:55.962856] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:07:08.047 [2024-11-26 23:40:55.963064] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:07:08.047 [2024-11-26 23:40:55.963124] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:07:08.047 [2024-11-26 23:40:55.963457] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:07:08.047 [2024-11-26 23:40:55.963658] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:07:08.047 [2024-11-26 23:40:55.963703] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:07:08.047 [2024-11-26 23:40:55.963877] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:08.047 23:40:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:08.047 23:40:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:07:08.047 23:40:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:08.047 23:40:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:08.047 23:40:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:07:08.047 23:40:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:07:08.047 23:40:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:08.047 23:40:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:08.047 23:40:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:08.047 23:40:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:08.047 23:40:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:08.047 23:40:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:08.047 23:40:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:08.047 23:40:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:08.047 23:40:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:08.047 23:40:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:08.047 23:40:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:08.047 "name": "raid_bdev1", 00:07:08.047 "uuid": "d8eddafd-d63f-438d-9223-8b4e73930d34", 00:07:08.047 "strip_size_kb": 0, 00:07:08.047 "state": "online", 00:07:08.047 "raid_level": "raid1", 00:07:08.047 "superblock": true, 00:07:08.047 "num_base_bdevs": 2, 00:07:08.047 "num_base_bdevs_discovered": 2, 00:07:08.047 "num_base_bdevs_operational": 2, 00:07:08.047 "base_bdevs_list": [ 00:07:08.047 { 00:07:08.047 "name": "pt1", 00:07:08.047 "uuid": "00000000-0000-0000-0000-000000000001", 00:07:08.047 "is_configured": true, 00:07:08.047 "data_offset": 2048, 00:07:08.047 "data_size": 63488 00:07:08.047 }, 00:07:08.047 { 00:07:08.047 "name": "pt2", 00:07:08.047 "uuid": "00000000-0000-0000-0000-000000000002", 00:07:08.047 "is_configured": true, 00:07:08.047 "data_offset": 2048, 00:07:08.047 "data_size": 63488 00:07:08.047 } 00:07:08.047 ] 00:07:08.047 }' 00:07:08.047 23:40:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:08.047 23:40:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:08.312 23:40:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:07:08.312 23:40:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:07:08.312 23:40:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:07:08.312 23:40:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:07:08.312 23:40:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:07:08.312 23:40:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:07:08.312 23:40:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:07:08.312 23:40:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:07:08.313 23:40:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:08.313 23:40:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:08.313 [2024-11-26 23:40:56.392443] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:08.313 23:40:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:08.313 23:40:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:07:08.313 "name": "raid_bdev1", 00:07:08.313 "aliases": [ 00:07:08.313 "d8eddafd-d63f-438d-9223-8b4e73930d34" 00:07:08.313 ], 00:07:08.313 "product_name": "Raid Volume", 00:07:08.313 "block_size": 512, 00:07:08.313 "num_blocks": 63488, 00:07:08.313 "uuid": "d8eddafd-d63f-438d-9223-8b4e73930d34", 00:07:08.313 "assigned_rate_limits": { 00:07:08.313 "rw_ios_per_sec": 0, 00:07:08.313 "rw_mbytes_per_sec": 0, 00:07:08.313 "r_mbytes_per_sec": 0, 00:07:08.313 "w_mbytes_per_sec": 0 00:07:08.313 }, 00:07:08.313 "claimed": false, 00:07:08.313 "zoned": false, 00:07:08.313 "supported_io_types": { 00:07:08.313 "read": true, 00:07:08.313 "write": true, 00:07:08.313 "unmap": false, 00:07:08.313 "flush": false, 00:07:08.313 "reset": true, 00:07:08.313 "nvme_admin": false, 00:07:08.313 "nvme_io": false, 00:07:08.313 "nvme_io_md": false, 00:07:08.313 "write_zeroes": true, 00:07:08.313 "zcopy": false, 00:07:08.313 "get_zone_info": false, 00:07:08.313 "zone_management": false, 00:07:08.313 "zone_append": false, 00:07:08.313 "compare": false, 00:07:08.313 "compare_and_write": false, 00:07:08.313 "abort": false, 00:07:08.313 "seek_hole": false, 00:07:08.313 "seek_data": false, 00:07:08.313 "copy": false, 00:07:08.313 "nvme_iov_md": false 00:07:08.313 }, 00:07:08.313 "memory_domains": [ 00:07:08.313 { 00:07:08.313 "dma_device_id": "system", 00:07:08.313 "dma_device_type": 1 00:07:08.313 }, 00:07:08.313 { 00:07:08.313 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:08.313 "dma_device_type": 2 00:07:08.313 }, 00:07:08.313 { 00:07:08.313 "dma_device_id": "system", 00:07:08.313 "dma_device_type": 1 00:07:08.313 }, 00:07:08.313 { 00:07:08.313 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:08.313 "dma_device_type": 2 00:07:08.313 } 00:07:08.313 ], 00:07:08.313 "driver_specific": { 00:07:08.313 "raid": { 00:07:08.313 "uuid": "d8eddafd-d63f-438d-9223-8b4e73930d34", 00:07:08.313 "strip_size_kb": 0, 00:07:08.313 "state": "online", 00:07:08.313 "raid_level": "raid1", 00:07:08.313 "superblock": true, 00:07:08.313 "num_base_bdevs": 2, 00:07:08.313 "num_base_bdevs_discovered": 2, 00:07:08.313 "num_base_bdevs_operational": 2, 00:07:08.313 "base_bdevs_list": [ 00:07:08.313 { 00:07:08.313 "name": "pt1", 00:07:08.313 "uuid": "00000000-0000-0000-0000-000000000001", 00:07:08.313 "is_configured": true, 00:07:08.313 "data_offset": 2048, 00:07:08.313 "data_size": 63488 00:07:08.313 }, 00:07:08.313 { 00:07:08.313 "name": "pt2", 00:07:08.313 "uuid": "00000000-0000-0000-0000-000000000002", 00:07:08.313 "is_configured": true, 00:07:08.313 "data_offset": 2048, 00:07:08.313 "data_size": 63488 00:07:08.313 } 00:07:08.313 ] 00:07:08.313 } 00:07:08.313 } 00:07:08.313 }' 00:07:08.313 23:40:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:07:08.572 23:40:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:07:08.572 pt2' 00:07:08.572 23:40:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:08.572 23:40:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:07:08.572 23:40:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:08.572 23:40:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:08.572 23:40:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:07:08.572 23:40:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:08.572 23:40:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:08.572 23:40:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:08.572 23:40:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:08.572 23:40:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:08.572 23:40:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:08.572 23:40:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:07:08.572 23:40:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:08.572 23:40:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:08.572 23:40:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:08.572 23:40:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:08.572 23:40:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:08.572 23:40:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:08.572 23:40:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:07:08.572 23:40:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:07:08.572 23:40:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:08.572 23:40:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:08.572 [2024-11-26 23:40:56.631937] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:08.572 23:40:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:08.572 23:40:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=d8eddafd-d63f-438d-9223-8b4e73930d34 00:07:08.572 23:40:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z d8eddafd-d63f-438d-9223-8b4e73930d34 ']' 00:07:08.572 23:40:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:07:08.572 23:40:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:08.572 23:40:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:08.572 [2024-11-26 23:40:56.667603] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:07:08.572 [2024-11-26 23:40:56.667671] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:08.572 [2024-11-26 23:40:56.667771] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:08.572 [2024-11-26 23:40:56.667853] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:08.572 [2024-11-26 23:40:56.667885] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:07:08.572 23:40:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:08.572 23:40:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:08.572 23:40:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:08.572 23:40:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:08.572 23:40:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:07:08.572 23:40:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:08.833 23:40:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:07:08.833 23:40:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:07:08.833 23:40:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:07:08.833 23:40:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:07:08.833 23:40:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:08.833 23:40:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:08.833 23:40:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:08.833 23:40:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:07:08.833 23:40:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:07:08.833 23:40:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:08.833 23:40:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:08.833 23:40:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:08.833 23:40:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:07:08.833 23:40:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:07:08.833 23:40:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:08.833 23:40:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:08.833 23:40:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:08.833 23:40:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:07:08.833 23:40:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:07:08.833 23:40:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@652 -- # local es=0 00:07:08.833 23:40:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:07:08.833 23:40:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:07:08.833 23:40:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:07:08.833 23:40:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:07:08.833 23:40:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:07:08.833 23:40:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:07:08.833 23:40:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:08.833 23:40:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:08.833 [2024-11-26 23:40:56.803439] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:07:08.833 [2024-11-26 23:40:56.805307] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:07:08.833 [2024-11-26 23:40:56.805476] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:07:08.833 [2024-11-26 23:40:56.805582] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:07:08.833 [2024-11-26 23:40:56.805604] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:07:08.833 [2024-11-26 23:40:56.805614] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state configuring 00:07:08.833 request: 00:07:08.833 { 00:07:08.833 "name": "raid_bdev1", 00:07:08.833 "raid_level": "raid1", 00:07:08.833 "base_bdevs": [ 00:07:08.833 "malloc1", 00:07:08.833 "malloc2" 00:07:08.833 ], 00:07:08.833 "superblock": false, 00:07:08.833 "method": "bdev_raid_create", 00:07:08.833 "req_id": 1 00:07:08.833 } 00:07:08.833 Got JSON-RPC error response 00:07:08.833 response: 00:07:08.833 { 00:07:08.833 "code": -17, 00:07:08.833 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:07:08.833 } 00:07:08.833 23:40:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:07:08.833 23:40:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # es=1 00:07:08.833 23:40:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:07:08.833 23:40:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:07:08.833 23:40:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:07:08.833 23:40:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:07:08.833 23:40:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:08.833 23:40:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:08.833 23:40:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:08.833 23:40:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:08.833 23:40:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:07:08.833 23:40:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:07:08.833 23:40:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:07:08.833 23:40:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:08.833 23:40:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:08.833 [2024-11-26 23:40:56.855310] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:07:08.833 [2024-11-26 23:40:56.855424] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:08.833 [2024-11-26 23:40:56.855465] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:07:08.833 [2024-11-26 23:40:56.855493] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:08.833 [2024-11-26 23:40:56.857642] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:08.833 [2024-11-26 23:40:56.857712] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:07:08.833 [2024-11-26 23:40:56.857799] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:07:08.833 [2024-11-26 23:40:56.857880] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:07:08.833 pt1 00:07:08.833 23:40:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:08.833 23:40:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 2 00:07:08.833 23:40:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:08.833 23:40:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:08.833 23:40:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:07:08.833 23:40:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:07:08.833 23:40:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:08.833 23:40:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:08.833 23:40:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:08.833 23:40:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:08.833 23:40:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:08.833 23:40:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:08.833 23:40:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:08.833 23:40:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:08.833 23:40:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:08.833 23:40:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:08.833 23:40:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:08.833 "name": "raid_bdev1", 00:07:08.833 "uuid": "d8eddafd-d63f-438d-9223-8b4e73930d34", 00:07:08.833 "strip_size_kb": 0, 00:07:08.833 "state": "configuring", 00:07:08.833 "raid_level": "raid1", 00:07:08.833 "superblock": true, 00:07:08.833 "num_base_bdevs": 2, 00:07:08.833 "num_base_bdevs_discovered": 1, 00:07:08.833 "num_base_bdevs_operational": 2, 00:07:08.833 "base_bdevs_list": [ 00:07:08.833 { 00:07:08.833 "name": "pt1", 00:07:08.833 "uuid": "00000000-0000-0000-0000-000000000001", 00:07:08.833 "is_configured": true, 00:07:08.833 "data_offset": 2048, 00:07:08.833 "data_size": 63488 00:07:08.833 }, 00:07:08.833 { 00:07:08.833 "name": null, 00:07:08.833 "uuid": "00000000-0000-0000-0000-000000000002", 00:07:08.833 "is_configured": false, 00:07:08.833 "data_offset": 2048, 00:07:08.833 "data_size": 63488 00:07:08.833 } 00:07:08.833 ] 00:07:08.833 }' 00:07:08.833 23:40:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:08.833 23:40:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:09.093 23:40:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 2 -gt 2 ']' 00:07:09.093 23:40:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:07:09.093 23:40:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:07:09.093 23:40:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:07:09.093 23:40:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:09.093 23:40:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:09.351 [2024-11-26 23:40:57.222681] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:07:09.351 [2024-11-26 23:40:57.222785] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:09.351 [2024-11-26 23:40:57.222811] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:07:09.351 [2024-11-26 23:40:57.222819] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:09.351 [2024-11-26 23:40:57.223220] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:09.351 [2024-11-26 23:40:57.223239] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:07:09.351 [2024-11-26 23:40:57.223307] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:07:09.351 [2024-11-26 23:40:57.223327] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:07:09.351 [2024-11-26 23:40:57.223456] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:07:09.351 [2024-11-26 23:40:57.223467] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:07:09.351 [2024-11-26 23:40:57.223711] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:07:09.351 [2024-11-26 23:40:57.223829] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:07:09.351 [2024-11-26 23:40:57.223844] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:07:09.351 [2024-11-26 23:40:57.223948] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:09.351 pt2 00:07:09.351 23:40:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:09.351 23:40:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:07:09.351 23:40:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:07:09.351 23:40:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:07:09.351 23:40:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:09.351 23:40:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:09.351 23:40:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:07:09.351 23:40:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:07:09.351 23:40:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:09.351 23:40:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:09.351 23:40:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:09.351 23:40:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:09.351 23:40:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:09.351 23:40:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:09.351 23:40:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:09.351 23:40:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:09.351 23:40:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:09.351 23:40:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:09.351 23:40:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:09.351 "name": "raid_bdev1", 00:07:09.351 "uuid": "d8eddafd-d63f-438d-9223-8b4e73930d34", 00:07:09.351 "strip_size_kb": 0, 00:07:09.351 "state": "online", 00:07:09.351 "raid_level": "raid1", 00:07:09.351 "superblock": true, 00:07:09.351 "num_base_bdevs": 2, 00:07:09.351 "num_base_bdevs_discovered": 2, 00:07:09.351 "num_base_bdevs_operational": 2, 00:07:09.351 "base_bdevs_list": [ 00:07:09.351 { 00:07:09.351 "name": "pt1", 00:07:09.351 "uuid": "00000000-0000-0000-0000-000000000001", 00:07:09.351 "is_configured": true, 00:07:09.351 "data_offset": 2048, 00:07:09.351 "data_size": 63488 00:07:09.351 }, 00:07:09.351 { 00:07:09.351 "name": "pt2", 00:07:09.351 "uuid": "00000000-0000-0000-0000-000000000002", 00:07:09.351 "is_configured": true, 00:07:09.351 "data_offset": 2048, 00:07:09.351 "data_size": 63488 00:07:09.351 } 00:07:09.351 ] 00:07:09.351 }' 00:07:09.351 23:40:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:09.351 23:40:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:09.622 23:40:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:07:09.622 23:40:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:07:09.622 23:40:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:07:09.622 23:40:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:07:09.622 23:40:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:07:09.622 23:40:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:07:09.622 23:40:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:07:09.622 23:40:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:07:09.622 23:40:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:09.622 23:40:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:09.622 [2024-11-26 23:40:57.694480] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:09.622 23:40:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:09.622 23:40:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:07:09.622 "name": "raid_bdev1", 00:07:09.622 "aliases": [ 00:07:09.622 "d8eddafd-d63f-438d-9223-8b4e73930d34" 00:07:09.622 ], 00:07:09.622 "product_name": "Raid Volume", 00:07:09.622 "block_size": 512, 00:07:09.622 "num_blocks": 63488, 00:07:09.622 "uuid": "d8eddafd-d63f-438d-9223-8b4e73930d34", 00:07:09.622 "assigned_rate_limits": { 00:07:09.622 "rw_ios_per_sec": 0, 00:07:09.622 "rw_mbytes_per_sec": 0, 00:07:09.622 "r_mbytes_per_sec": 0, 00:07:09.622 "w_mbytes_per_sec": 0 00:07:09.622 }, 00:07:09.622 "claimed": false, 00:07:09.622 "zoned": false, 00:07:09.622 "supported_io_types": { 00:07:09.622 "read": true, 00:07:09.622 "write": true, 00:07:09.622 "unmap": false, 00:07:09.622 "flush": false, 00:07:09.622 "reset": true, 00:07:09.622 "nvme_admin": false, 00:07:09.622 "nvme_io": false, 00:07:09.622 "nvme_io_md": false, 00:07:09.622 "write_zeroes": true, 00:07:09.622 "zcopy": false, 00:07:09.622 "get_zone_info": false, 00:07:09.622 "zone_management": false, 00:07:09.622 "zone_append": false, 00:07:09.622 "compare": false, 00:07:09.622 "compare_and_write": false, 00:07:09.622 "abort": false, 00:07:09.622 "seek_hole": false, 00:07:09.622 "seek_data": false, 00:07:09.622 "copy": false, 00:07:09.622 "nvme_iov_md": false 00:07:09.622 }, 00:07:09.622 "memory_domains": [ 00:07:09.622 { 00:07:09.622 "dma_device_id": "system", 00:07:09.622 "dma_device_type": 1 00:07:09.622 }, 00:07:09.622 { 00:07:09.622 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:09.622 "dma_device_type": 2 00:07:09.622 }, 00:07:09.622 { 00:07:09.622 "dma_device_id": "system", 00:07:09.622 "dma_device_type": 1 00:07:09.622 }, 00:07:09.622 { 00:07:09.622 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:09.622 "dma_device_type": 2 00:07:09.622 } 00:07:09.622 ], 00:07:09.622 "driver_specific": { 00:07:09.622 "raid": { 00:07:09.622 "uuid": "d8eddafd-d63f-438d-9223-8b4e73930d34", 00:07:09.622 "strip_size_kb": 0, 00:07:09.622 "state": "online", 00:07:09.622 "raid_level": "raid1", 00:07:09.622 "superblock": true, 00:07:09.622 "num_base_bdevs": 2, 00:07:09.622 "num_base_bdevs_discovered": 2, 00:07:09.622 "num_base_bdevs_operational": 2, 00:07:09.622 "base_bdevs_list": [ 00:07:09.622 { 00:07:09.622 "name": "pt1", 00:07:09.622 "uuid": "00000000-0000-0000-0000-000000000001", 00:07:09.622 "is_configured": true, 00:07:09.622 "data_offset": 2048, 00:07:09.622 "data_size": 63488 00:07:09.622 }, 00:07:09.622 { 00:07:09.622 "name": "pt2", 00:07:09.622 "uuid": "00000000-0000-0000-0000-000000000002", 00:07:09.622 "is_configured": true, 00:07:09.622 "data_offset": 2048, 00:07:09.622 "data_size": 63488 00:07:09.622 } 00:07:09.622 ] 00:07:09.622 } 00:07:09.622 } 00:07:09.622 }' 00:07:09.622 23:40:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:07:09.881 23:40:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:07:09.881 pt2' 00:07:09.881 23:40:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:09.881 23:40:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:07:09.881 23:40:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:09.881 23:40:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:09.881 23:40:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:07:09.881 23:40:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:09.881 23:40:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:09.881 23:40:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:09.881 23:40:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:09.881 23:40:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:09.881 23:40:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:09.881 23:40:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:09.881 23:40:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:07:09.881 23:40:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:09.881 23:40:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:09.881 23:40:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:09.881 23:40:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:09.881 23:40:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:09.881 23:40:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:07:09.881 23:40:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:07:09.881 23:40:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:09.881 23:40:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:09.881 [2024-11-26 23:40:57.910475] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:09.881 23:40:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:09.881 23:40:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' d8eddafd-d63f-438d-9223-8b4e73930d34 '!=' d8eddafd-d63f-438d-9223-8b4e73930d34 ']' 00:07:09.881 23:40:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy raid1 00:07:09.881 23:40:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:07:09.881 23:40:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@199 -- # return 0 00:07:09.881 23:40:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@493 -- # rpc_cmd bdev_passthru_delete pt1 00:07:09.881 23:40:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:09.881 23:40:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:09.881 [2024-11-26 23:40:57.958176] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt1 00:07:09.881 23:40:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:09.881 23:40:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@496 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:07:09.881 23:40:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:09.881 23:40:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:09.881 23:40:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:07:09.881 23:40:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:07:09.881 23:40:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:07:09.881 23:40:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:09.881 23:40:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:09.881 23:40:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:09.881 23:40:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:09.881 23:40:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:09.881 23:40:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:09.881 23:40:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:09.881 23:40:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:09.881 23:40:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:09.881 23:40:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:09.881 "name": "raid_bdev1", 00:07:09.881 "uuid": "d8eddafd-d63f-438d-9223-8b4e73930d34", 00:07:09.881 "strip_size_kb": 0, 00:07:09.881 "state": "online", 00:07:09.881 "raid_level": "raid1", 00:07:09.881 "superblock": true, 00:07:09.881 "num_base_bdevs": 2, 00:07:09.881 "num_base_bdevs_discovered": 1, 00:07:09.881 "num_base_bdevs_operational": 1, 00:07:09.881 "base_bdevs_list": [ 00:07:09.881 { 00:07:09.881 "name": null, 00:07:09.881 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:09.881 "is_configured": false, 00:07:09.881 "data_offset": 0, 00:07:09.881 "data_size": 63488 00:07:09.881 }, 00:07:09.881 { 00:07:09.881 "name": "pt2", 00:07:09.881 "uuid": "00000000-0000-0000-0000-000000000002", 00:07:09.881 "is_configured": true, 00:07:09.881 "data_offset": 2048, 00:07:09.881 "data_size": 63488 00:07:09.881 } 00:07:09.881 ] 00:07:09.881 }' 00:07:09.881 23:40:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:09.881 23:40:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:10.450 23:40:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@499 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:07:10.450 23:40:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:10.450 23:40:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:10.450 [2024-11-26 23:40:58.381442] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:07:10.450 [2024-11-26 23:40:58.381525] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:10.450 [2024-11-26 23:40:58.381631] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:10.450 [2024-11-26 23:40:58.381717] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:10.450 [2024-11-26 23:40:58.381751] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:07:10.450 23:40:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:10.450 23:40:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # jq -r '.[]' 00:07:10.450 23:40:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:10.450 23:40:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:10.450 23:40:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:10.451 23:40:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:10.451 23:40:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # raid_bdev= 00:07:10.451 23:40:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@501 -- # '[' -n '' ']' 00:07:10.451 23:40:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i = 1 )) 00:07:10.451 23:40:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:07:10.451 23:40:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt2 00:07:10.451 23:40:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:10.451 23:40:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:10.451 23:40:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:10.451 23:40:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:07:10.451 23:40:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:07:10.451 23:40:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i = 1 )) 00:07:10.451 23:40:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:07:10.451 23:40:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@519 -- # i=1 00:07:10.451 23:40:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@520 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:07:10.451 23:40:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:10.451 23:40:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:10.451 [2024-11-26 23:40:58.433326] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:07:10.451 [2024-11-26 23:40:58.433391] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:10.451 [2024-11-26 23:40:58.433412] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008780 00:07:10.451 [2024-11-26 23:40:58.433420] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:10.451 [2024-11-26 23:40:58.435734] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:10.451 [2024-11-26 23:40:58.435771] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:07:10.451 [2024-11-26 23:40:58.435845] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:07:10.451 [2024-11-26 23:40:58.435875] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:07:10.451 [2024-11-26 23:40:58.435950] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:07:10.451 [2024-11-26 23:40:58.435958] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:07:10.451 [2024-11-26 23:40:58.436186] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:07:10.451 [2024-11-26 23:40:58.436318] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:07:10.451 [2024-11-26 23:40:58.436328] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001c80 00:07:10.451 [2024-11-26 23:40:58.436454] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:10.451 pt2 00:07:10.451 23:40:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:10.451 23:40:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@523 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:07:10.451 23:40:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:10.451 23:40:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:10.451 23:40:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:07:10.451 23:40:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:07:10.451 23:40:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:07:10.451 23:40:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:10.451 23:40:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:10.451 23:40:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:10.451 23:40:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:10.451 23:40:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:10.451 23:40:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:10.451 23:40:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:10.451 23:40:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:10.451 23:40:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:10.451 23:40:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:10.451 "name": "raid_bdev1", 00:07:10.451 "uuid": "d8eddafd-d63f-438d-9223-8b4e73930d34", 00:07:10.451 "strip_size_kb": 0, 00:07:10.451 "state": "online", 00:07:10.451 "raid_level": "raid1", 00:07:10.451 "superblock": true, 00:07:10.451 "num_base_bdevs": 2, 00:07:10.451 "num_base_bdevs_discovered": 1, 00:07:10.451 "num_base_bdevs_operational": 1, 00:07:10.451 "base_bdevs_list": [ 00:07:10.451 { 00:07:10.451 "name": null, 00:07:10.451 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:10.451 "is_configured": false, 00:07:10.451 "data_offset": 2048, 00:07:10.451 "data_size": 63488 00:07:10.451 }, 00:07:10.451 { 00:07:10.451 "name": "pt2", 00:07:10.451 "uuid": "00000000-0000-0000-0000-000000000002", 00:07:10.451 "is_configured": true, 00:07:10.451 "data_offset": 2048, 00:07:10.451 "data_size": 63488 00:07:10.451 } 00:07:10.451 ] 00:07:10.451 }' 00:07:10.451 23:40:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:10.451 23:40:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:11.020 23:40:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@526 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:07:11.020 23:40:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:11.020 23:40:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:11.020 [2024-11-26 23:40:58.860614] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:07:11.020 [2024-11-26 23:40:58.860642] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:11.020 [2024-11-26 23:40:58.860712] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:11.020 [2024-11-26 23:40:58.860758] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:11.020 [2024-11-26 23:40:58.860769] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name raid_bdev1, state offline 00:07:11.020 23:40:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:11.020 23:40:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # jq -r '.[]' 00:07:11.020 23:40:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:11.020 23:40:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:11.020 23:40:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:11.020 23:40:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:11.020 23:40:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # raid_bdev= 00:07:11.020 23:40:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@528 -- # '[' -n '' ']' 00:07:11.020 23:40:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@532 -- # '[' 2 -gt 2 ']' 00:07:11.020 23:40:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@540 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:07:11.020 23:40:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:11.020 23:40:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:11.020 [2024-11-26 23:40:58.904548] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:07:11.020 [2024-11-26 23:40:58.904630] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:11.020 [2024-11-26 23:40:58.904655] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008d80 00:07:11.020 [2024-11-26 23:40:58.904674] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:11.020 [2024-11-26 23:40:58.907208] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:11.020 [2024-11-26 23:40:58.907308] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:07:11.020 [2024-11-26 23:40:58.907414] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:07:11.020 [2024-11-26 23:40:58.907459] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:07:11.020 [2024-11-26 23:40:58.907560] bdev_raid.c:3685:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev pt2 (4) greater than existing raid bdev raid_bdev1 (2) 00:07:11.020 [2024-11-26 23:40:58.907582] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:07:11.020 [2024-11-26 23:40:58.907602] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002000 name raid_bdev1, state configuring 00:07:11.020 [2024-11-26 23:40:58.907632] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:07:11.020 [2024-11-26 23:40:58.907701] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000002380 00:07:11.020 [2024-11-26 23:40:58.907712] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:07:11.020 [2024-11-26 23:40:58.907949] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:07:11.020 [2024-11-26 23:40:58.908060] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000002380 00:07:11.020 [2024-11-26 23:40:58.908068] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000002380 00:07:11.020 [2024-11-26 23:40:58.908180] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:11.020 pt1 00:07:11.020 23:40:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:11.020 23:40:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@542 -- # '[' 2 -gt 2 ']' 00:07:11.020 23:40:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@554 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:07:11.020 23:40:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:11.020 23:40:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:11.020 23:40:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:07:11.020 23:40:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:07:11.020 23:40:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:07:11.020 23:40:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:11.020 23:40:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:11.020 23:40:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:11.020 23:40:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:11.020 23:40:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:11.020 23:40:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:11.020 23:40:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:11.020 23:40:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:11.020 23:40:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:11.020 23:40:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:11.020 "name": "raid_bdev1", 00:07:11.020 "uuid": "d8eddafd-d63f-438d-9223-8b4e73930d34", 00:07:11.020 "strip_size_kb": 0, 00:07:11.020 "state": "online", 00:07:11.020 "raid_level": "raid1", 00:07:11.020 "superblock": true, 00:07:11.020 "num_base_bdevs": 2, 00:07:11.020 "num_base_bdevs_discovered": 1, 00:07:11.020 "num_base_bdevs_operational": 1, 00:07:11.020 "base_bdevs_list": [ 00:07:11.020 { 00:07:11.020 "name": null, 00:07:11.020 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:11.020 "is_configured": false, 00:07:11.020 "data_offset": 2048, 00:07:11.020 "data_size": 63488 00:07:11.020 }, 00:07:11.020 { 00:07:11.020 "name": "pt2", 00:07:11.020 "uuid": "00000000-0000-0000-0000-000000000002", 00:07:11.020 "is_configured": true, 00:07:11.020 "data_offset": 2048, 00:07:11.020 "data_size": 63488 00:07:11.020 } 00:07:11.020 ] 00:07:11.020 }' 00:07:11.020 23:40:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:11.020 23:40:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:11.279 23:40:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@555 -- # rpc_cmd bdev_raid_get_bdevs online 00:07:11.279 23:40:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@555 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:07:11.279 23:40:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:11.279 23:40:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:11.279 23:40:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:11.279 23:40:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@555 -- # [[ false == \f\a\l\s\e ]] 00:07:11.279 23:40:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@558 -- # jq -r '.[] | .uuid' 00:07:11.279 23:40:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@558 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:07:11.279 23:40:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:11.279 23:40:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:11.279 [2024-11-26 23:40:59.395938] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:11.538 23:40:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:11.538 23:40:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@558 -- # '[' d8eddafd-d63f-438d-9223-8b4e73930d34 '!=' d8eddafd-d63f-438d-9223-8b4e73930d34 ']' 00:07:11.538 23:40:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 74179 00:07:11.538 23:40:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # '[' -z 74179 ']' 00:07:11.538 23:40:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@958 -- # kill -0 74179 00:07:11.538 23:40:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # uname 00:07:11.538 23:40:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:07:11.538 23:40:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 74179 00:07:11.538 23:40:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:07:11.538 23:40:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:07:11.538 23:40:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 74179' 00:07:11.538 killing process with pid 74179 00:07:11.538 23:40:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@973 -- # kill 74179 00:07:11.538 [2024-11-26 23:40:59.464387] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:11.538 [2024-11-26 23:40:59.464518] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:11.538 23:40:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@978 -- # wait 74179 00:07:11.538 [2024-11-26 23:40:59.464599] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:11.538 [2024-11-26 23:40:59.464615] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002380 name raid_bdev1, state offline 00:07:11.538 [2024-11-26 23:40:59.487814] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:11.798 23:40:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:07:11.798 00:07:11.798 real 0m4.716s 00:07:11.798 user 0m7.758s 00:07:11.798 sys 0m0.906s 00:07:11.798 23:40:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:07:11.798 ************************************ 00:07:11.798 END TEST raid_superblock_test 00:07:11.798 ************************************ 00:07:11.798 23:40:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:11.798 23:40:59 bdev_raid -- bdev/bdev_raid.sh@971 -- # run_test raid_read_error_test raid_io_error_test raid1 2 read 00:07:11.798 23:40:59 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:07:11.798 23:40:59 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:11.798 23:40:59 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:11.798 ************************************ 00:07:11.798 START TEST raid_read_error_test 00:07:11.798 ************************************ 00:07:11.798 23:40:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test raid1 2 read 00:07:11.798 23:40:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid1 00:07:11.798 23:40:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=2 00:07:11.799 23:40:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=read 00:07:11.799 23:40:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:07:11.799 23:40:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:07:11.799 23:40:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:07:11.799 23:40:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:07:11.799 23:40:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:07:11.799 23:40:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:07:11.799 23:40:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:07:11.799 23:40:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:07:11.799 23:40:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:07:11.799 23:40:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:07:11.799 23:40:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:07:11.799 23:40:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:07:11.799 23:40:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:07:11.799 23:40:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:07:11.799 23:40:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:07:11.799 23:40:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid1 '!=' raid1 ']' 00:07:11.799 23:40:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@804 -- # strip_size=0 00:07:11.799 23:40:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:07:11.799 23:40:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.Qg5abCcYTc 00:07:11.799 23:40:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=74487 00:07:11.799 23:40:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 74487 00:07:11.799 23:40:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:07:11.799 23:40:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@835 -- # '[' -z 74487 ']' 00:07:11.799 23:40:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:11.799 23:40:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:11.799 23:40:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:11.799 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:11.799 23:40:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:11.799 23:40:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:11.799 [2024-11-26 23:40:59.871032] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:07:11.799 [2024-11-26 23:40:59.871224] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid74487 ] 00:07:12.059 [2024-11-26 23:41:00.027597] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:12.059 [2024-11-26 23:41:00.053647] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:12.059 [2024-11-26 23:41:00.098150] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:12.059 [2024-11-26 23:41:00.098186] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:12.628 23:41:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:12.628 23:41:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@868 -- # return 0 00:07:12.628 23:41:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:07:12.628 23:41:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:07:12.628 23:41:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:12.628 23:41:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:12.628 BaseBdev1_malloc 00:07:12.628 23:41:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:12.628 23:41:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:07:12.628 23:41:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:12.628 23:41:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:12.628 true 00:07:12.628 23:41:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:12.628 23:41:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:07:12.628 23:41:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:12.628 23:41:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:12.628 [2024-11-26 23:41:00.714025] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:07:12.628 [2024-11-26 23:41:00.714152] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:12.628 [2024-11-26 23:41:00.714190] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006980 00:07:12.628 [2024-11-26 23:41:00.714218] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:12.628 [2024-11-26 23:41:00.716323] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:12.629 [2024-11-26 23:41:00.716421] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:07:12.629 BaseBdev1 00:07:12.629 23:41:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:12.629 23:41:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:07:12.629 23:41:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:07:12.629 23:41:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:12.629 23:41:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:12.629 BaseBdev2_malloc 00:07:12.629 23:41:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:12.629 23:41:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:07:12.629 23:41:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:12.629 23:41:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:12.629 true 00:07:12.629 23:41:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:12.629 23:41:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:07:12.629 23:41:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:12.629 23:41:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:12.629 [2024-11-26 23:41:00.754633] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:07:12.629 [2024-11-26 23:41:00.754719] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:12.629 [2024-11-26 23:41:00.754753] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:07:12.629 [2024-11-26 23:41:00.754790] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:12.629 [2024-11-26 23:41:00.756889] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:12.629 [2024-11-26 23:41:00.756961] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:07:12.889 BaseBdev2 00:07:12.889 23:41:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:12.889 23:41:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 -s 00:07:12.889 23:41:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:12.889 23:41:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:12.889 [2024-11-26 23:41:00.766651] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:12.889 [2024-11-26 23:41:00.768529] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:07:12.889 [2024-11-26 23:41:00.768753] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:07:12.889 [2024-11-26 23:41:00.768788] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:07:12.889 [2024-11-26 23:41:00.769068] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002870 00:07:12.889 [2024-11-26 23:41:00.769259] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:07:12.889 [2024-11-26 23:41:00.769306] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:07:12.889 [2024-11-26 23:41:00.769472] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:12.889 23:41:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:12.889 23:41:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:07:12.889 23:41:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:12.889 23:41:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:12.889 23:41:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:07:12.889 23:41:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:07:12.889 23:41:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:12.889 23:41:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:12.889 23:41:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:12.889 23:41:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:12.889 23:41:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:12.889 23:41:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:12.889 23:41:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:12.889 23:41:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:12.889 23:41:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:12.889 23:41:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:12.889 23:41:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:12.889 "name": "raid_bdev1", 00:07:12.889 "uuid": "07e05ad6-8483-4d56-9371-bebfc5992657", 00:07:12.889 "strip_size_kb": 0, 00:07:12.889 "state": "online", 00:07:12.889 "raid_level": "raid1", 00:07:12.889 "superblock": true, 00:07:12.889 "num_base_bdevs": 2, 00:07:12.889 "num_base_bdevs_discovered": 2, 00:07:12.889 "num_base_bdevs_operational": 2, 00:07:12.889 "base_bdevs_list": [ 00:07:12.889 { 00:07:12.889 "name": "BaseBdev1", 00:07:12.889 "uuid": "981c7482-a3df-5be4-acc7-0befda0e66a9", 00:07:12.889 "is_configured": true, 00:07:12.889 "data_offset": 2048, 00:07:12.889 "data_size": 63488 00:07:12.889 }, 00:07:12.889 { 00:07:12.889 "name": "BaseBdev2", 00:07:12.890 "uuid": "955eeecb-eb59-5b18-9485-3150929f548a", 00:07:12.890 "is_configured": true, 00:07:12.890 "data_offset": 2048, 00:07:12.890 "data_size": 63488 00:07:12.890 } 00:07:12.890 ] 00:07:12.890 }' 00:07:12.890 23:41:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:12.890 23:41:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:13.150 23:41:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:07:13.150 23:41:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:07:13.150 [2024-11-26 23:41:01.254301] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002a10 00:07:14.090 23:41:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:07:14.090 23:41:02 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:14.091 23:41:02 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:14.091 23:41:02 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:14.091 23:41:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:07:14.091 23:41:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid1 = \r\a\i\d\1 ]] 00:07:14.091 23:41:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ read = \w\r\i\t\e ]] 00:07:14.091 23:41:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=2 00:07:14.091 23:41:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:07:14.091 23:41:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:14.091 23:41:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:14.091 23:41:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:07:14.091 23:41:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:07:14.091 23:41:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:14.091 23:41:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:14.091 23:41:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:14.091 23:41:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:14.091 23:41:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:14.091 23:41:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:14.091 23:41:02 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:14.091 23:41:02 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:14.091 23:41:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:14.091 23:41:02 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:14.350 23:41:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:14.350 "name": "raid_bdev1", 00:07:14.350 "uuid": "07e05ad6-8483-4d56-9371-bebfc5992657", 00:07:14.350 "strip_size_kb": 0, 00:07:14.350 "state": "online", 00:07:14.350 "raid_level": "raid1", 00:07:14.350 "superblock": true, 00:07:14.350 "num_base_bdevs": 2, 00:07:14.350 "num_base_bdevs_discovered": 2, 00:07:14.350 "num_base_bdevs_operational": 2, 00:07:14.350 "base_bdevs_list": [ 00:07:14.350 { 00:07:14.350 "name": "BaseBdev1", 00:07:14.350 "uuid": "981c7482-a3df-5be4-acc7-0befda0e66a9", 00:07:14.350 "is_configured": true, 00:07:14.350 "data_offset": 2048, 00:07:14.350 "data_size": 63488 00:07:14.350 }, 00:07:14.350 { 00:07:14.350 "name": "BaseBdev2", 00:07:14.350 "uuid": "955eeecb-eb59-5b18-9485-3150929f548a", 00:07:14.350 "is_configured": true, 00:07:14.350 "data_offset": 2048, 00:07:14.350 "data_size": 63488 00:07:14.350 } 00:07:14.350 ] 00:07:14.350 }' 00:07:14.351 23:41:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:14.351 23:41:02 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:14.610 23:41:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:07:14.610 23:41:02 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:14.610 23:41:02 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:14.610 [2024-11-26 23:41:02.654013] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:07:14.611 [2024-11-26 23:41:02.654123] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:14.611 [2024-11-26 23:41:02.656671] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:14.611 [2024-11-26 23:41:02.656766] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:14.611 [2024-11-26 23:41:02.656870] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:14.611 [2024-11-26 23:41:02.656933] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:07:14.611 { 00:07:14.611 "results": [ 00:07:14.611 { 00:07:14.611 "job": "raid_bdev1", 00:07:14.611 "core_mask": "0x1", 00:07:14.611 "workload": "randrw", 00:07:14.611 "percentage": 50, 00:07:14.611 "status": "finished", 00:07:14.611 "queue_depth": 1, 00:07:14.611 "io_size": 131072, 00:07:14.611 "runtime": 1.400832, 00:07:14.611 "iops": 19620.48268457602, 00:07:14.611 "mibps": 2452.5603355720027, 00:07:14.611 "io_failed": 0, 00:07:14.611 "io_timeout": 0, 00:07:14.611 "avg_latency_us": 48.36864646297742, 00:07:14.611 "min_latency_us": 22.134497816593885, 00:07:14.611 "max_latency_us": 1416.6078602620087 00:07:14.611 } 00:07:14.611 ], 00:07:14.611 "core_count": 1 00:07:14.611 } 00:07:14.611 23:41:02 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:14.611 23:41:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 74487 00:07:14.611 23:41:02 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # '[' -z 74487 ']' 00:07:14.611 23:41:02 bdev_raid.raid_read_error_test -- common/autotest_common.sh@958 -- # kill -0 74487 00:07:14.611 23:41:02 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # uname 00:07:14.611 23:41:02 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:07:14.611 23:41:02 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 74487 00:07:14.611 23:41:02 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:07:14.611 23:41:02 bdev_raid.raid_read_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:07:14.611 23:41:02 bdev_raid.raid_read_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 74487' 00:07:14.611 killing process with pid 74487 00:07:14.611 23:41:02 bdev_raid.raid_read_error_test -- common/autotest_common.sh@973 -- # kill 74487 00:07:14.611 [2024-11-26 23:41:02.693731] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:14.611 23:41:02 bdev_raid.raid_read_error_test -- common/autotest_common.sh@978 -- # wait 74487 00:07:14.611 [2024-11-26 23:41:02.709938] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:14.883 23:41:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:07:14.883 23:41:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.Qg5abCcYTc 00:07:14.883 23:41:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:07:14.883 23:41:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.00 00:07:14.883 23:41:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid1 00:07:14.883 ************************************ 00:07:14.883 END TEST raid_read_error_test 00:07:14.883 ************************************ 00:07:14.883 23:41:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:07:14.883 23:41:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@199 -- # return 0 00:07:14.883 23:41:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@847 -- # [[ 0.00 = \0\.\0\0 ]] 00:07:14.883 00:07:14.883 real 0m3.151s 00:07:14.883 user 0m4.022s 00:07:14.883 sys 0m0.466s 00:07:14.883 23:41:02 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:07:14.883 23:41:02 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:14.883 23:41:02 bdev_raid -- bdev/bdev_raid.sh@972 -- # run_test raid_write_error_test raid_io_error_test raid1 2 write 00:07:14.883 23:41:02 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:07:14.883 23:41:02 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:14.883 23:41:02 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:14.883 ************************************ 00:07:14.883 START TEST raid_write_error_test 00:07:14.883 ************************************ 00:07:14.883 23:41:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test raid1 2 write 00:07:14.883 23:41:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid1 00:07:14.883 23:41:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=2 00:07:14.883 23:41:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=write 00:07:14.883 23:41:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:07:14.883 23:41:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:07:14.883 23:41:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:07:14.883 23:41:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:07:14.883 23:41:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:07:14.883 23:41:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:07:14.883 23:41:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:07:14.883 23:41:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:07:14.883 23:41:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:07:14.883 23:41:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:07:14.883 23:41:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:07:14.883 23:41:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:07:14.883 23:41:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:07:14.883 23:41:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:07:14.883 23:41:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:07:14.883 23:41:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid1 '!=' raid1 ']' 00:07:14.883 23:41:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@804 -- # strip_size=0 00:07:14.883 23:41:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:07:14.883 23:41:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.ojEeJ8muvh 00:07:14.883 23:41:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=74616 00:07:14.883 23:41:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 74616 00:07:14.883 23:41:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:07:14.883 23:41:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@835 -- # '[' -z 74616 ']' 00:07:14.883 23:41:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:14.883 23:41:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:14.883 23:41:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:14.883 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:14.883 23:41:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:14.883 23:41:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:15.145 [2024-11-26 23:41:03.079751] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:07:15.146 [2024-11-26 23:41:03.079891] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid74616 ] 00:07:15.146 [2024-11-26 23:41:03.233561] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:15.146 [2024-11-26 23:41:03.258535] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:15.405 [2024-11-26 23:41:03.301520] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:15.405 [2024-11-26 23:41:03.301633] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:15.977 23:41:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:15.977 23:41:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@868 -- # return 0 00:07:15.977 23:41:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:07:15.977 23:41:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:07:15.977 23:41:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:15.977 23:41:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:15.977 BaseBdev1_malloc 00:07:15.977 23:41:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:15.977 23:41:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:07:15.977 23:41:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:15.977 23:41:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:15.977 true 00:07:15.977 23:41:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:15.977 23:41:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:07:15.977 23:41:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:15.977 23:41:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:15.977 [2024-11-26 23:41:03.937501] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:07:15.977 [2024-11-26 23:41:03.937556] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:15.977 [2024-11-26 23:41:03.937598] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006980 00:07:15.977 [2024-11-26 23:41:03.937607] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:15.977 [2024-11-26 23:41:03.939699] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:15.977 [2024-11-26 23:41:03.939793] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:07:15.977 BaseBdev1 00:07:15.977 23:41:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:15.977 23:41:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:07:15.977 23:41:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:07:15.977 23:41:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:15.977 23:41:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:15.977 BaseBdev2_malloc 00:07:15.977 23:41:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:15.977 23:41:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:07:15.977 23:41:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:15.977 23:41:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:15.977 true 00:07:15.977 23:41:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:15.977 23:41:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:07:15.977 23:41:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:15.977 23:41:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:15.977 [2024-11-26 23:41:03.978065] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:07:15.977 [2024-11-26 23:41:03.978188] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:15.977 [2024-11-26 23:41:03.978211] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:07:15.977 [2024-11-26 23:41:03.978228] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:15.977 [2024-11-26 23:41:03.980313] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:15.977 [2024-11-26 23:41:03.980372] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:07:15.977 BaseBdev2 00:07:15.977 23:41:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:15.977 23:41:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 -s 00:07:15.977 23:41:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:15.977 23:41:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:15.977 [2024-11-26 23:41:03.990085] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:15.977 [2024-11-26 23:41:03.991871] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:07:15.977 [2024-11-26 23:41:03.992043] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:07:15.977 [2024-11-26 23:41:03.992056] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:07:15.977 [2024-11-26 23:41:03.992299] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002870 00:07:15.977 [2024-11-26 23:41:03.992474] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:07:15.977 [2024-11-26 23:41:03.992488] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:07:15.977 [2024-11-26 23:41:03.992625] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:15.977 23:41:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:15.977 23:41:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:07:15.977 23:41:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:15.977 23:41:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:15.977 23:41:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:07:15.977 23:41:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:07:15.977 23:41:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:15.977 23:41:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:15.977 23:41:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:15.977 23:41:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:15.977 23:41:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:15.977 23:41:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:15.977 23:41:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:15.977 23:41:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:15.977 23:41:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:15.977 23:41:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:15.977 23:41:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:15.977 "name": "raid_bdev1", 00:07:15.977 "uuid": "9a738da4-66dc-4e5c-9a64-50198cea6683", 00:07:15.977 "strip_size_kb": 0, 00:07:15.977 "state": "online", 00:07:15.977 "raid_level": "raid1", 00:07:15.977 "superblock": true, 00:07:15.977 "num_base_bdevs": 2, 00:07:15.977 "num_base_bdevs_discovered": 2, 00:07:15.977 "num_base_bdevs_operational": 2, 00:07:15.977 "base_bdevs_list": [ 00:07:15.977 { 00:07:15.977 "name": "BaseBdev1", 00:07:15.977 "uuid": "9742a43b-af76-5bc4-b9fd-f04047644596", 00:07:15.977 "is_configured": true, 00:07:15.977 "data_offset": 2048, 00:07:15.977 "data_size": 63488 00:07:15.977 }, 00:07:15.977 { 00:07:15.977 "name": "BaseBdev2", 00:07:15.977 "uuid": "6ea8f04f-1669-5b78-86bc-3f12926c3f6a", 00:07:15.977 "is_configured": true, 00:07:15.977 "data_offset": 2048, 00:07:15.977 "data_size": 63488 00:07:15.977 } 00:07:15.977 ] 00:07:15.977 }' 00:07:15.977 23:41:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:15.977 23:41:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:16.548 23:41:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:07:16.548 23:41:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:07:16.548 [2024-11-26 23:41:04.489635] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002a10 00:07:17.486 23:41:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:07:17.486 23:41:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:17.486 23:41:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:17.486 [2024-11-26 23:41:05.413751] bdev_raid.c:2276:_raid_bdev_fail_base_bdev: *NOTICE*: Failing base bdev in slot 0 ('BaseBdev1') of raid bdev 'raid_bdev1' 00:07:17.486 [2024-11-26 23:41:05.413873] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:07:17.486 [2024-11-26 23:41:05.414134] bdev_raid.c:1974:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 0 raid_ch: 0x60d000002a10 00:07:17.486 23:41:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:17.486 23:41:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:07:17.486 23:41:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid1 = \r\a\i\d\1 ]] 00:07:17.486 23:41:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ write = \w\r\i\t\e ]] 00:07:17.486 23:41:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@833 -- # expected_num_base_bdevs=1 00:07:17.486 23:41:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:07:17.486 23:41:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:17.486 23:41:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:17.486 23:41:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:07:17.486 23:41:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:07:17.486 23:41:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:07:17.486 23:41:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:17.486 23:41:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:17.486 23:41:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:17.486 23:41:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:17.486 23:41:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:17.486 23:41:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:17.486 23:41:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:17.486 23:41:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:17.486 23:41:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:17.486 23:41:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:17.486 "name": "raid_bdev1", 00:07:17.486 "uuid": "9a738da4-66dc-4e5c-9a64-50198cea6683", 00:07:17.486 "strip_size_kb": 0, 00:07:17.486 "state": "online", 00:07:17.486 "raid_level": "raid1", 00:07:17.486 "superblock": true, 00:07:17.486 "num_base_bdevs": 2, 00:07:17.486 "num_base_bdevs_discovered": 1, 00:07:17.486 "num_base_bdevs_operational": 1, 00:07:17.486 "base_bdevs_list": [ 00:07:17.486 { 00:07:17.486 "name": null, 00:07:17.486 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:17.486 "is_configured": false, 00:07:17.486 "data_offset": 0, 00:07:17.486 "data_size": 63488 00:07:17.486 }, 00:07:17.486 { 00:07:17.486 "name": "BaseBdev2", 00:07:17.486 "uuid": "6ea8f04f-1669-5b78-86bc-3f12926c3f6a", 00:07:17.486 "is_configured": true, 00:07:17.486 "data_offset": 2048, 00:07:17.486 "data_size": 63488 00:07:17.486 } 00:07:17.486 ] 00:07:17.486 }' 00:07:17.486 23:41:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:17.486 23:41:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:17.747 23:41:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:07:17.747 23:41:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:17.747 23:41:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:17.747 [2024-11-26 23:41:05.854710] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:07:17.747 [2024-11-26 23:41:05.854745] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:17.747 [2024-11-26 23:41:05.857235] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:17.747 [2024-11-26 23:41:05.857288] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:17.747 [2024-11-26 23:41:05.857339] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:17.747 [2024-11-26 23:41:05.857350] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:07:17.747 23:41:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:17.747 { 00:07:17.747 "results": [ 00:07:17.747 { 00:07:17.747 "job": "raid_bdev1", 00:07:17.747 "core_mask": "0x1", 00:07:17.747 "workload": "randrw", 00:07:17.747 "percentage": 50, 00:07:17.747 "status": "finished", 00:07:17.747 "queue_depth": 1, 00:07:17.747 "io_size": 131072, 00:07:17.747 "runtime": 1.365885, 00:07:17.747 "iops": 22551.67894808128, 00:07:17.747 "mibps": 2818.95986851016, 00:07:17.747 "io_failed": 0, 00:07:17.747 "io_timeout": 0, 00:07:17.747 "avg_latency_us": 41.70050583458453, 00:07:17.747 "min_latency_us": 21.687336244541484, 00:07:17.747 "max_latency_us": 1445.2262008733624 00:07:17.747 } 00:07:17.747 ], 00:07:17.747 "core_count": 1 00:07:17.747 } 00:07:17.747 23:41:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 74616 00:07:17.747 23:41:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # '[' -z 74616 ']' 00:07:17.747 23:41:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@958 -- # kill -0 74616 00:07:17.747 23:41:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # uname 00:07:17.747 23:41:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:07:17.747 23:41:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 74616 00:07:18.007 23:41:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:07:18.007 23:41:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:07:18.007 23:41:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 74616' 00:07:18.007 killing process with pid 74616 00:07:18.007 23:41:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@973 -- # kill 74616 00:07:18.007 [2024-11-26 23:41:05.906249] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:18.007 23:41:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@978 -- # wait 74616 00:07:18.007 [2024-11-26 23:41:05.921639] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:18.007 23:41:06 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:07:18.007 23:41:06 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.ojEeJ8muvh 00:07:18.007 23:41:06 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:07:18.007 23:41:06 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.00 00:07:18.007 23:41:06 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid1 00:07:18.007 ************************************ 00:07:18.007 END TEST raid_write_error_test 00:07:18.007 ************************************ 00:07:18.007 23:41:06 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:07:18.007 23:41:06 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@199 -- # return 0 00:07:18.007 23:41:06 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@847 -- # [[ 0.00 = \0\.\0\0 ]] 00:07:18.007 00:07:18.007 real 0m3.144s 00:07:18.007 user 0m3.983s 00:07:18.007 sys 0m0.482s 00:07:18.007 23:41:06 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:07:18.007 23:41:06 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:18.267 23:41:06 bdev_raid -- bdev/bdev_raid.sh@966 -- # for n in {2..4} 00:07:18.267 23:41:06 bdev_raid -- bdev/bdev_raid.sh@967 -- # for level in raid0 concat raid1 00:07:18.267 23:41:06 bdev_raid -- bdev/bdev_raid.sh@968 -- # run_test raid_state_function_test raid_state_function_test raid0 3 false 00:07:18.267 23:41:06 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:07:18.267 23:41:06 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:18.267 23:41:06 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:18.267 ************************************ 00:07:18.267 START TEST raid_state_function_test 00:07:18.267 ************************************ 00:07:18.267 23:41:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1129 -- # raid_state_function_test raid0 3 false 00:07:18.267 23:41:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=raid0 00:07:18.267 23:41:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=3 00:07:18.267 23:41:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:07:18.267 23:41:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:07:18.267 23:41:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:07:18.267 23:41:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:18.267 23:41:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:07:18.267 23:41:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:07:18.268 23:41:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:18.268 23:41:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:07:18.268 23:41:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:07:18.268 23:41:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:18.268 23:41:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:07:18.268 23:41:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:07:18.268 23:41:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:18.268 23:41:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:07:18.268 23:41:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:07:18.268 23:41:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:07:18.268 23:41:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:07:18.268 23:41:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:07:18.268 23:41:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:07:18.268 23:41:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' raid0 '!=' raid1 ']' 00:07:18.268 23:41:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:07:18.268 23:41:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:07:18.268 23:41:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:07:18.268 23:41:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:07:18.268 23:41:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=74748 00:07:18.268 23:41:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:07:18.268 Process raid pid: 74748 00:07:18.268 23:41:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 74748' 00:07:18.268 23:41:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 74748 00:07:18.268 23:41:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@835 -- # '[' -z 74748 ']' 00:07:18.268 23:41:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:18.268 23:41:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:18.268 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:18.268 23:41:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:18.268 23:41:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:18.268 23:41:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:18.268 [2024-11-26 23:41:06.298233] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:07:18.268 [2024-11-26 23:41:06.298496] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:07:18.528 [2024-11-26 23:41:06.451727] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:18.528 [2024-11-26 23:41:06.476596] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:18.528 [2024-11-26 23:41:06.519445] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:18.528 [2024-11-26 23:41:06.519500] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:19.100 23:41:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:19.100 23:41:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@868 -- # return 0 00:07:19.100 23:41:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:07:19.100 23:41:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:19.100 23:41:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:19.100 [2024-11-26 23:41:07.118430] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:07:19.100 [2024-11-26 23:41:07.118490] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:07:19.100 [2024-11-26 23:41:07.118506] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:19.100 [2024-11-26 23:41:07.118517] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:19.100 [2024-11-26 23:41:07.118523] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:07:19.100 [2024-11-26 23:41:07.118533] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:07:19.100 23:41:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:19.100 23:41:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:07:19.100 23:41:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:19.100 23:41:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:19.100 23:41:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:19.100 23:41:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:19.100 23:41:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:07:19.100 23:41:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:19.100 23:41:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:19.100 23:41:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:19.100 23:41:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:19.101 23:41:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:19.101 23:41:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:19.101 23:41:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:19.101 23:41:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:19.101 23:41:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:19.101 23:41:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:19.101 "name": "Existed_Raid", 00:07:19.101 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:19.101 "strip_size_kb": 64, 00:07:19.101 "state": "configuring", 00:07:19.101 "raid_level": "raid0", 00:07:19.101 "superblock": false, 00:07:19.101 "num_base_bdevs": 3, 00:07:19.101 "num_base_bdevs_discovered": 0, 00:07:19.101 "num_base_bdevs_operational": 3, 00:07:19.101 "base_bdevs_list": [ 00:07:19.101 { 00:07:19.101 "name": "BaseBdev1", 00:07:19.101 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:19.101 "is_configured": false, 00:07:19.101 "data_offset": 0, 00:07:19.101 "data_size": 0 00:07:19.101 }, 00:07:19.101 { 00:07:19.101 "name": "BaseBdev2", 00:07:19.101 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:19.101 "is_configured": false, 00:07:19.101 "data_offset": 0, 00:07:19.101 "data_size": 0 00:07:19.101 }, 00:07:19.101 { 00:07:19.101 "name": "BaseBdev3", 00:07:19.101 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:19.101 "is_configured": false, 00:07:19.101 "data_offset": 0, 00:07:19.101 "data_size": 0 00:07:19.101 } 00:07:19.101 ] 00:07:19.101 }' 00:07:19.101 23:41:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:19.101 23:41:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:19.671 23:41:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:07:19.671 23:41:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:19.671 23:41:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:19.671 [2024-11-26 23:41:07.565588] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:07:19.671 [2024-11-26 23:41:07.565670] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:07:19.671 23:41:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:19.671 23:41:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:07:19.671 23:41:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:19.671 23:41:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:19.671 [2024-11-26 23:41:07.577571] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:07:19.671 [2024-11-26 23:41:07.577655] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:07:19.671 [2024-11-26 23:41:07.577697] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:19.671 [2024-11-26 23:41:07.577805] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:19.671 [2024-11-26 23:41:07.577853] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:07:19.671 [2024-11-26 23:41:07.577876] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:07:19.671 23:41:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:19.671 23:41:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:07:19.671 23:41:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:19.671 23:41:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:19.671 [2024-11-26 23:41:07.598395] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:19.671 BaseBdev1 00:07:19.671 23:41:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:19.671 23:41:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:07:19.671 23:41:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:07:19.671 23:41:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:07:19.671 23:41:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:07:19.671 23:41:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:07:19.671 23:41:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:07:19.671 23:41:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:07:19.671 23:41:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:19.671 23:41:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:19.671 23:41:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:19.671 23:41:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:07:19.671 23:41:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:19.671 23:41:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:19.671 [ 00:07:19.671 { 00:07:19.671 "name": "BaseBdev1", 00:07:19.671 "aliases": [ 00:07:19.671 "14b50235-66a4-46aa-b287-862a1005a7a1" 00:07:19.671 ], 00:07:19.671 "product_name": "Malloc disk", 00:07:19.671 "block_size": 512, 00:07:19.671 "num_blocks": 65536, 00:07:19.671 "uuid": "14b50235-66a4-46aa-b287-862a1005a7a1", 00:07:19.671 "assigned_rate_limits": { 00:07:19.671 "rw_ios_per_sec": 0, 00:07:19.671 "rw_mbytes_per_sec": 0, 00:07:19.671 "r_mbytes_per_sec": 0, 00:07:19.671 "w_mbytes_per_sec": 0 00:07:19.671 }, 00:07:19.671 "claimed": true, 00:07:19.671 "claim_type": "exclusive_write", 00:07:19.671 "zoned": false, 00:07:19.671 "supported_io_types": { 00:07:19.671 "read": true, 00:07:19.671 "write": true, 00:07:19.671 "unmap": true, 00:07:19.671 "flush": true, 00:07:19.671 "reset": true, 00:07:19.671 "nvme_admin": false, 00:07:19.671 "nvme_io": false, 00:07:19.671 "nvme_io_md": false, 00:07:19.671 "write_zeroes": true, 00:07:19.671 "zcopy": true, 00:07:19.671 "get_zone_info": false, 00:07:19.671 "zone_management": false, 00:07:19.671 "zone_append": false, 00:07:19.671 "compare": false, 00:07:19.671 "compare_and_write": false, 00:07:19.671 "abort": true, 00:07:19.671 "seek_hole": false, 00:07:19.671 "seek_data": false, 00:07:19.671 "copy": true, 00:07:19.671 "nvme_iov_md": false 00:07:19.671 }, 00:07:19.671 "memory_domains": [ 00:07:19.671 { 00:07:19.671 "dma_device_id": "system", 00:07:19.671 "dma_device_type": 1 00:07:19.671 }, 00:07:19.671 { 00:07:19.671 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:19.671 "dma_device_type": 2 00:07:19.671 } 00:07:19.671 ], 00:07:19.671 "driver_specific": {} 00:07:19.671 } 00:07:19.671 ] 00:07:19.671 23:41:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:19.671 23:41:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:07:19.671 23:41:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:07:19.671 23:41:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:19.671 23:41:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:19.671 23:41:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:19.671 23:41:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:19.671 23:41:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:07:19.671 23:41:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:19.671 23:41:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:19.671 23:41:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:19.671 23:41:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:19.671 23:41:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:19.671 23:41:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:19.671 23:41:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:19.671 23:41:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:19.671 23:41:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:19.671 23:41:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:19.671 "name": "Existed_Raid", 00:07:19.671 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:19.671 "strip_size_kb": 64, 00:07:19.671 "state": "configuring", 00:07:19.671 "raid_level": "raid0", 00:07:19.671 "superblock": false, 00:07:19.671 "num_base_bdevs": 3, 00:07:19.671 "num_base_bdevs_discovered": 1, 00:07:19.671 "num_base_bdevs_operational": 3, 00:07:19.671 "base_bdevs_list": [ 00:07:19.671 { 00:07:19.671 "name": "BaseBdev1", 00:07:19.671 "uuid": "14b50235-66a4-46aa-b287-862a1005a7a1", 00:07:19.671 "is_configured": true, 00:07:19.671 "data_offset": 0, 00:07:19.672 "data_size": 65536 00:07:19.672 }, 00:07:19.672 { 00:07:19.672 "name": "BaseBdev2", 00:07:19.672 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:19.672 "is_configured": false, 00:07:19.672 "data_offset": 0, 00:07:19.672 "data_size": 0 00:07:19.672 }, 00:07:19.672 { 00:07:19.672 "name": "BaseBdev3", 00:07:19.672 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:19.672 "is_configured": false, 00:07:19.672 "data_offset": 0, 00:07:19.672 "data_size": 0 00:07:19.672 } 00:07:19.672 ] 00:07:19.672 }' 00:07:19.672 23:41:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:19.672 23:41:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:19.932 23:41:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:07:19.932 23:41:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:19.932 23:41:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:19.932 [2024-11-26 23:41:08.025709] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:07:19.932 [2024-11-26 23:41:08.025759] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:07:19.932 23:41:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:19.932 23:41:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:07:19.932 23:41:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:19.932 23:41:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:19.932 [2024-11-26 23:41:08.033723] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:19.932 [2024-11-26 23:41:08.035613] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:19.932 [2024-11-26 23:41:08.035690] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:19.932 [2024-11-26 23:41:08.035735] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:07:19.932 [2024-11-26 23:41:08.035760] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:07:19.932 23:41:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:19.932 23:41:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:07:19.932 23:41:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:07:19.932 23:41:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:07:19.932 23:41:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:19.932 23:41:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:19.932 23:41:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:19.932 23:41:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:19.932 23:41:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:07:19.932 23:41:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:19.932 23:41:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:19.932 23:41:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:19.932 23:41:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:19.932 23:41:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:19.932 23:41:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:19.932 23:41:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:19.932 23:41:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:20.191 23:41:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:20.191 23:41:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:20.191 "name": "Existed_Raid", 00:07:20.191 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:20.191 "strip_size_kb": 64, 00:07:20.191 "state": "configuring", 00:07:20.191 "raid_level": "raid0", 00:07:20.191 "superblock": false, 00:07:20.191 "num_base_bdevs": 3, 00:07:20.191 "num_base_bdevs_discovered": 1, 00:07:20.191 "num_base_bdevs_operational": 3, 00:07:20.191 "base_bdevs_list": [ 00:07:20.191 { 00:07:20.191 "name": "BaseBdev1", 00:07:20.191 "uuid": "14b50235-66a4-46aa-b287-862a1005a7a1", 00:07:20.191 "is_configured": true, 00:07:20.191 "data_offset": 0, 00:07:20.191 "data_size": 65536 00:07:20.191 }, 00:07:20.191 { 00:07:20.191 "name": "BaseBdev2", 00:07:20.191 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:20.191 "is_configured": false, 00:07:20.191 "data_offset": 0, 00:07:20.191 "data_size": 0 00:07:20.191 }, 00:07:20.191 { 00:07:20.191 "name": "BaseBdev3", 00:07:20.191 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:20.191 "is_configured": false, 00:07:20.191 "data_offset": 0, 00:07:20.191 "data_size": 0 00:07:20.191 } 00:07:20.191 ] 00:07:20.191 }' 00:07:20.191 23:41:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:20.191 23:41:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:20.451 23:41:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:07:20.451 23:41:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:20.452 23:41:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:20.452 [2024-11-26 23:41:08.484137] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:07:20.452 BaseBdev2 00:07:20.452 23:41:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:20.452 23:41:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:07:20.452 23:41:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:07:20.452 23:41:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:07:20.452 23:41:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:07:20.452 23:41:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:07:20.452 23:41:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:07:20.452 23:41:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:07:20.452 23:41:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:20.452 23:41:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:20.452 23:41:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:20.452 23:41:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:07:20.452 23:41:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:20.452 23:41:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:20.452 [ 00:07:20.452 { 00:07:20.452 "name": "BaseBdev2", 00:07:20.452 "aliases": [ 00:07:20.452 "2d12925b-bd41-4d54-a462-2adb2a60dc24" 00:07:20.452 ], 00:07:20.452 "product_name": "Malloc disk", 00:07:20.452 "block_size": 512, 00:07:20.452 "num_blocks": 65536, 00:07:20.452 "uuid": "2d12925b-bd41-4d54-a462-2adb2a60dc24", 00:07:20.452 "assigned_rate_limits": { 00:07:20.452 "rw_ios_per_sec": 0, 00:07:20.452 "rw_mbytes_per_sec": 0, 00:07:20.452 "r_mbytes_per_sec": 0, 00:07:20.452 "w_mbytes_per_sec": 0 00:07:20.452 }, 00:07:20.452 "claimed": true, 00:07:20.452 "claim_type": "exclusive_write", 00:07:20.452 "zoned": false, 00:07:20.452 "supported_io_types": { 00:07:20.452 "read": true, 00:07:20.452 "write": true, 00:07:20.452 "unmap": true, 00:07:20.452 "flush": true, 00:07:20.452 "reset": true, 00:07:20.452 "nvme_admin": false, 00:07:20.452 "nvme_io": false, 00:07:20.452 "nvme_io_md": false, 00:07:20.452 "write_zeroes": true, 00:07:20.452 "zcopy": true, 00:07:20.452 "get_zone_info": false, 00:07:20.452 "zone_management": false, 00:07:20.452 "zone_append": false, 00:07:20.452 "compare": false, 00:07:20.452 "compare_and_write": false, 00:07:20.452 "abort": true, 00:07:20.452 "seek_hole": false, 00:07:20.452 "seek_data": false, 00:07:20.452 "copy": true, 00:07:20.452 "nvme_iov_md": false 00:07:20.452 }, 00:07:20.452 "memory_domains": [ 00:07:20.452 { 00:07:20.452 "dma_device_id": "system", 00:07:20.452 "dma_device_type": 1 00:07:20.452 }, 00:07:20.452 { 00:07:20.452 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:20.452 "dma_device_type": 2 00:07:20.452 } 00:07:20.452 ], 00:07:20.452 "driver_specific": {} 00:07:20.452 } 00:07:20.452 ] 00:07:20.452 23:41:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:20.452 23:41:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:07:20.452 23:41:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:07:20.452 23:41:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:07:20.452 23:41:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:07:20.452 23:41:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:20.452 23:41:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:20.452 23:41:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:20.452 23:41:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:20.452 23:41:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:07:20.452 23:41:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:20.452 23:41:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:20.452 23:41:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:20.452 23:41:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:20.452 23:41:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:20.452 23:41:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:20.452 23:41:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:20.452 23:41:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:20.452 23:41:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:20.452 23:41:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:20.452 "name": "Existed_Raid", 00:07:20.452 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:20.452 "strip_size_kb": 64, 00:07:20.452 "state": "configuring", 00:07:20.452 "raid_level": "raid0", 00:07:20.452 "superblock": false, 00:07:20.452 "num_base_bdevs": 3, 00:07:20.452 "num_base_bdevs_discovered": 2, 00:07:20.452 "num_base_bdevs_operational": 3, 00:07:20.452 "base_bdevs_list": [ 00:07:20.452 { 00:07:20.452 "name": "BaseBdev1", 00:07:20.452 "uuid": "14b50235-66a4-46aa-b287-862a1005a7a1", 00:07:20.452 "is_configured": true, 00:07:20.452 "data_offset": 0, 00:07:20.452 "data_size": 65536 00:07:20.452 }, 00:07:20.452 { 00:07:20.452 "name": "BaseBdev2", 00:07:20.452 "uuid": "2d12925b-bd41-4d54-a462-2adb2a60dc24", 00:07:20.452 "is_configured": true, 00:07:20.452 "data_offset": 0, 00:07:20.452 "data_size": 65536 00:07:20.452 }, 00:07:20.452 { 00:07:20.452 "name": "BaseBdev3", 00:07:20.452 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:20.452 "is_configured": false, 00:07:20.452 "data_offset": 0, 00:07:20.452 "data_size": 0 00:07:20.452 } 00:07:20.452 ] 00:07:20.452 }' 00:07:20.452 23:41:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:20.452 23:41:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:21.023 23:41:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:07:21.023 23:41:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:21.023 23:41:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:21.023 [2024-11-26 23:41:08.981911] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:07:21.023 [2024-11-26 23:41:08.982016] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:07:21.023 [2024-11-26 23:41:08.982055] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 196608, blocklen 512 00:07:21.023 [2024-11-26 23:41:08.983108] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:07:21.023 [2024-11-26 23:41:08.983658] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:07:21.023 [2024-11-26 23:41:08.983719] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:07:21.023 [2024-11-26 23:41:08.984383] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:21.023 BaseBdev3 00:07:21.023 23:41:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:21.023 23:41:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:07:21.023 23:41:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:07:21.023 23:41:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:07:21.023 23:41:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:07:21.023 23:41:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:07:21.023 23:41:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:07:21.023 23:41:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:07:21.023 23:41:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:21.023 23:41:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:21.023 23:41:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:21.023 23:41:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:07:21.023 23:41:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:21.023 23:41:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:21.023 [ 00:07:21.023 { 00:07:21.023 "name": "BaseBdev3", 00:07:21.023 "aliases": [ 00:07:21.023 "7d428d3e-075a-4212-b243-03310f029293" 00:07:21.023 ], 00:07:21.023 "product_name": "Malloc disk", 00:07:21.023 "block_size": 512, 00:07:21.023 "num_blocks": 65536, 00:07:21.023 "uuid": "7d428d3e-075a-4212-b243-03310f029293", 00:07:21.023 "assigned_rate_limits": { 00:07:21.023 "rw_ios_per_sec": 0, 00:07:21.023 "rw_mbytes_per_sec": 0, 00:07:21.023 "r_mbytes_per_sec": 0, 00:07:21.023 "w_mbytes_per_sec": 0 00:07:21.023 }, 00:07:21.023 "claimed": true, 00:07:21.023 "claim_type": "exclusive_write", 00:07:21.023 "zoned": false, 00:07:21.023 "supported_io_types": { 00:07:21.023 "read": true, 00:07:21.023 "write": true, 00:07:21.023 "unmap": true, 00:07:21.023 "flush": true, 00:07:21.023 "reset": true, 00:07:21.023 "nvme_admin": false, 00:07:21.023 "nvme_io": false, 00:07:21.023 "nvme_io_md": false, 00:07:21.023 "write_zeroes": true, 00:07:21.023 "zcopy": true, 00:07:21.023 "get_zone_info": false, 00:07:21.023 "zone_management": false, 00:07:21.023 "zone_append": false, 00:07:21.023 "compare": false, 00:07:21.023 "compare_and_write": false, 00:07:21.023 "abort": true, 00:07:21.023 "seek_hole": false, 00:07:21.023 "seek_data": false, 00:07:21.023 "copy": true, 00:07:21.023 "nvme_iov_md": false 00:07:21.023 }, 00:07:21.023 "memory_domains": [ 00:07:21.023 { 00:07:21.023 "dma_device_id": "system", 00:07:21.023 "dma_device_type": 1 00:07:21.023 }, 00:07:21.024 { 00:07:21.024 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:21.024 "dma_device_type": 2 00:07:21.024 } 00:07:21.024 ], 00:07:21.024 "driver_specific": {} 00:07:21.024 } 00:07:21.024 ] 00:07:21.024 23:41:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:21.024 23:41:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:07:21.024 23:41:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:07:21.024 23:41:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:07:21.024 23:41:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid0 64 3 00:07:21.024 23:41:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:21.024 23:41:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:21.024 23:41:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:21.024 23:41:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:21.024 23:41:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:07:21.024 23:41:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:21.024 23:41:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:21.024 23:41:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:21.024 23:41:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:21.024 23:41:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:21.024 23:41:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:21.024 23:41:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:21.024 23:41:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:21.024 23:41:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:21.024 23:41:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:21.024 "name": "Existed_Raid", 00:07:21.024 "uuid": "768c6071-1d58-43b1-99d2-86b7f9ca2d91", 00:07:21.024 "strip_size_kb": 64, 00:07:21.024 "state": "online", 00:07:21.024 "raid_level": "raid0", 00:07:21.024 "superblock": false, 00:07:21.024 "num_base_bdevs": 3, 00:07:21.024 "num_base_bdevs_discovered": 3, 00:07:21.024 "num_base_bdevs_operational": 3, 00:07:21.024 "base_bdevs_list": [ 00:07:21.024 { 00:07:21.024 "name": "BaseBdev1", 00:07:21.024 "uuid": "14b50235-66a4-46aa-b287-862a1005a7a1", 00:07:21.024 "is_configured": true, 00:07:21.024 "data_offset": 0, 00:07:21.024 "data_size": 65536 00:07:21.024 }, 00:07:21.024 { 00:07:21.024 "name": "BaseBdev2", 00:07:21.024 "uuid": "2d12925b-bd41-4d54-a462-2adb2a60dc24", 00:07:21.024 "is_configured": true, 00:07:21.024 "data_offset": 0, 00:07:21.024 "data_size": 65536 00:07:21.024 }, 00:07:21.024 { 00:07:21.024 "name": "BaseBdev3", 00:07:21.024 "uuid": "7d428d3e-075a-4212-b243-03310f029293", 00:07:21.024 "is_configured": true, 00:07:21.024 "data_offset": 0, 00:07:21.024 "data_size": 65536 00:07:21.024 } 00:07:21.024 ] 00:07:21.024 }' 00:07:21.024 23:41:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:21.024 23:41:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:21.594 23:41:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:07:21.594 23:41:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:07:21.594 23:41:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:07:21.594 23:41:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:07:21.594 23:41:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:07:21.594 23:41:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:07:21.594 23:41:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:07:21.594 23:41:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:21.594 23:41:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:21.594 23:41:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:07:21.594 [2024-11-26 23:41:09.465298] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:21.594 23:41:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:21.594 23:41:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:07:21.594 "name": "Existed_Raid", 00:07:21.594 "aliases": [ 00:07:21.594 "768c6071-1d58-43b1-99d2-86b7f9ca2d91" 00:07:21.594 ], 00:07:21.594 "product_name": "Raid Volume", 00:07:21.594 "block_size": 512, 00:07:21.594 "num_blocks": 196608, 00:07:21.594 "uuid": "768c6071-1d58-43b1-99d2-86b7f9ca2d91", 00:07:21.594 "assigned_rate_limits": { 00:07:21.594 "rw_ios_per_sec": 0, 00:07:21.594 "rw_mbytes_per_sec": 0, 00:07:21.594 "r_mbytes_per_sec": 0, 00:07:21.594 "w_mbytes_per_sec": 0 00:07:21.594 }, 00:07:21.594 "claimed": false, 00:07:21.594 "zoned": false, 00:07:21.594 "supported_io_types": { 00:07:21.594 "read": true, 00:07:21.594 "write": true, 00:07:21.594 "unmap": true, 00:07:21.594 "flush": true, 00:07:21.594 "reset": true, 00:07:21.594 "nvme_admin": false, 00:07:21.594 "nvme_io": false, 00:07:21.594 "nvme_io_md": false, 00:07:21.594 "write_zeroes": true, 00:07:21.594 "zcopy": false, 00:07:21.594 "get_zone_info": false, 00:07:21.594 "zone_management": false, 00:07:21.594 "zone_append": false, 00:07:21.594 "compare": false, 00:07:21.594 "compare_and_write": false, 00:07:21.594 "abort": false, 00:07:21.594 "seek_hole": false, 00:07:21.594 "seek_data": false, 00:07:21.594 "copy": false, 00:07:21.594 "nvme_iov_md": false 00:07:21.594 }, 00:07:21.594 "memory_domains": [ 00:07:21.594 { 00:07:21.594 "dma_device_id": "system", 00:07:21.594 "dma_device_type": 1 00:07:21.594 }, 00:07:21.594 { 00:07:21.594 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:21.594 "dma_device_type": 2 00:07:21.594 }, 00:07:21.594 { 00:07:21.594 "dma_device_id": "system", 00:07:21.594 "dma_device_type": 1 00:07:21.594 }, 00:07:21.594 { 00:07:21.594 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:21.594 "dma_device_type": 2 00:07:21.594 }, 00:07:21.594 { 00:07:21.594 "dma_device_id": "system", 00:07:21.594 "dma_device_type": 1 00:07:21.594 }, 00:07:21.594 { 00:07:21.594 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:21.594 "dma_device_type": 2 00:07:21.594 } 00:07:21.594 ], 00:07:21.594 "driver_specific": { 00:07:21.594 "raid": { 00:07:21.594 "uuid": "768c6071-1d58-43b1-99d2-86b7f9ca2d91", 00:07:21.594 "strip_size_kb": 64, 00:07:21.594 "state": "online", 00:07:21.594 "raid_level": "raid0", 00:07:21.595 "superblock": false, 00:07:21.595 "num_base_bdevs": 3, 00:07:21.595 "num_base_bdevs_discovered": 3, 00:07:21.595 "num_base_bdevs_operational": 3, 00:07:21.595 "base_bdevs_list": [ 00:07:21.595 { 00:07:21.595 "name": "BaseBdev1", 00:07:21.595 "uuid": "14b50235-66a4-46aa-b287-862a1005a7a1", 00:07:21.595 "is_configured": true, 00:07:21.595 "data_offset": 0, 00:07:21.595 "data_size": 65536 00:07:21.595 }, 00:07:21.595 { 00:07:21.595 "name": "BaseBdev2", 00:07:21.595 "uuid": "2d12925b-bd41-4d54-a462-2adb2a60dc24", 00:07:21.595 "is_configured": true, 00:07:21.595 "data_offset": 0, 00:07:21.595 "data_size": 65536 00:07:21.595 }, 00:07:21.595 { 00:07:21.595 "name": "BaseBdev3", 00:07:21.595 "uuid": "7d428d3e-075a-4212-b243-03310f029293", 00:07:21.595 "is_configured": true, 00:07:21.595 "data_offset": 0, 00:07:21.595 "data_size": 65536 00:07:21.595 } 00:07:21.595 ] 00:07:21.595 } 00:07:21.595 } 00:07:21.595 }' 00:07:21.595 23:41:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:07:21.595 23:41:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:07:21.595 BaseBdev2 00:07:21.595 BaseBdev3' 00:07:21.595 23:41:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:21.595 23:41:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:07:21.595 23:41:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:21.595 23:41:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:07:21.595 23:41:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:21.595 23:41:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:21.595 23:41:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:21.595 23:41:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:21.595 23:41:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:21.595 23:41:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:21.595 23:41:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:21.595 23:41:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:21.595 23:41:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:07:21.595 23:41:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:21.595 23:41:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:21.595 23:41:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:21.595 23:41:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:21.595 23:41:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:21.595 23:41:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:21.595 23:41:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:21.595 23:41:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:07:21.595 23:41:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:21.595 23:41:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:21.595 23:41:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:21.595 23:41:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:21.595 23:41:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:21.595 23:41:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:07:21.595 23:41:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:21.595 23:41:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:21.595 [2024-11-26 23:41:09.700649] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:07:21.595 [2024-11-26 23:41:09.700682] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:21.595 [2024-11-26 23:41:09.700735] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:21.595 23:41:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:21.595 23:41:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:07:21.595 23:41:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy raid0 00:07:21.595 23:41:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:07:21.595 23:41:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # return 1 00:07:21.595 23:41:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:07:21.595 23:41:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline raid0 64 2 00:07:21.595 23:41:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:21.595 23:41:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:07:21.595 23:41:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:21.595 23:41:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:21.595 23:41:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:21.595 23:41:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:21.595 23:41:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:21.595 23:41:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:21.595 23:41:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:21.595 23:41:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:21.595 23:41:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:21.595 23:41:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:21.595 23:41:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:21.856 23:41:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:21.856 23:41:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:21.856 "name": "Existed_Raid", 00:07:21.856 "uuid": "768c6071-1d58-43b1-99d2-86b7f9ca2d91", 00:07:21.856 "strip_size_kb": 64, 00:07:21.856 "state": "offline", 00:07:21.856 "raid_level": "raid0", 00:07:21.856 "superblock": false, 00:07:21.856 "num_base_bdevs": 3, 00:07:21.856 "num_base_bdevs_discovered": 2, 00:07:21.856 "num_base_bdevs_operational": 2, 00:07:21.856 "base_bdevs_list": [ 00:07:21.856 { 00:07:21.856 "name": null, 00:07:21.856 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:21.856 "is_configured": false, 00:07:21.856 "data_offset": 0, 00:07:21.856 "data_size": 65536 00:07:21.856 }, 00:07:21.856 { 00:07:21.856 "name": "BaseBdev2", 00:07:21.856 "uuid": "2d12925b-bd41-4d54-a462-2adb2a60dc24", 00:07:21.856 "is_configured": true, 00:07:21.856 "data_offset": 0, 00:07:21.856 "data_size": 65536 00:07:21.856 }, 00:07:21.856 { 00:07:21.856 "name": "BaseBdev3", 00:07:21.856 "uuid": "7d428d3e-075a-4212-b243-03310f029293", 00:07:21.856 "is_configured": true, 00:07:21.856 "data_offset": 0, 00:07:21.856 "data_size": 65536 00:07:21.856 } 00:07:21.856 ] 00:07:21.856 }' 00:07:21.856 23:41:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:21.856 23:41:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:22.116 23:41:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:07:22.116 23:41:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:07:22.116 23:41:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:22.116 23:41:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:22.116 23:41:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:22.116 23:41:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:07:22.116 23:41:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:22.116 23:41:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:07:22.116 23:41:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:07:22.116 23:41:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:07:22.116 23:41:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:22.116 23:41:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:22.116 [2024-11-26 23:41:10.179206] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:07:22.116 23:41:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:22.116 23:41:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:07:22.116 23:41:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:07:22.116 23:41:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:22.116 23:41:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:22.116 23:41:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:07:22.116 23:41:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:22.116 23:41:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:22.116 23:41:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:07:22.116 23:41:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:07:22.377 23:41:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:07:22.377 23:41:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:22.377 23:41:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:22.377 [2024-11-26 23:41:10.250447] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:07:22.377 [2024-11-26 23:41:10.250494] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:07:22.377 23:41:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:22.377 23:41:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:07:22.377 23:41:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:07:22.377 23:41:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:22.377 23:41:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:22.377 23:41:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:22.377 23:41:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:07:22.378 23:41:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:22.378 23:41:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:07:22.378 23:41:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:07:22.378 23:41:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 3 -gt 2 ']' 00:07:22.378 23:41:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:07:22.378 23:41:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:07:22.378 23:41:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:07:22.378 23:41:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:22.378 23:41:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:22.378 BaseBdev2 00:07:22.378 23:41:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:22.378 23:41:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:07:22.378 23:41:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:07:22.378 23:41:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:07:22.378 23:41:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:07:22.378 23:41:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:07:22.378 23:41:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:07:22.378 23:41:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:07:22.378 23:41:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:22.378 23:41:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:22.378 23:41:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:22.378 23:41:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:07:22.378 23:41:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:22.378 23:41:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:22.378 [ 00:07:22.378 { 00:07:22.378 "name": "BaseBdev2", 00:07:22.378 "aliases": [ 00:07:22.378 "9c5727a2-1496-4c68-8a1e-69260743ba76" 00:07:22.378 ], 00:07:22.378 "product_name": "Malloc disk", 00:07:22.378 "block_size": 512, 00:07:22.378 "num_blocks": 65536, 00:07:22.378 "uuid": "9c5727a2-1496-4c68-8a1e-69260743ba76", 00:07:22.378 "assigned_rate_limits": { 00:07:22.378 "rw_ios_per_sec": 0, 00:07:22.378 "rw_mbytes_per_sec": 0, 00:07:22.378 "r_mbytes_per_sec": 0, 00:07:22.378 "w_mbytes_per_sec": 0 00:07:22.378 }, 00:07:22.378 "claimed": false, 00:07:22.378 "zoned": false, 00:07:22.378 "supported_io_types": { 00:07:22.378 "read": true, 00:07:22.378 "write": true, 00:07:22.378 "unmap": true, 00:07:22.378 "flush": true, 00:07:22.378 "reset": true, 00:07:22.378 "nvme_admin": false, 00:07:22.378 "nvme_io": false, 00:07:22.378 "nvme_io_md": false, 00:07:22.378 "write_zeroes": true, 00:07:22.378 "zcopy": true, 00:07:22.378 "get_zone_info": false, 00:07:22.378 "zone_management": false, 00:07:22.378 "zone_append": false, 00:07:22.378 "compare": false, 00:07:22.378 "compare_and_write": false, 00:07:22.378 "abort": true, 00:07:22.378 "seek_hole": false, 00:07:22.378 "seek_data": false, 00:07:22.378 "copy": true, 00:07:22.378 "nvme_iov_md": false 00:07:22.378 }, 00:07:22.378 "memory_domains": [ 00:07:22.378 { 00:07:22.378 "dma_device_id": "system", 00:07:22.378 "dma_device_type": 1 00:07:22.378 }, 00:07:22.378 { 00:07:22.378 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:22.378 "dma_device_type": 2 00:07:22.378 } 00:07:22.378 ], 00:07:22.378 "driver_specific": {} 00:07:22.378 } 00:07:22.378 ] 00:07:22.378 23:41:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:22.378 23:41:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:07:22.378 23:41:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:07:22.378 23:41:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:07:22.378 23:41:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:07:22.378 23:41:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:22.378 23:41:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:22.378 BaseBdev3 00:07:22.378 23:41:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:22.378 23:41:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:07:22.378 23:41:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:07:22.378 23:41:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:07:22.378 23:41:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:07:22.378 23:41:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:07:22.378 23:41:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:07:22.378 23:41:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:07:22.378 23:41:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:22.378 23:41:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:22.378 23:41:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:22.378 23:41:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:07:22.378 23:41:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:22.378 23:41:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:22.378 [ 00:07:22.378 { 00:07:22.378 "name": "BaseBdev3", 00:07:22.378 "aliases": [ 00:07:22.378 "2c9302db-2499-4154-9a9d-48345d29d96e" 00:07:22.378 ], 00:07:22.378 "product_name": "Malloc disk", 00:07:22.378 "block_size": 512, 00:07:22.378 "num_blocks": 65536, 00:07:22.378 "uuid": "2c9302db-2499-4154-9a9d-48345d29d96e", 00:07:22.378 "assigned_rate_limits": { 00:07:22.378 "rw_ios_per_sec": 0, 00:07:22.378 "rw_mbytes_per_sec": 0, 00:07:22.378 "r_mbytes_per_sec": 0, 00:07:22.378 "w_mbytes_per_sec": 0 00:07:22.378 }, 00:07:22.378 "claimed": false, 00:07:22.378 "zoned": false, 00:07:22.378 "supported_io_types": { 00:07:22.378 "read": true, 00:07:22.378 "write": true, 00:07:22.378 "unmap": true, 00:07:22.378 "flush": true, 00:07:22.378 "reset": true, 00:07:22.378 "nvme_admin": false, 00:07:22.378 "nvme_io": false, 00:07:22.378 "nvme_io_md": false, 00:07:22.378 "write_zeroes": true, 00:07:22.378 "zcopy": true, 00:07:22.378 "get_zone_info": false, 00:07:22.378 "zone_management": false, 00:07:22.378 "zone_append": false, 00:07:22.378 "compare": false, 00:07:22.378 "compare_and_write": false, 00:07:22.378 "abort": true, 00:07:22.378 "seek_hole": false, 00:07:22.378 "seek_data": false, 00:07:22.378 "copy": true, 00:07:22.378 "nvme_iov_md": false 00:07:22.378 }, 00:07:22.378 "memory_domains": [ 00:07:22.378 { 00:07:22.378 "dma_device_id": "system", 00:07:22.378 "dma_device_type": 1 00:07:22.378 }, 00:07:22.378 { 00:07:22.378 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:22.378 "dma_device_type": 2 00:07:22.378 } 00:07:22.378 ], 00:07:22.378 "driver_specific": {} 00:07:22.378 } 00:07:22.378 ] 00:07:22.378 23:41:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:22.378 23:41:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:07:22.378 23:41:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:07:22.378 23:41:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:07:22.378 23:41:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:07:22.378 23:41:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:22.378 23:41:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:22.378 [2024-11-26 23:41:10.425695] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:07:22.378 [2024-11-26 23:41:10.425781] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:07:22.378 [2024-11-26 23:41:10.425823] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:07:22.378 [2024-11-26 23:41:10.427615] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:07:22.378 23:41:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:22.378 23:41:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:07:22.378 23:41:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:22.378 23:41:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:22.378 23:41:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:22.378 23:41:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:22.378 23:41:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:07:22.378 23:41:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:22.378 23:41:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:22.378 23:41:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:22.378 23:41:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:22.378 23:41:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:22.378 23:41:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:22.378 23:41:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:22.379 23:41:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:22.379 23:41:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:22.379 23:41:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:22.379 "name": "Existed_Raid", 00:07:22.379 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:22.379 "strip_size_kb": 64, 00:07:22.379 "state": "configuring", 00:07:22.379 "raid_level": "raid0", 00:07:22.379 "superblock": false, 00:07:22.379 "num_base_bdevs": 3, 00:07:22.379 "num_base_bdevs_discovered": 2, 00:07:22.379 "num_base_bdevs_operational": 3, 00:07:22.379 "base_bdevs_list": [ 00:07:22.379 { 00:07:22.379 "name": "BaseBdev1", 00:07:22.379 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:22.379 "is_configured": false, 00:07:22.379 "data_offset": 0, 00:07:22.379 "data_size": 0 00:07:22.379 }, 00:07:22.379 { 00:07:22.379 "name": "BaseBdev2", 00:07:22.379 "uuid": "9c5727a2-1496-4c68-8a1e-69260743ba76", 00:07:22.379 "is_configured": true, 00:07:22.379 "data_offset": 0, 00:07:22.379 "data_size": 65536 00:07:22.379 }, 00:07:22.379 { 00:07:22.379 "name": "BaseBdev3", 00:07:22.379 "uuid": "2c9302db-2499-4154-9a9d-48345d29d96e", 00:07:22.379 "is_configured": true, 00:07:22.379 "data_offset": 0, 00:07:22.379 "data_size": 65536 00:07:22.379 } 00:07:22.379 ] 00:07:22.379 }' 00:07:22.379 23:41:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:22.379 23:41:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:22.948 23:41:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:07:22.948 23:41:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:22.948 23:41:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:22.948 [2024-11-26 23:41:10.789160] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:07:22.948 23:41:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:22.948 23:41:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:07:22.948 23:41:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:22.948 23:41:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:22.948 23:41:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:22.948 23:41:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:22.948 23:41:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:07:22.948 23:41:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:22.948 23:41:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:22.948 23:41:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:22.948 23:41:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:22.948 23:41:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:22.948 23:41:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:22.948 23:41:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:22.949 23:41:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:22.949 23:41:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:22.949 23:41:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:22.949 "name": "Existed_Raid", 00:07:22.949 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:22.949 "strip_size_kb": 64, 00:07:22.949 "state": "configuring", 00:07:22.949 "raid_level": "raid0", 00:07:22.949 "superblock": false, 00:07:22.949 "num_base_bdevs": 3, 00:07:22.949 "num_base_bdevs_discovered": 1, 00:07:22.949 "num_base_bdevs_operational": 3, 00:07:22.949 "base_bdevs_list": [ 00:07:22.949 { 00:07:22.949 "name": "BaseBdev1", 00:07:22.949 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:22.949 "is_configured": false, 00:07:22.949 "data_offset": 0, 00:07:22.949 "data_size": 0 00:07:22.949 }, 00:07:22.949 { 00:07:22.949 "name": null, 00:07:22.949 "uuid": "9c5727a2-1496-4c68-8a1e-69260743ba76", 00:07:22.949 "is_configured": false, 00:07:22.949 "data_offset": 0, 00:07:22.949 "data_size": 65536 00:07:22.949 }, 00:07:22.949 { 00:07:22.949 "name": "BaseBdev3", 00:07:22.949 "uuid": "2c9302db-2499-4154-9a9d-48345d29d96e", 00:07:22.949 "is_configured": true, 00:07:22.949 "data_offset": 0, 00:07:22.949 "data_size": 65536 00:07:22.949 } 00:07:22.949 ] 00:07:22.949 }' 00:07:22.949 23:41:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:22.949 23:41:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:23.209 23:41:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:23.209 23:41:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:23.209 23:41:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:23.209 23:41:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:07:23.209 23:41:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:23.209 23:41:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:07:23.209 23:41:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:07:23.209 23:41:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:23.209 23:41:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:23.209 [2024-11-26 23:41:11.259589] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:23.209 BaseBdev1 00:07:23.209 23:41:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:23.209 23:41:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:07:23.209 23:41:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:07:23.209 23:41:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:07:23.209 23:41:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:07:23.209 23:41:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:07:23.209 23:41:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:07:23.209 23:41:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:07:23.209 23:41:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:23.209 23:41:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:23.209 23:41:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:23.209 23:41:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:07:23.209 23:41:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:23.209 23:41:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:23.209 [ 00:07:23.209 { 00:07:23.209 "name": "BaseBdev1", 00:07:23.209 "aliases": [ 00:07:23.209 "5a4fa438-802d-47e5-a418-fa61cf6baa15" 00:07:23.209 ], 00:07:23.209 "product_name": "Malloc disk", 00:07:23.209 "block_size": 512, 00:07:23.209 "num_blocks": 65536, 00:07:23.209 "uuid": "5a4fa438-802d-47e5-a418-fa61cf6baa15", 00:07:23.209 "assigned_rate_limits": { 00:07:23.209 "rw_ios_per_sec": 0, 00:07:23.209 "rw_mbytes_per_sec": 0, 00:07:23.209 "r_mbytes_per_sec": 0, 00:07:23.209 "w_mbytes_per_sec": 0 00:07:23.209 }, 00:07:23.209 "claimed": true, 00:07:23.209 "claim_type": "exclusive_write", 00:07:23.209 "zoned": false, 00:07:23.209 "supported_io_types": { 00:07:23.209 "read": true, 00:07:23.209 "write": true, 00:07:23.209 "unmap": true, 00:07:23.209 "flush": true, 00:07:23.209 "reset": true, 00:07:23.209 "nvme_admin": false, 00:07:23.209 "nvme_io": false, 00:07:23.209 "nvme_io_md": false, 00:07:23.209 "write_zeroes": true, 00:07:23.209 "zcopy": true, 00:07:23.209 "get_zone_info": false, 00:07:23.209 "zone_management": false, 00:07:23.209 "zone_append": false, 00:07:23.209 "compare": false, 00:07:23.209 "compare_and_write": false, 00:07:23.209 "abort": true, 00:07:23.209 "seek_hole": false, 00:07:23.209 "seek_data": false, 00:07:23.209 "copy": true, 00:07:23.209 "nvme_iov_md": false 00:07:23.209 }, 00:07:23.209 "memory_domains": [ 00:07:23.209 { 00:07:23.209 "dma_device_id": "system", 00:07:23.209 "dma_device_type": 1 00:07:23.209 }, 00:07:23.209 { 00:07:23.209 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:23.209 "dma_device_type": 2 00:07:23.209 } 00:07:23.209 ], 00:07:23.209 "driver_specific": {} 00:07:23.209 } 00:07:23.209 ] 00:07:23.209 23:41:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:23.209 23:41:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:07:23.209 23:41:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:07:23.209 23:41:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:23.209 23:41:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:23.209 23:41:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:23.209 23:41:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:23.209 23:41:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:07:23.209 23:41:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:23.209 23:41:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:23.209 23:41:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:23.209 23:41:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:23.209 23:41:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:23.209 23:41:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:23.209 23:41:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:23.209 23:41:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:23.209 23:41:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:23.470 23:41:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:23.470 "name": "Existed_Raid", 00:07:23.470 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:23.470 "strip_size_kb": 64, 00:07:23.470 "state": "configuring", 00:07:23.470 "raid_level": "raid0", 00:07:23.470 "superblock": false, 00:07:23.470 "num_base_bdevs": 3, 00:07:23.470 "num_base_bdevs_discovered": 2, 00:07:23.470 "num_base_bdevs_operational": 3, 00:07:23.470 "base_bdevs_list": [ 00:07:23.470 { 00:07:23.470 "name": "BaseBdev1", 00:07:23.470 "uuid": "5a4fa438-802d-47e5-a418-fa61cf6baa15", 00:07:23.470 "is_configured": true, 00:07:23.470 "data_offset": 0, 00:07:23.470 "data_size": 65536 00:07:23.470 }, 00:07:23.470 { 00:07:23.470 "name": null, 00:07:23.470 "uuid": "9c5727a2-1496-4c68-8a1e-69260743ba76", 00:07:23.470 "is_configured": false, 00:07:23.470 "data_offset": 0, 00:07:23.470 "data_size": 65536 00:07:23.470 }, 00:07:23.470 { 00:07:23.470 "name": "BaseBdev3", 00:07:23.470 "uuid": "2c9302db-2499-4154-9a9d-48345d29d96e", 00:07:23.470 "is_configured": true, 00:07:23.470 "data_offset": 0, 00:07:23.470 "data_size": 65536 00:07:23.470 } 00:07:23.470 ] 00:07:23.470 }' 00:07:23.470 23:41:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:23.470 23:41:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:23.730 23:41:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:23.730 23:41:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:07:23.730 23:41:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:23.730 23:41:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:23.730 23:41:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:23.730 23:41:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:07:23.730 23:41:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:07:23.730 23:41:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:23.730 23:41:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:23.730 [2024-11-26 23:41:11.790780] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:07:23.730 23:41:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:23.730 23:41:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:07:23.730 23:41:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:23.730 23:41:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:23.730 23:41:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:23.730 23:41:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:23.730 23:41:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:07:23.730 23:41:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:23.730 23:41:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:23.730 23:41:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:23.730 23:41:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:23.730 23:41:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:23.730 23:41:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:23.730 23:41:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:23.730 23:41:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:23.730 23:41:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:23.730 23:41:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:23.730 "name": "Existed_Raid", 00:07:23.730 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:23.730 "strip_size_kb": 64, 00:07:23.730 "state": "configuring", 00:07:23.730 "raid_level": "raid0", 00:07:23.730 "superblock": false, 00:07:23.730 "num_base_bdevs": 3, 00:07:23.730 "num_base_bdevs_discovered": 1, 00:07:23.730 "num_base_bdevs_operational": 3, 00:07:23.730 "base_bdevs_list": [ 00:07:23.730 { 00:07:23.730 "name": "BaseBdev1", 00:07:23.730 "uuid": "5a4fa438-802d-47e5-a418-fa61cf6baa15", 00:07:23.730 "is_configured": true, 00:07:23.730 "data_offset": 0, 00:07:23.730 "data_size": 65536 00:07:23.730 }, 00:07:23.730 { 00:07:23.730 "name": null, 00:07:23.730 "uuid": "9c5727a2-1496-4c68-8a1e-69260743ba76", 00:07:23.730 "is_configured": false, 00:07:23.730 "data_offset": 0, 00:07:23.730 "data_size": 65536 00:07:23.730 }, 00:07:23.730 { 00:07:23.730 "name": null, 00:07:23.730 "uuid": "2c9302db-2499-4154-9a9d-48345d29d96e", 00:07:23.730 "is_configured": false, 00:07:23.730 "data_offset": 0, 00:07:23.730 "data_size": 65536 00:07:23.730 } 00:07:23.730 ] 00:07:23.730 }' 00:07:23.730 23:41:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:23.730 23:41:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:24.314 23:41:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:24.314 23:41:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:24.314 23:41:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:24.314 23:41:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:07:24.314 23:41:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:24.314 23:41:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:07:24.314 23:41:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:07:24.314 23:41:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:24.314 23:41:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:24.314 [2024-11-26 23:41:12.333902] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:07:24.314 23:41:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:24.314 23:41:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:07:24.314 23:41:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:24.314 23:41:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:24.314 23:41:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:24.314 23:41:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:24.314 23:41:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:07:24.314 23:41:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:24.314 23:41:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:24.314 23:41:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:24.314 23:41:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:24.314 23:41:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:24.314 23:41:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:24.314 23:41:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:24.314 23:41:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:24.314 23:41:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:24.314 23:41:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:24.314 "name": "Existed_Raid", 00:07:24.314 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:24.314 "strip_size_kb": 64, 00:07:24.314 "state": "configuring", 00:07:24.314 "raid_level": "raid0", 00:07:24.314 "superblock": false, 00:07:24.314 "num_base_bdevs": 3, 00:07:24.314 "num_base_bdevs_discovered": 2, 00:07:24.314 "num_base_bdevs_operational": 3, 00:07:24.314 "base_bdevs_list": [ 00:07:24.314 { 00:07:24.314 "name": "BaseBdev1", 00:07:24.314 "uuid": "5a4fa438-802d-47e5-a418-fa61cf6baa15", 00:07:24.314 "is_configured": true, 00:07:24.314 "data_offset": 0, 00:07:24.314 "data_size": 65536 00:07:24.314 }, 00:07:24.314 { 00:07:24.314 "name": null, 00:07:24.314 "uuid": "9c5727a2-1496-4c68-8a1e-69260743ba76", 00:07:24.314 "is_configured": false, 00:07:24.314 "data_offset": 0, 00:07:24.314 "data_size": 65536 00:07:24.314 }, 00:07:24.314 { 00:07:24.314 "name": "BaseBdev3", 00:07:24.314 "uuid": "2c9302db-2499-4154-9a9d-48345d29d96e", 00:07:24.314 "is_configured": true, 00:07:24.314 "data_offset": 0, 00:07:24.314 "data_size": 65536 00:07:24.314 } 00:07:24.314 ] 00:07:24.314 }' 00:07:24.314 23:41:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:24.314 23:41:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:24.884 23:41:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:24.884 23:41:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:07:24.884 23:41:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:24.884 23:41:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:24.884 23:41:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:24.884 23:41:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:07:24.884 23:41:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:07:24.884 23:41:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:24.884 23:41:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:24.884 [2024-11-26 23:41:12.833105] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:07:24.884 23:41:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:24.884 23:41:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:07:24.884 23:41:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:24.884 23:41:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:24.884 23:41:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:24.884 23:41:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:24.884 23:41:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:07:24.884 23:41:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:24.884 23:41:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:24.884 23:41:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:24.884 23:41:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:24.884 23:41:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:24.884 23:41:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:24.884 23:41:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:24.884 23:41:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:24.884 23:41:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:24.884 23:41:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:24.884 "name": "Existed_Raid", 00:07:24.884 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:24.884 "strip_size_kb": 64, 00:07:24.884 "state": "configuring", 00:07:24.884 "raid_level": "raid0", 00:07:24.884 "superblock": false, 00:07:24.884 "num_base_bdevs": 3, 00:07:24.884 "num_base_bdevs_discovered": 1, 00:07:24.884 "num_base_bdevs_operational": 3, 00:07:24.884 "base_bdevs_list": [ 00:07:24.884 { 00:07:24.884 "name": null, 00:07:24.884 "uuid": "5a4fa438-802d-47e5-a418-fa61cf6baa15", 00:07:24.884 "is_configured": false, 00:07:24.884 "data_offset": 0, 00:07:24.884 "data_size": 65536 00:07:24.884 }, 00:07:24.884 { 00:07:24.884 "name": null, 00:07:24.884 "uuid": "9c5727a2-1496-4c68-8a1e-69260743ba76", 00:07:24.884 "is_configured": false, 00:07:24.884 "data_offset": 0, 00:07:24.884 "data_size": 65536 00:07:24.884 }, 00:07:24.884 { 00:07:24.884 "name": "BaseBdev3", 00:07:24.884 "uuid": "2c9302db-2499-4154-9a9d-48345d29d96e", 00:07:24.884 "is_configured": true, 00:07:24.884 "data_offset": 0, 00:07:24.884 "data_size": 65536 00:07:24.884 } 00:07:24.884 ] 00:07:24.884 }' 00:07:24.884 23:41:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:24.884 23:41:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:25.452 23:41:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:25.452 23:41:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:07:25.452 23:41:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:25.452 23:41:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:25.452 23:41:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:25.452 23:41:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:07:25.452 23:41:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:07:25.452 23:41:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:25.452 23:41:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:25.452 [2024-11-26 23:41:13.326976] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:07:25.452 23:41:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:25.452 23:41:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:07:25.452 23:41:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:25.452 23:41:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:25.452 23:41:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:25.452 23:41:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:25.452 23:41:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:07:25.452 23:41:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:25.452 23:41:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:25.452 23:41:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:25.452 23:41:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:25.452 23:41:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:25.452 23:41:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:25.452 23:41:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:25.452 23:41:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:25.452 23:41:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:25.452 23:41:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:25.452 "name": "Existed_Raid", 00:07:25.452 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:25.452 "strip_size_kb": 64, 00:07:25.452 "state": "configuring", 00:07:25.452 "raid_level": "raid0", 00:07:25.452 "superblock": false, 00:07:25.452 "num_base_bdevs": 3, 00:07:25.452 "num_base_bdevs_discovered": 2, 00:07:25.452 "num_base_bdevs_operational": 3, 00:07:25.452 "base_bdevs_list": [ 00:07:25.452 { 00:07:25.452 "name": null, 00:07:25.452 "uuid": "5a4fa438-802d-47e5-a418-fa61cf6baa15", 00:07:25.452 "is_configured": false, 00:07:25.452 "data_offset": 0, 00:07:25.452 "data_size": 65536 00:07:25.452 }, 00:07:25.452 { 00:07:25.452 "name": "BaseBdev2", 00:07:25.452 "uuid": "9c5727a2-1496-4c68-8a1e-69260743ba76", 00:07:25.452 "is_configured": true, 00:07:25.452 "data_offset": 0, 00:07:25.452 "data_size": 65536 00:07:25.452 }, 00:07:25.452 { 00:07:25.452 "name": "BaseBdev3", 00:07:25.452 "uuid": "2c9302db-2499-4154-9a9d-48345d29d96e", 00:07:25.452 "is_configured": true, 00:07:25.452 "data_offset": 0, 00:07:25.452 "data_size": 65536 00:07:25.452 } 00:07:25.452 ] 00:07:25.452 }' 00:07:25.452 23:41:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:25.452 23:41:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:25.711 23:41:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:07:25.711 23:41:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:25.711 23:41:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:25.711 23:41:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:25.711 23:41:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:25.970 23:41:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:07:25.970 23:41:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:25.970 23:41:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:07:25.970 23:41:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:25.970 23:41:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:25.970 23:41:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:25.970 23:41:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 5a4fa438-802d-47e5-a418-fa61cf6baa15 00:07:25.970 23:41:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:25.970 23:41:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:25.970 [2024-11-26 23:41:13.921058] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:07:25.970 [2024-11-26 23:41:13.921184] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:07:25.970 [2024-11-26 23:41:13.921199] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 196608, blocklen 512 00:07:25.970 [2024-11-26 23:41:13.921474] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002870 00:07:25.970 [2024-11-26 23:41:13.921601] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:07:25.970 [2024-11-26 23:41:13.921611] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001c80 00:07:25.970 [2024-11-26 23:41:13.921785] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:25.970 NewBaseBdev 00:07:25.971 23:41:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:25.971 23:41:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:07:25.971 23:41:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:07:25.971 23:41:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:07:25.971 23:41:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:07:25.971 23:41:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:07:25.971 23:41:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:07:25.971 23:41:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:07:25.971 23:41:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:25.971 23:41:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:25.971 23:41:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:25.971 23:41:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:07:25.971 23:41:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:25.971 23:41:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:25.971 [ 00:07:25.971 { 00:07:25.971 "name": "NewBaseBdev", 00:07:25.971 "aliases": [ 00:07:25.971 "5a4fa438-802d-47e5-a418-fa61cf6baa15" 00:07:25.971 ], 00:07:25.971 "product_name": "Malloc disk", 00:07:25.971 "block_size": 512, 00:07:25.971 "num_blocks": 65536, 00:07:25.971 "uuid": "5a4fa438-802d-47e5-a418-fa61cf6baa15", 00:07:25.971 "assigned_rate_limits": { 00:07:25.971 "rw_ios_per_sec": 0, 00:07:25.971 "rw_mbytes_per_sec": 0, 00:07:25.971 "r_mbytes_per_sec": 0, 00:07:25.971 "w_mbytes_per_sec": 0 00:07:25.971 }, 00:07:25.971 "claimed": true, 00:07:25.971 "claim_type": "exclusive_write", 00:07:25.971 "zoned": false, 00:07:25.971 "supported_io_types": { 00:07:25.971 "read": true, 00:07:25.971 "write": true, 00:07:25.971 "unmap": true, 00:07:25.971 "flush": true, 00:07:25.971 "reset": true, 00:07:25.971 "nvme_admin": false, 00:07:25.971 "nvme_io": false, 00:07:25.971 "nvme_io_md": false, 00:07:25.971 "write_zeroes": true, 00:07:25.971 "zcopy": true, 00:07:25.971 "get_zone_info": false, 00:07:25.971 "zone_management": false, 00:07:25.971 "zone_append": false, 00:07:25.971 "compare": false, 00:07:25.971 "compare_and_write": false, 00:07:25.971 "abort": true, 00:07:25.971 "seek_hole": false, 00:07:25.971 "seek_data": false, 00:07:25.971 "copy": true, 00:07:25.971 "nvme_iov_md": false 00:07:25.971 }, 00:07:25.971 "memory_domains": [ 00:07:25.971 { 00:07:25.971 "dma_device_id": "system", 00:07:25.971 "dma_device_type": 1 00:07:25.971 }, 00:07:25.971 { 00:07:25.971 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:25.971 "dma_device_type": 2 00:07:25.971 } 00:07:25.971 ], 00:07:25.971 "driver_specific": {} 00:07:25.971 } 00:07:25.971 ] 00:07:25.971 23:41:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:25.971 23:41:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:07:25.971 23:41:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid0 64 3 00:07:25.971 23:41:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:25.971 23:41:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:25.971 23:41:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:25.971 23:41:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:25.971 23:41:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:07:25.971 23:41:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:25.971 23:41:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:25.971 23:41:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:25.971 23:41:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:25.971 23:41:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:25.971 23:41:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:25.971 23:41:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:25.971 23:41:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:25.971 23:41:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:25.971 23:41:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:25.971 "name": "Existed_Raid", 00:07:25.971 "uuid": "681501ec-545d-4a6e-a672-fea1abc658cf", 00:07:25.971 "strip_size_kb": 64, 00:07:25.971 "state": "online", 00:07:25.971 "raid_level": "raid0", 00:07:25.971 "superblock": false, 00:07:25.971 "num_base_bdevs": 3, 00:07:25.971 "num_base_bdevs_discovered": 3, 00:07:25.971 "num_base_bdevs_operational": 3, 00:07:25.971 "base_bdevs_list": [ 00:07:25.971 { 00:07:25.971 "name": "NewBaseBdev", 00:07:25.971 "uuid": "5a4fa438-802d-47e5-a418-fa61cf6baa15", 00:07:25.971 "is_configured": true, 00:07:25.971 "data_offset": 0, 00:07:25.971 "data_size": 65536 00:07:25.971 }, 00:07:25.971 { 00:07:25.971 "name": "BaseBdev2", 00:07:25.971 "uuid": "9c5727a2-1496-4c68-8a1e-69260743ba76", 00:07:25.971 "is_configured": true, 00:07:25.971 "data_offset": 0, 00:07:25.971 "data_size": 65536 00:07:25.971 }, 00:07:25.971 { 00:07:25.971 "name": "BaseBdev3", 00:07:25.971 "uuid": "2c9302db-2499-4154-9a9d-48345d29d96e", 00:07:25.971 "is_configured": true, 00:07:25.971 "data_offset": 0, 00:07:25.971 "data_size": 65536 00:07:25.971 } 00:07:25.971 ] 00:07:25.971 }' 00:07:25.971 23:41:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:25.971 23:41:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:26.231 23:41:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:07:26.231 23:41:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:07:26.231 23:41:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:07:26.231 23:41:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:07:26.231 23:41:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:07:26.231 23:41:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:07:26.231 23:41:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:07:26.231 23:41:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:26.231 23:41:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:26.231 23:41:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:07:26.492 [2024-11-26 23:41:14.364697] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:26.492 23:41:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:26.492 23:41:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:07:26.492 "name": "Existed_Raid", 00:07:26.492 "aliases": [ 00:07:26.492 "681501ec-545d-4a6e-a672-fea1abc658cf" 00:07:26.492 ], 00:07:26.492 "product_name": "Raid Volume", 00:07:26.492 "block_size": 512, 00:07:26.492 "num_blocks": 196608, 00:07:26.492 "uuid": "681501ec-545d-4a6e-a672-fea1abc658cf", 00:07:26.492 "assigned_rate_limits": { 00:07:26.492 "rw_ios_per_sec": 0, 00:07:26.492 "rw_mbytes_per_sec": 0, 00:07:26.492 "r_mbytes_per_sec": 0, 00:07:26.492 "w_mbytes_per_sec": 0 00:07:26.492 }, 00:07:26.492 "claimed": false, 00:07:26.492 "zoned": false, 00:07:26.492 "supported_io_types": { 00:07:26.492 "read": true, 00:07:26.492 "write": true, 00:07:26.492 "unmap": true, 00:07:26.492 "flush": true, 00:07:26.492 "reset": true, 00:07:26.492 "nvme_admin": false, 00:07:26.492 "nvme_io": false, 00:07:26.492 "nvme_io_md": false, 00:07:26.492 "write_zeroes": true, 00:07:26.492 "zcopy": false, 00:07:26.492 "get_zone_info": false, 00:07:26.492 "zone_management": false, 00:07:26.492 "zone_append": false, 00:07:26.492 "compare": false, 00:07:26.492 "compare_and_write": false, 00:07:26.492 "abort": false, 00:07:26.492 "seek_hole": false, 00:07:26.492 "seek_data": false, 00:07:26.492 "copy": false, 00:07:26.492 "nvme_iov_md": false 00:07:26.492 }, 00:07:26.492 "memory_domains": [ 00:07:26.492 { 00:07:26.492 "dma_device_id": "system", 00:07:26.492 "dma_device_type": 1 00:07:26.492 }, 00:07:26.492 { 00:07:26.492 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:26.492 "dma_device_type": 2 00:07:26.492 }, 00:07:26.492 { 00:07:26.492 "dma_device_id": "system", 00:07:26.492 "dma_device_type": 1 00:07:26.492 }, 00:07:26.492 { 00:07:26.492 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:26.492 "dma_device_type": 2 00:07:26.492 }, 00:07:26.492 { 00:07:26.492 "dma_device_id": "system", 00:07:26.492 "dma_device_type": 1 00:07:26.492 }, 00:07:26.492 { 00:07:26.492 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:26.492 "dma_device_type": 2 00:07:26.492 } 00:07:26.492 ], 00:07:26.492 "driver_specific": { 00:07:26.492 "raid": { 00:07:26.492 "uuid": "681501ec-545d-4a6e-a672-fea1abc658cf", 00:07:26.492 "strip_size_kb": 64, 00:07:26.492 "state": "online", 00:07:26.492 "raid_level": "raid0", 00:07:26.492 "superblock": false, 00:07:26.492 "num_base_bdevs": 3, 00:07:26.492 "num_base_bdevs_discovered": 3, 00:07:26.492 "num_base_bdevs_operational": 3, 00:07:26.492 "base_bdevs_list": [ 00:07:26.492 { 00:07:26.492 "name": "NewBaseBdev", 00:07:26.492 "uuid": "5a4fa438-802d-47e5-a418-fa61cf6baa15", 00:07:26.492 "is_configured": true, 00:07:26.492 "data_offset": 0, 00:07:26.492 "data_size": 65536 00:07:26.492 }, 00:07:26.492 { 00:07:26.492 "name": "BaseBdev2", 00:07:26.492 "uuid": "9c5727a2-1496-4c68-8a1e-69260743ba76", 00:07:26.492 "is_configured": true, 00:07:26.492 "data_offset": 0, 00:07:26.492 "data_size": 65536 00:07:26.492 }, 00:07:26.492 { 00:07:26.492 "name": "BaseBdev3", 00:07:26.492 "uuid": "2c9302db-2499-4154-9a9d-48345d29d96e", 00:07:26.492 "is_configured": true, 00:07:26.492 "data_offset": 0, 00:07:26.492 "data_size": 65536 00:07:26.492 } 00:07:26.493 ] 00:07:26.493 } 00:07:26.493 } 00:07:26.493 }' 00:07:26.493 23:41:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:07:26.493 23:41:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:07:26.493 BaseBdev2 00:07:26.493 BaseBdev3' 00:07:26.493 23:41:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:26.493 23:41:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:07:26.493 23:41:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:26.493 23:41:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:07:26.493 23:41:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:26.493 23:41:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:26.493 23:41:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:26.493 23:41:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:26.493 23:41:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:26.493 23:41:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:26.493 23:41:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:26.493 23:41:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:07:26.493 23:41:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:26.493 23:41:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:26.493 23:41:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:26.493 23:41:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:26.493 23:41:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:26.493 23:41:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:26.493 23:41:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:26.493 23:41:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:07:26.493 23:41:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:26.493 23:41:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:26.493 23:41:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:26.493 23:41:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:26.753 23:41:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:26.753 23:41:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:26.753 23:41:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:07:26.753 23:41:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:26.753 23:41:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:26.753 [2024-11-26 23:41:14.651868] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:07:26.753 [2024-11-26 23:41:14.651896] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:26.753 [2024-11-26 23:41:14.651964] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:26.753 [2024-11-26 23:41:14.652018] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:26.753 [2024-11-26 23:41:14.652038] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name Existed_Raid, state offline 00:07:26.753 23:41:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:26.753 23:41:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 74748 00:07:26.753 23:41:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # '[' -z 74748 ']' 00:07:26.753 23:41:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@958 -- # kill -0 74748 00:07:26.753 23:41:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # uname 00:07:26.753 23:41:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:07:26.753 23:41:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 74748 00:07:26.753 killing process with pid 74748 00:07:26.753 23:41:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:07:26.753 23:41:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:07:26.753 23:41:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 74748' 00:07:26.753 23:41:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@973 -- # kill 74748 00:07:26.753 [2024-11-26 23:41:14.702167] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:26.753 23:41:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@978 -- # wait 74748 00:07:26.753 [2024-11-26 23:41:14.732490] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:27.012 ************************************ 00:07:27.013 END TEST raid_state_function_test 00:07:27.013 ************************************ 00:07:27.013 23:41:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:07:27.013 00:07:27.013 real 0m8.745s 00:07:27.013 user 0m14.929s 00:07:27.013 sys 0m1.712s 00:07:27.013 23:41:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:07:27.013 23:41:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:27.013 23:41:15 bdev_raid -- bdev/bdev_raid.sh@969 -- # run_test raid_state_function_test_sb raid_state_function_test raid0 3 true 00:07:27.013 23:41:15 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:07:27.013 23:41:15 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:27.013 23:41:15 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:27.013 ************************************ 00:07:27.013 START TEST raid_state_function_test_sb 00:07:27.013 ************************************ 00:07:27.013 23:41:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1129 -- # raid_state_function_test raid0 3 true 00:07:27.013 23:41:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=raid0 00:07:27.013 23:41:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=3 00:07:27.013 23:41:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:07:27.013 23:41:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:07:27.013 23:41:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:07:27.013 23:41:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:27.013 23:41:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:07:27.013 23:41:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:07:27.013 23:41:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:27.013 23:41:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:07:27.013 23:41:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:07:27.013 23:41:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:27.013 23:41:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:07:27.013 23:41:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:07:27.013 23:41:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:27.013 23:41:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:07:27.013 23:41:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:07:27.013 23:41:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:07:27.013 23:41:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:07:27.013 23:41:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:07:27.013 23:41:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:07:27.013 23:41:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' raid0 '!=' raid1 ']' 00:07:27.013 23:41:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:07:27.013 23:41:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:07:27.013 23:41:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:07:27.013 23:41:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:07:27.013 23:41:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=75353 00:07:27.013 23:41:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:07:27.013 23:41:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 75353' 00:07:27.013 Process raid pid: 75353 00:07:27.013 23:41:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 75353 00:07:27.013 23:41:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@835 -- # '[' -z 75353 ']' 00:07:27.013 23:41:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:27.013 23:41:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:27.013 23:41:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:27.013 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:27.013 23:41:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:27.013 23:41:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:27.013 [2024-11-26 23:41:15.113440] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:07:27.013 [2024-11-26 23:41:15.113659] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:07:27.273 [2024-11-26 23:41:15.263192] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:27.273 [2024-11-26 23:41:15.288319] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:27.273 [2024-11-26 23:41:15.330153] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:27.273 [2024-11-26 23:41:15.330265] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:28.210 23:41:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:28.210 23:41:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@868 -- # return 0 00:07:28.210 23:41:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:07:28.210 23:41:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:28.210 23:41:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:28.210 [2024-11-26 23:41:15.980261] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:07:28.210 [2024-11-26 23:41:15.980321] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:07:28.210 [2024-11-26 23:41:15.980331] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:28.210 [2024-11-26 23:41:15.980351] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:28.210 [2024-11-26 23:41:15.980358] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:07:28.210 [2024-11-26 23:41:15.980369] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:07:28.210 23:41:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:28.210 23:41:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:07:28.211 23:41:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:28.211 23:41:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:28.211 23:41:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:28.211 23:41:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:28.211 23:41:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:07:28.211 23:41:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:28.211 23:41:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:28.211 23:41:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:28.211 23:41:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:28.211 23:41:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:28.211 23:41:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:28.211 23:41:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:28.211 23:41:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:28.211 23:41:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:28.211 23:41:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:28.211 "name": "Existed_Raid", 00:07:28.211 "uuid": "701df528-90bb-43ad-ba33-a18d1175e8b7", 00:07:28.211 "strip_size_kb": 64, 00:07:28.211 "state": "configuring", 00:07:28.211 "raid_level": "raid0", 00:07:28.211 "superblock": true, 00:07:28.211 "num_base_bdevs": 3, 00:07:28.211 "num_base_bdevs_discovered": 0, 00:07:28.211 "num_base_bdevs_operational": 3, 00:07:28.211 "base_bdevs_list": [ 00:07:28.211 { 00:07:28.211 "name": "BaseBdev1", 00:07:28.211 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:28.211 "is_configured": false, 00:07:28.211 "data_offset": 0, 00:07:28.211 "data_size": 0 00:07:28.211 }, 00:07:28.211 { 00:07:28.211 "name": "BaseBdev2", 00:07:28.211 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:28.211 "is_configured": false, 00:07:28.211 "data_offset": 0, 00:07:28.211 "data_size": 0 00:07:28.211 }, 00:07:28.211 { 00:07:28.211 "name": "BaseBdev3", 00:07:28.211 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:28.211 "is_configured": false, 00:07:28.211 "data_offset": 0, 00:07:28.211 "data_size": 0 00:07:28.211 } 00:07:28.211 ] 00:07:28.211 }' 00:07:28.211 23:41:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:28.211 23:41:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:28.470 23:41:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:07:28.470 23:41:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:28.470 23:41:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:28.470 [2024-11-26 23:41:16.447386] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:07:28.470 [2024-11-26 23:41:16.447502] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:07:28.470 23:41:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:28.470 23:41:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:07:28.470 23:41:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:28.470 23:41:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:28.470 [2024-11-26 23:41:16.459371] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:07:28.470 [2024-11-26 23:41:16.459460] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:07:28.470 [2024-11-26 23:41:16.459488] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:28.470 [2024-11-26 23:41:16.459511] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:28.470 [2024-11-26 23:41:16.459528] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:07:28.470 [2024-11-26 23:41:16.459549] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:07:28.470 23:41:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:28.470 23:41:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:07:28.470 23:41:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:28.471 23:41:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:28.471 [2024-11-26 23:41:16.480080] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:28.471 BaseBdev1 00:07:28.471 23:41:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:28.471 23:41:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:07:28.471 23:41:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:07:28.471 23:41:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:07:28.471 23:41:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:07:28.471 23:41:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:07:28.471 23:41:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:07:28.471 23:41:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:07:28.471 23:41:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:28.471 23:41:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:28.471 23:41:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:28.471 23:41:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:07:28.471 23:41:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:28.471 23:41:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:28.471 [ 00:07:28.471 { 00:07:28.471 "name": "BaseBdev1", 00:07:28.471 "aliases": [ 00:07:28.471 "38e62c0d-295f-4497-aad6-67d6851006c3" 00:07:28.471 ], 00:07:28.471 "product_name": "Malloc disk", 00:07:28.471 "block_size": 512, 00:07:28.471 "num_blocks": 65536, 00:07:28.471 "uuid": "38e62c0d-295f-4497-aad6-67d6851006c3", 00:07:28.471 "assigned_rate_limits": { 00:07:28.471 "rw_ios_per_sec": 0, 00:07:28.471 "rw_mbytes_per_sec": 0, 00:07:28.471 "r_mbytes_per_sec": 0, 00:07:28.471 "w_mbytes_per_sec": 0 00:07:28.471 }, 00:07:28.471 "claimed": true, 00:07:28.471 "claim_type": "exclusive_write", 00:07:28.471 "zoned": false, 00:07:28.471 "supported_io_types": { 00:07:28.471 "read": true, 00:07:28.471 "write": true, 00:07:28.471 "unmap": true, 00:07:28.471 "flush": true, 00:07:28.471 "reset": true, 00:07:28.471 "nvme_admin": false, 00:07:28.471 "nvme_io": false, 00:07:28.471 "nvme_io_md": false, 00:07:28.471 "write_zeroes": true, 00:07:28.471 "zcopy": true, 00:07:28.471 "get_zone_info": false, 00:07:28.471 "zone_management": false, 00:07:28.471 "zone_append": false, 00:07:28.471 "compare": false, 00:07:28.471 "compare_and_write": false, 00:07:28.471 "abort": true, 00:07:28.471 "seek_hole": false, 00:07:28.471 "seek_data": false, 00:07:28.471 "copy": true, 00:07:28.471 "nvme_iov_md": false 00:07:28.471 }, 00:07:28.471 "memory_domains": [ 00:07:28.471 { 00:07:28.471 "dma_device_id": "system", 00:07:28.471 "dma_device_type": 1 00:07:28.471 }, 00:07:28.471 { 00:07:28.471 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:28.471 "dma_device_type": 2 00:07:28.471 } 00:07:28.471 ], 00:07:28.471 "driver_specific": {} 00:07:28.471 } 00:07:28.471 ] 00:07:28.471 23:41:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:28.471 23:41:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:07:28.471 23:41:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:07:28.471 23:41:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:28.471 23:41:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:28.471 23:41:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:28.471 23:41:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:28.471 23:41:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:07:28.471 23:41:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:28.471 23:41:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:28.471 23:41:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:28.471 23:41:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:28.471 23:41:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:28.471 23:41:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:28.471 23:41:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:28.471 23:41:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:28.471 23:41:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:28.471 23:41:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:28.471 "name": "Existed_Raid", 00:07:28.471 "uuid": "12ea54ea-3b3d-4a07-bb3f-3858051f330a", 00:07:28.471 "strip_size_kb": 64, 00:07:28.471 "state": "configuring", 00:07:28.471 "raid_level": "raid0", 00:07:28.471 "superblock": true, 00:07:28.471 "num_base_bdevs": 3, 00:07:28.471 "num_base_bdevs_discovered": 1, 00:07:28.471 "num_base_bdevs_operational": 3, 00:07:28.471 "base_bdevs_list": [ 00:07:28.471 { 00:07:28.471 "name": "BaseBdev1", 00:07:28.471 "uuid": "38e62c0d-295f-4497-aad6-67d6851006c3", 00:07:28.471 "is_configured": true, 00:07:28.471 "data_offset": 2048, 00:07:28.471 "data_size": 63488 00:07:28.471 }, 00:07:28.471 { 00:07:28.471 "name": "BaseBdev2", 00:07:28.471 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:28.471 "is_configured": false, 00:07:28.471 "data_offset": 0, 00:07:28.471 "data_size": 0 00:07:28.471 }, 00:07:28.471 { 00:07:28.471 "name": "BaseBdev3", 00:07:28.471 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:28.471 "is_configured": false, 00:07:28.471 "data_offset": 0, 00:07:28.471 "data_size": 0 00:07:28.471 } 00:07:28.471 ] 00:07:28.471 }' 00:07:28.471 23:41:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:28.471 23:41:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:29.041 23:41:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:07:29.041 23:41:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:29.041 23:41:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:29.041 [2024-11-26 23:41:16.935388] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:07:29.041 [2024-11-26 23:41:16.935443] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:07:29.041 23:41:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:29.041 23:41:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:07:29.041 23:41:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:29.041 23:41:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:29.041 [2024-11-26 23:41:16.943401] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:29.041 [2024-11-26 23:41:16.945264] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:29.041 [2024-11-26 23:41:16.945305] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:29.041 [2024-11-26 23:41:16.945315] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:07:29.041 [2024-11-26 23:41:16.945325] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:07:29.041 23:41:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:29.041 23:41:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:07:29.041 23:41:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:07:29.041 23:41:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:07:29.041 23:41:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:29.041 23:41:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:29.041 23:41:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:29.041 23:41:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:29.041 23:41:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:07:29.041 23:41:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:29.041 23:41:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:29.041 23:41:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:29.041 23:41:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:29.041 23:41:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:29.041 23:41:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:29.041 23:41:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:29.041 23:41:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:29.041 23:41:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:29.041 23:41:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:29.041 "name": "Existed_Raid", 00:07:29.041 "uuid": "149fb2b5-ed82-4c95-8711-3125d7cec04f", 00:07:29.041 "strip_size_kb": 64, 00:07:29.041 "state": "configuring", 00:07:29.041 "raid_level": "raid0", 00:07:29.041 "superblock": true, 00:07:29.041 "num_base_bdevs": 3, 00:07:29.041 "num_base_bdevs_discovered": 1, 00:07:29.041 "num_base_bdevs_operational": 3, 00:07:29.041 "base_bdevs_list": [ 00:07:29.041 { 00:07:29.041 "name": "BaseBdev1", 00:07:29.041 "uuid": "38e62c0d-295f-4497-aad6-67d6851006c3", 00:07:29.041 "is_configured": true, 00:07:29.041 "data_offset": 2048, 00:07:29.041 "data_size": 63488 00:07:29.041 }, 00:07:29.041 { 00:07:29.041 "name": "BaseBdev2", 00:07:29.041 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:29.041 "is_configured": false, 00:07:29.041 "data_offset": 0, 00:07:29.041 "data_size": 0 00:07:29.041 }, 00:07:29.041 { 00:07:29.041 "name": "BaseBdev3", 00:07:29.041 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:29.041 "is_configured": false, 00:07:29.041 "data_offset": 0, 00:07:29.041 "data_size": 0 00:07:29.041 } 00:07:29.041 ] 00:07:29.041 }' 00:07:29.041 23:41:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:29.041 23:41:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:29.302 23:41:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:07:29.302 23:41:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:29.302 23:41:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:29.302 [2024-11-26 23:41:17.345753] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:07:29.302 BaseBdev2 00:07:29.302 23:41:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:29.302 23:41:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:07:29.302 23:41:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:07:29.302 23:41:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:07:29.302 23:41:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:07:29.302 23:41:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:07:29.302 23:41:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:07:29.302 23:41:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:07:29.302 23:41:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:29.302 23:41:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:29.302 23:41:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:29.302 23:41:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:07:29.302 23:41:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:29.302 23:41:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:29.302 [ 00:07:29.302 { 00:07:29.302 "name": "BaseBdev2", 00:07:29.302 "aliases": [ 00:07:29.302 "03ef36ad-03a2-493e-922d-02aae7bef418" 00:07:29.302 ], 00:07:29.302 "product_name": "Malloc disk", 00:07:29.302 "block_size": 512, 00:07:29.302 "num_blocks": 65536, 00:07:29.302 "uuid": "03ef36ad-03a2-493e-922d-02aae7bef418", 00:07:29.302 "assigned_rate_limits": { 00:07:29.302 "rw_ios_per_sec": 0, 00:07:29.302 "rw_mbytes_per_sec": 0, 00:07:29.302 "r_mbytes_per_sec": 0, 00:07:29.302 "w_mbytes_per_sec": 0 00:07:29.302 }, 00:07:29.302 "claimed": true, 00:07:29.302 "claim_type": "exclusive_write", 00:07:29.302 "zoned": false, 00:07:29.302 "supported_io_types": { 00:07:29.302 "read": true, 00:07:29.302 "write": true, 00:07:29.302 "unmap": true, 00:07:29.302 "flush": true, 00:07:29.302 "reset": true, 00:07:29.302 "nvme_admin": false, 00:07:29.302 "nvme_io": false, 00:07:29.302 "nvme_io_md": false, 00:07:29.302 "write_zeroes": true, 00:07:29.302 "zcopy": true, 00:07:29.302 "get_zone_info": false, 00:07:29.302 "zone_management": false, 00:07:29.302 "zone_append": false, 00:07:29.302 "compare": false, 00:07:29.302 "compare_and_write": false, 00:07:29.302 "abort": true, 00:07:29.302 "seek_hole": false, 00:07:29.302 "seek_data": false, 00:07:29.302 "copy": true, 00:07:29.302 "nvme_iov_md": false 00:07:29.302 }, 00:07:29.302 "memory_domains": [ 00:07:29.302 { 00:07:29.302 "dma_device_id": "system", 00:07:29.302 "dma_device_type": 1 00:07:29.302 }, 00:07:29.302 { 00:07:29.302 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:29.302 "dma_device_type": 2 00:07:29.302 } 00:07:29.302 ], 00:07:29.302 "driver_specific": {} 00:07:29.302 } 00:07:29.302 ] 00:07:29.302 23:41:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:29.302 23:41:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:07:29.302 23:41:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:07:29.302 23:41:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:07:29.302 23:41:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:07:29.302 23:41:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:29.302 23:41:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:29.302 23:41:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:29.302 23:41:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:29.302 23:41:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:07:29.302 23:41:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:29.302 23:41:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:29.302 23:41:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:29.302 23:41:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:29.302 23:41:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:29.302 23:41:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:29.302 23:41:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:29.302 23:41:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:29.302 23:41:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:29.562 23:41:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:29.562 "name": "Existed_Raid", 00:07:29.562 "uuid": "149fb2b5-ed82-4c95-8711-3125d7cec04f", 00:07:29.562 "strip_size_kb": 64, 00:07:29.562 "state": "configuring", 00:07:29.562 "raid_level": "raid0", 00:07:29.562 "superblock": true, 00:07:29.562 "num_base_bdevs": 3, 00:07:29.562 "num_base_bdevs_discovered": 2, 00:07:29.562 "num_base_bdevs_operational": 3, 00:07:29.562 "base_bdevs_list": [ 00:07:29.562 { 00:07:29.562 "name": "BaseBdev1", 00:07:29.562 "uuid": "38e62c0d-295f-4497-aad6-67d6851006c3", 00:07:29.562 "is_configured": true, 00:07:29.562 "data_offset": 2048, 00:07:29.562 "data_size": 63488 00:07:29.562 }, 00:07:29.562 { 00:07:29.562 "name": "BaseBdev2", 00:07:29.562 "uuid": "03ef36ad-03a2-493e-922d-02aae7bef418", 00:07:29.562 "is_configured": true, 00:07:29.562 "data_offset": 2048, 00:07:29.562 "data_size": 63488 00:07:29.562 }, 00:07:29.562 { 00:07:29.562 "name": "BaseBdev3", 00:07:29.562 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:29.562 "is_configured": false, 00:07:29.562 "data_offset": 0, 00:07:29.562 "data_size": 0 00:07:29.562 } 00:07:29.562 ] 00:07:29.562 }' 00:07:29.562 23:41:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:29.562 23:41:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:29.822 23:41:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:07:29.822 23:41:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:29.822 23:41:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:29.822 [2024-11-26 23:41:17.862457] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:07:29.822 [2024-11-26 23:41:17.862951] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:07:29.822 [2024-11-26 23:41:17.863008] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:07:29.822 BaseBdev3 00:07:29.822 [2024-11-26 23:41:17.863941] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:07:29.822 23:41:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:29.822 [2024-11-26 23:41:17.864433] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:07:29.822 [2024-11-26 23:41:17.864474] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:07:29.822 23:41:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:07:29.822 [2024-11-26 23:41:17.864853] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:29.822 23:41:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:07:29.822 23:41:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:07:29.822 23:41:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:07:29.822 23:41:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:07:29.822 23:41:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:07:29.822 23:41:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:07:29.822 23:41:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:29.822 23:41:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:29.822 23:41:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:29.822 23:41:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:07:29.822 23:41:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:29.822 23:41:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:29.822 [ 00:07:29.822 { 00:07:29.822 "name": "BaseBdev3", 00:07:29.822 "aliases": [ 00:07:29.822 "f4794e2c-5e53-4903-bd7a-171440b2a6b6" 00:07:29.822 ], 00:07:29.822 "product_name": "Malloc disk", 00:07:29.822 "block_size": 512, 00:07:29.822 "num_blocks": 65536, 00:07:29.822 "uuid": "f4794e2c-5e53-4903-bd7a-171440b2a6b6", 00:07:29.822 "assigned_rate_limits": { 00:07:29.822 "rw_ios_per_sec": 0, 00:07:29.822 "rw_mbytes_per_sec": 0, 00:07:29.822 "r_mbytes_per_sec": 0, 00:07:29.822 "w_mbytes_per_sec": 0 00:07:29.822 }, 00:07:29.822 "claimed": true, 00:07:29.822 "claim_type": "exclusive_write", 00:07:29.822 "zoned": false, 00:07:29.822 "supported_io_types": { 00:07:29.822 "read": true, 00:07:29.822 "write": true, 00:07:29.822 "unmap": true, 00:07:29.822 "flush": true, 00:07:29.822 "reset": true, 00:07:29.822 "nvme_admin": false, 00:07:29.822 "nvme_io": false, 00:07:29.822 "nvme_io_md": false, 00:07:29.822 "write_zeroes": true, 00:07:29.822 "zcopy": true, 00:07:29.822 "get_zone_info": false, 00:07:29.822 "zone_management": false, 00:07:29.822 "zone_append": false, 00:07:29.822 "compare": false, 00:07:29.822 "compare_and_write": false, 00:07:29.822 "abort": true, 00:07:29.822 "seek_hole": false, 00:07:29.822 "seek_data": false, 00:07:29.822 "copy": true, 00:07:29.822 "nvme_iov_md": false 00:07:29.822 }, 00:07:29.822 "memory_domains": [ 00:07:29.822 { 00:07:29.822 "dma_device_id": "system", 00:07:29.822 "dma_device_type": 1 00:07:29.822 }, 00:07:29.822 { 00:07:29.822 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:29.822 "dma_device_type": 2 00:07:29.822 } 00:07:29.822 ], 00:07:29.822 "driver_specific": {} 00:07:29.822 } 00:07:29.822 ] 00:07:29.822 23:41:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:29.822 23:41:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:07:29.822 23:41:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:07:29.822 23:41:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:07:29.822 23:41:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid0 64 3 00:07:29.822 23:41:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:29.822 23:41:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:29.822 23:41:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:29.822 23:41:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:29.822 23:41:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:07:29.822 23:41:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:29.822 23:41:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:29.822 23:41:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:29.822 23:41:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:29.822 23:41:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:29.822 23:41:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:29.822 23:41:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:29.822 23:41:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:29.822 23:41:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:30.083 23:41:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:30.083 "name": "Existed_Raid", 00:07:30.083 "uuid": "149fb2b5-ed82-4c95-8711-3125d7cec04f", 00:07:30.083 "strip_size_kb": 64, 00:07:30.083 "state": "online", 00:07:30.083 "raid_level": "raid0", 00:07:30.083 "superblock": true, 00:07:30.083 "num_base_bdevs": 3, 00:07:30.083 "num_base_bdevs_discovered": 3, 00:07:30.083 "num_base_bdevs_operational": 3, 00:07:30.083 "base_bdevs_list": [ 00:07:30.083 { 00:07:30.083 "name": "BaseBdev1", 00:07:30.083 "uuid": "38e62c0d-295f-4497-aad6-67d6851006c3", 00:07:30.083 "is_configured": true, 00:07:30.083 "data_offset": 2048, 00:07:30.083 "data_size": 63488 00:07:30.083 }, 00:07:30.083 { 00:07:30.083 "name": "BaseBdev2", 00:07:30.083 "uuid": "03ef36ad-03a2-493e-922d-02aae7bef418", 00:07:30.083 "is_configured": true, 00:07:30.083 "data_offset": 2048, 00:07:30.083 "data_size": 63488 00:07:30.083 }, 00:07:30.083 { 00:07:30.083 "name": "BaseBdev3", 00:07:30.083 "uuid": "f4794e2c-5e53-4903-bd7a-171440b2a6b6", 00:07:30.083 "is_configured": true, 00:07:30.083 "data_offset": 2048, 00:07:30.083 "data_size": 63488 00:07:30.083 } 00:07:30.083 ] 00:07:30.083 }' 00:07:30.083 23:41:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:30.083 23:41:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:30.343 23:41:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:07:30.343 23:41:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:07:30.343 23:41:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:07:30.343 23:41:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:07:30.343 23:41:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:07:30.343 23:41:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:07:30.343 23:41:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:07:30.343 23:41:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:30.343 23:41:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:30.343 23:41:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:07:30.343 [2024-11-26 23:41:18.397909] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:30.343 23:41:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:30.343 23:41:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:07:30.343 "name": "Existed_Raid", 00:07:30.343 "aliases": [ 00:07:30.343 "149fb2b5-ed82-4c95-8711-3125d7cec04f" 00:07:30.343 ], 00:07:30.343 "product_name": "Raid Volume", 00:07:30.343 "block_size": 512, 00:07:30.343 "num_blocks": 190464, 00:07:30.343 "uuid": "149fb2b5-ed82-4c95-8711-3125d7cec04f", 00:07:30.343 "assigned_rate_limits": { 00:07:30.343 "rw_ios_per_sec": 0, 00:07:30.343 "rw_mbytes_per_sec": 0, 00:07:30.343 "r_mbytes_per_sec": 0, 00:07:30.343 "w_mbytes_per_sec": 0 00:07:30.343 }, 00:07:30.343 "claimed": false, 00:07:30.343 "zoned": false, 00:07:30.343 "supported_io_types": { 00:07:30.343 "read": true, 00:07:30.343 "write": true, 00:07:30.343 "unmap": true, 00:07:30.343 "flush": true, 00:07:30.343 "reset": true, 00:07:30.343 "nvme_admin": false, 00:07:30.343 "nvme_io": false, 00:07:30.343 "nvme_io_md": false, 00:07:30.343 "write_zeroes": true, 00:07:30.343 "zcopy": false, 00:07:30.343 "get_zone_info": false, 00:07:30.343 "zone_management": false, 00:07:30.343 "zone_append": false, 00:07:30.343 "compare": false, 00:07:30.343 "compare_and_write": false, 00:07:30.343 "abort": false, 00:07:30.343 "seek_hole": false, 00:07:30.343 "seek_data": false, 00:07:30.343 "copy": false, 00:07:30.343 "nvme_iov_md": false 00:07:30.343 }, 00:07:30.343 "memory_domains": [ 00:07:30.343 { 00:07:30.343 "dma_device_id": "system", 00:07:30.343 "dma_device_type": 1 00:07:30.343 }, 00:07:30.343 { 00:07:30.343 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:30.343 "dma_device_type": 2 00:07:30.343 }, 00:07:30.343 { 00:07:30.343 "dma_device_id": "system", 00:07:30.343 "dma_device_type": 1 00:07:30.343 }, 00:07:30.343 { 00:07:30.343 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:30.343 "dma_device_type": 2 00:07:30.343 }, 00:07:30.343 { 00:07:30.343 "dma_device_id": "system", 00:07:30.343 "dma_device_type": 1 00:07:30.343 }, 00:07:30.343 { 00:07:30.343 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:30.343 "dma_device_type": 2 00:07:30.343 } 00:07:30.343 ], 00:07:30.343 "driver_specific": { 00:07:30.343 "raid": { 00:07:30.343 "uuid": "149fb2b5-ed82-4c95-8711-3125d7cec04f", 00:07:30.343 "strip_size_kb": 64, 00:07:30.343 "state": "online", 00:07:30.343 "raid_level": "raid0", 00:07:30.344 "superblock": true, 00:07:30.344 "num_base_bdevs": 3, 00:07:30.344 "num_base_bdevs_discovered": 3, 00:07:30.344 "num_base_bdevs_operational": 3, 00:07:30.344 "base_bdevs_list": [ 00:07:30.344 { 00:07:30.344 "name": "BaseBdev1", 00:07:30.344 "uuid": "38e62c0d-295f-4497-aad6-67d6851006c3", 00:07:30.344 "is_configured": true, 00:07:30.344 "data_offset": 2048, 00:07:30.344 "data_size": 63488 00:07:30.344 }, 00:07:30.344 { 00:07:30.344 "name": "BaseBdev2", 00:07:30.344 "uuid": "03ef36ad-03a2-493e-922d-02aae7bef418", 00:07:30.344 "is_configured": true, 00:07:30.344 "data_offset": 2048, 00:07:30.344 "data_size": 63488 00:07:30.344 }, 00:07:30.344 { 00:07:30.344 "name": "BaseBdev3", 00:07:30.344 "uuid": "f4794e2c-5e53-4903-bd7a-171440b2a6b6", 00:07:30.344 "is_configured": true, 00:07:30.344 "data_offset": 2048, 00:07:30.344 "data_size": 63488 00:07:30.344 } 00:07:30.344 ] 00:07:30.344 } 00:07:30.344 } 00:07:30.344 }' 00:07:30.344 23:41:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:07:30.604 23:41:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:07:30.604 BaseBdev2 00:07:30.604 BaseBdev3' 00:07:30.604 23:41:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:30.604 23:41:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:07:30.604 23:41:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:30.604 23:41:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:30.604 23:41:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:07:30.604 23:41:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:30.604 23:41:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:30.604 23:41:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:30.604 23:41:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:30.604 23:41:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:30.604 23:41:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:30.604 23:41:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:07:30.604 23:41:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:30.604 23:41:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:30.604 23:41:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:30.604 23:41:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:30.604 23:41:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:30.604 23:41:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:30.604 23:41:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:30.604 23:41:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:07:30.604 23:41:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:30.604 23:41:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:30.604 23:41:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:30.604 23:41:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:30.604 23:41:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:30.604 23:41:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:30.604 23:41:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:07:30.604 23:41:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:30.604 23:41:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:30.604 [2024-11-26 23:41:18.697121] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:07:30.604 [2024-11-26 23:41:18.697151] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:30.604 [2024-11-26 23:41:18.697210] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:30.604 23:41:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:30.604 23:41:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:07:30.604 23:41:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy raid0 00:07:30.604 23:41:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:07:30.604 23:41:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # return 1 00:07:30.604 23:41:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:07:30.604 23:41:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline raid0 64 2 00:07:30.604 23:41:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:30.604 23:41:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:07:30.604 23:41:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:30.604 23:41:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:30.604 23:41:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:30.604 23:41:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:30.604 23:41:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:30.604 23:41:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:30.604 23:41:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:30.604 23:41:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:30.604 23:41:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:30.604 23:41:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:30.604 23:41:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:30.864 23:41:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:30.864 23:41:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:30.864 "name": "Existed_Raid", 00:07:30.864 "uuid": "149fb2b5-ed82-4c95-8711-3125d7cec04f", 00:07:30.864 "strip_size_kb": 64, 00:07:30.864 "state": "offline", 00:07:30.864 "raid_level": "raid0", 00:07:30.864 "superblock": true, 00:07:30.864 "num_base_bdevs": 3, 00:07:30.864 "num_base_bdevs_discovered": 2, 00:07:30.864 "num_base_bdevs_operational": 2, 00:07:30.864 "base_bdevs_list": [ 00:07:30.864 { 00:07:30.864 "name": null, 00:07:30.864 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:30.864 "is_configured": false, 00:07:30.864 "data_offset": 0, 00:07:30.864 "data_size": 63488 00:07:30.864 }, 00:07:30.864 { 00:07:30.864 "name": "BaseBdev2", 00:07:30.864 "uuid": "03ef36ad-03a2-493e-922d-02aae7bef418", 00:07:30.864 "is_configured": true, 00:07:30.864 "data_offset": 2048, 00:07:30.864 "data_size": 63488 00:07:30.864 }, 00:07:30.864 { 00:07:30.864 "name": "BaseBdev3", 00:07:30.864 "uuid": "f4794e2c-5e53-4903-bd7a-171440b2a6b6", 00:07:30.864 "is_configured": true, 00:07:30.864 "data_offset": 2048, 00:07:30.864 "data_size": 63488 00:07:30.864 } 00:07:30.864 ] 00:07:30.864 }' 00:07:30.864 23:41:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:30.864 23:41:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:31.123 23:41:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:07:31.123 23:41:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:07:31.123 23:41:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:07:31.123 23:41:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:31.123 23:41:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:31.123 23:41:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:31.124 23:41:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:31.124 23:41:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:07:31.124 23:41:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:07:31.124 23:41:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:07:31.124 23:41:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:31.124 23:41:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:31.124 [2024-11-26 23:41:19.183913] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:07:31.124 23:41:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:31.124 23:41:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:07:31.124 23:41:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:07:31.124 23:41:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:07:31.124 23:41:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:31.124 23:41:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:31.124 23:41:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:31.124 23:41:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:31.124 23:41:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:07:31.124 23:41:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:07:31.124 23:41:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:07:31.124 23:41:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:31.124 23:41:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:31.383 [2024-11-26 23:41:19.255040] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:07:31.383 [2024-11-26 23:41:19.255147] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:07:31.383 23:41:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:31.383 23:41:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:07:31.383 23:41:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:07:31.383 23:41:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:31.383 23:41:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:31.383 23:41:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:31.383 23:41:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:07:31.383 23:41:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:31.383 23:41:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:07:31.383 23:41:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:07:31.383 23:41:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 3 -gt 2 ']' 00:07:31.383 23:41:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:07:31.383 23:41:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:07:31.383 23:41:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:07:31.383 23:41:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:31.383 23:41:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:31.383 BaseBdev2 00:07:31.383 23:41:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:31.383 23:41:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:07:31.383 23:41:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:07:31.383 23:41:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:07:31.383 23:41:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:07:31.383 23:41:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:07:31.383 23:41:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:07:31.383 23:41:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:07:31.383 23:41:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:31.383 23:41:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:31.383 23:41:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:31.383 23:41:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:07:31.383 23:41:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:31.383 23:41:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:31.383 [ 00:07:31.383 { 00:07:31.383 "name": "BaseBdev2", 00:07:31.383 "aliases": [ 00:07:31.383 "93f7819c-9cd4-4e31-b094-0ce39ef71b54" 00:07:31.383 ], 00:07:31.383 "product_name": "Malloc disk", 00:07:31.383 "block_size": 512, 00:07:31.383 "num_blocks": 65536, 00:07:31.383 "uuid": "93f7819c-9cd4-4e31-b094-0ce39ef71b54", 00:07:31.383 "assigned_rate_limits": { 00:07:31.383 "rw_ios_per_sec": 0, 00:07:31.383 "rw_mbytes_per_sec": 0, 00:07:31.383 "r_mbytes_per_sec": 0, 00:07:31.383 "w_mbytes_per_sec": 0 00:07:31.383 }, 00:07:31.383 "claimed": false, 00:07:31.383 "zoned": false, 00:07:31.383 "supported_io_types": { 00:07:31.383 "read": true, 00:07:31.383 "write": true, 00:07:31.383 "unmap": true, 00:07:31.383 "flush": true, 00:07:31.383 "reset": true, 00:07:31.383 "nvme_admin": false, 00:07:31.383 "nvme_io": false, 00:07:31.383 "nvme_io_md": false, 00:07:31.383 "write_zeroes": true, 00:07:31.383 "zcopy": true, 00:07:31.383 "get_zone_info": false, 00:07:31.383 "zone_management": false, 00:07:31.383 "zone_append": false, 00:07:31.383 "compare": false, 00:07:31.383 "compare_and_write": false, 00:07:31.383 "abort": true, 00:07:31.383 "seek_hole": false, 00:07:31.383 "seek_data": false, 00:07:31.383 "copy": true, 00:07:31.383 "nvme_iov_md": false 00:07:31.383 }, 00:07:31.383 "memory_domains": [ 00:07:31.383 { 00:07:31.383 "dma_device_id": "system", 00:07:31.383 "dma_device_type": 1 00:07:31.383 }, 00:07:31.383 { 00:07:31.383 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:31.383 "dma_device_type": 2 00:07:31.383 } 00:07:31.383 ], 00:07:31.383 "driver_specific": {} 00:07:31.383 } 00:07:31.383 ] 00:07:31.383 23:41:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:31.383 23:41:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:07:31.383 23:41:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:07:31.383 23:41:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:07:31.383 23:41:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:07:31.383 23:41:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:31.383 23:41:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:31.383 BaseBdev3 00:07:31.383 23:41:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:31.383 23:41:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:07:31.383 23:41:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:07:31.383 23:41:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:07:31.383 23:41:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:07:31.383 23:41:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:07:31.383 23:41:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:07:31.383 23:41:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:07:31.383 23:41:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:31.383 23:41:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:31.383 23:41:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:31.383 23:41:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:07:31.383 23:41:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:31.383 23:41:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:31.383 [ 00:07:31.383 { 00:07:31.383 "name": "BaseBdev3", 00:07:31.383 "aliases": [ 00:07:31.383 "1e8f9c9f-316a-4ab4-a75f-726bb6740431" 00:07:31.383 ], 00:07:31.383 "product_name": "Malloc disk", 00:07:31.383 "block_size": 512, 00:07:31.383 "num_blocks": 65536, 00:07:31.383 "uuid": "1e8f9c9f-316a-4ab4-a75f-726bb6740431", 00:07:31.383 "assigned_rate_limits": { 00:07:31.383 "rw_ios_per_sec": 0, 00:07:31.383 "rw_mbytes_per_sec": 0, 00:07:31.383 "r_mbytes_per_sec": 0, 00:07:31.383 "w_mbytes_per_sec": 0 00:07:31.383 }, 00:07:31.383 "claimed": false, 00:07:31.383 "zoned": false, 00:07:31.383 "supported_io_types": { 00:07:31.383 "read": true, 00:07:31.383 "write": true, 00:07:31.383 "unmap": true, 00:07:31.383 "flush": true, 00:07:31.383 "reset": true, 00:07:31.383 "nvme_admin": false, 00:07:31.383 "nvme_io": false, 00:07:31.383 "nvme_io_md": false, 00:07:31.383 "write_zeroes": true, 00:07:31.383 "zcopy": true, 00:07:31.383 "get_zone_info": false, 00:07:31.383 "zone_management": false, 00:07:31.383 "zone_append": false, 00:07:31.383 "compare": false, 00:07:31.383 "compare_and_write": false, 00:07:31.383 "abort": true, 00:07:31.383 "seek_hole": false, 00:07:31.383 "seek_data": false, 00:07:31.383 "copy": true, 00:07:31.383 "nvme_iov_md": false 00:07:31.383 }, 00:07:31.383 "memory_domains": [ 00:07:31.383 { 00:07:31.384 "dma_device_id": "system", 00:07:31.384 "dma_device_type": 1 00:07:31.384 }, 00:07:31.384 { 00:07:31.384 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:31.384 "dma_device_type": 2 00:07:31.384 } 00:07:31.384 ], 00:07:31.384 "driver_specific": {} 00:07:31.384 } 00:07:31.384 ] 00:07:31.384 23:41:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:31.384 23:41:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:07:31.384 23:41:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:07:31.384 23:41:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:07:31.384 23:41:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:07:31.384 23:41:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:31.384 23:41:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:31.384 [2024-11-26 23:41:19.431095] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:07:31.384 [2024-11-26 23:41:19.431228] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:07:31.384 [2024-11-26 23:41:19.431275] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:07:31.384 [2024-11-26 23:41:19.433144] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:07:31.384 23:41:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:31.384 23:41:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:07:31.384 23:41:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:31.384 23:41:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:31.384 23:41:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:31.384 23:41:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:31.384 23:41:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:07:31.384 23:41:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:31.384 23:41:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:31.384 23:41:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:31.384 23:41:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:31.384 23:41:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:31.384 23:41:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:31.384 23:41:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:31.384 23:41:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:31.384 23:41:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:31.384 23:41:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:31.384 "name": "Existed_Raid", 00:07:31.384 "uuid": "6e346487-78f6-4ca3-8331-530c0e45e621", 00:07:31.384 "strip_size_kb": 64, 00:07:31.384 "state": "configuring", 00:07:31.384 "raid_level": "raid0", 00:07:31.384 "superblock": true, 00:07:31.384 "num_base_bdevs": 3, 00:07:31.384 "num_base_bdevs_discovered": 2, 00:07:31.384 "num_base_bdevs_operational": 3, 00:07:31.384 "base_bdevs_list": [ 00:07:31.384 { 00:07:31.384 "name": "BaseBdev1", 00:07:31.384 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:31.384 "is_configured": false, 00:07:31.384 "data_offset": 0, 00:07:31.384 "data_size": 0 00:07:31.384 }, 00:07:31.384 { 00:07:31.384 "name": "BaseBdev2", 00:07:31.384 "uuid": "93f7819c-9cd4-4e31-b094-0ce39ef71b54", 00:07:31.384 "is_configured": true, 00:07:31.384 "data_offset": 2048, 00:07:31.384 "data_size": 63488 00:07:31.384 }, 00:07:31.384 { 00:07:31.384 "name": "BaseBdev3", 00:07:31.384 "uuid": "1e8f9c9f-316a-4ab4-a75f-726bb6740431", 00:07:31.384 "is_configured": true, 00:07:31.384 "data_offset": 2048, 00:07:31.384 "data_size": 63488 00:07:31.384 } 00:07:31.384 ] 00:07:31.384 }' 00:07:31.384 23:41:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:31.384 23:41:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:31.964 23:41:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:07:31.964 23:41:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:31.964 23:41:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:31.964 [2024-11-26 23:41:19.862408] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:07:31.964 23:41:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:31.964 23:41:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:07:31.964 23:41:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:31.964 23:41:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:31.964 23:41:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:31.964 23:41:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:31.964 23:41:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:07:31.964 23:41:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:31.964 23:41:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:31.964 23:41:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:31.964 23:41:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:31.964 23:41:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:31.964 23:41:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:31.964 23:41:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:31.964 23:41:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:31.964 23:41:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:31.964 23:41:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:31.964 "name": "Existed_Raid", 00:07:31.964 "uuid": "6e346487-78f6-4ca3-8331-530c0e45e621", 00:07:31.964 "strip_size_kb": 64, 00:07:31.964 "state": "configuring", 00:07:31.964 "raid_level": "raid0", 00:07:31.964 "superblock": true, 00:07:31.964 "num_base_bdevs": 3, 00:07:31.964 "num_base_bdevs_discovered": 1, 00:07:31.964 "num_base_bdevs_operational": 3, 00:07:31.964 "base_bdevs_list": [ 00:07:31.964 { 00:07:31.964 "name": "BaseBdev1", 00:07:31.964 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:31.964 "is_configured": false, 00:07:31.964 "data_offset": 0, 00:07:31.964 "data_size": 0 00:07:31.964 }, 00:07:31.964 { 00:07:31.964 "name": null, 00:07:31.964 "uuid": "93f7819c-9cd4-4e31-b094-0ce39ef71b54", 00:07:31.964 "is_configured": false, 00:07:31.964 "data_offset": 0, 00:07:31.964 "data_size": 63488 00:07:31.964 }, 00:07:31.964 { 00:07:31.964 "name": "BaseBdev3", 00:07:31.964 "uuid": "1e8f9c9f-316a-4ab4-a75f-726bb6740431", 00:07:31.964 "is_configured": true, 00:07:31.964 "data_offset": 2048, 00:07:31.964 "data_size": 63488 00:07:31.964 } 00:07:31.964 ] 00:07:31.964 }' 00:07:31.964 23:41:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:31.964 23:41:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:32.233 23:41:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:32.233 23:41:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:32.233 23:41:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:32.233 23:41:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:07:32.233 23:41:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:32.233 23:41:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:07:32.233 23:41:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:07:32.233 23:41:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:32.233 23:41:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:32.233 [2024-11-26 23:41:20.344861] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:32.233 BaseBdev1 00:07:32.233 23:41:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:32.233 23:41:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:07:32.233 23:41:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:07:32.233 23:41:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:07:32.233 23:41:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:07:32.233 23:41:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:07:32.233 23:41:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:07:32.234 23:41:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:07:32.234 23:41:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:32.234 23:41:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:32.508 23:41:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:32.508 23:41:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:07:32.508 23:41:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:32.508 23:41:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:32.508 [ 00:07:32.508 { 00:07:32.508 "name": "BaseBdev1", 00:07:32.508 "aliases": [ 00:07:32.508 "defb2327-d2e1-4997-808e-708a6d899cd5" 00:07:32.508 ], 00:07:32.508 "product_name": "Malloc disk", 00:07:32.508 "block_size": 512, 00:07:32.508 "num_blocks": 65536, 00:07:32.508 "uuid": "defb2327-d2e1-4997-808e-708a6d899cd5", 00:07:32.508 "assigned_rate_limits": { 00:07:32.508 "rw_ios_per_sec": 0, 00:07:32.508 "rw_mbytes_per_sec": 0, 00:07:32.508 "r_mbytes_per_sec": 0, 00:07:32.508 "w_mbytes_per_sec": 0 00:07:32.508 }, 00:07:32.508 "claimed": true, 00:07:32.508 "claim_type": "exclusive_write", 00:07:32.508 "zoned": false, 00:07:32.508 "supported_io_types": { 00:07:32.508 "read": true, 00:07:32.508 "write": true, 00:07:32.508 "unmap": true, 00:07:32.508 "flush": true, 00:07:32.508 "reset": true, 00:07:32.508 "nvme_admin": false, 00:07:32.508 "nvme_io": false, 00:07:32.508 "nvme_io_md": false, 00:07:32.508 "write_zeroes": true, 00:07:32.508 "zcopy": true, 00:07:32.508 "get_zone_info": false, 00:07:32.508 "zone_management": false, 00:07:32.508 "zone_append": false, 00:07:32.508 "compare": false, 00:07:32.508 "compare_and_write": false, 00:07:32.508 "abort": true, 00:07:32.508 "seek_hole": false, 00:07:32.508 "seek_data": false, 00:07:32.508 "copy": true, 00:07:32.508 "nvme_iov_md": false 00:07:32.508 }, 00:07:32.508 "memory_domains": [ 00:07:32.508 { 00:07:32.508 "dma_device_id": "system", 00:07:32.508 "dma_device_type": 1 00:07:32.508 }, 00:07:32.508 { 00:07:32.508 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:32.508 "dma_device_type": 2 00:07:32.508 } 00:07:32.508 ], 00:07:32.508 "driver_specific": {} 00:07:32.508 } 00:07:32.508 ] 00:07:32.508 23:41:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:32.508 23:41:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:07:32.508 23:41:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:07:32.508 23:41:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:32.509 23:41:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:32.509 23:41:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:32.509 23:41:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:32.509 23:41:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:07:32.509 23:41:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:32.509 23:41:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:32.509 23:41:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:32.509 23:41:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:32.509 23:41:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:32.509 23:41:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:32.509 23:41:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:32.509 23:41:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:32.509 23:41:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:32.509 23:41:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:32.509 "name": "Existed_Raid", 00:07:32.509 "uuid": "6e346487-78f6-4ca3-8331-530c0e45e621", 00:07:32.509 "strip_size_kb": 64, 00:07:32.509 "state": "configuring", 00:07:32.509 "raid_level": "raid0", 00:07:32.509 "superblock": true, 00:07:32.509 "num_base_bdevs": 3, 00:07:32.509 "num_base_bdevs_discovered": 2, 00:07:32.509 "num_base_bdevs_operational": 3, 00:07:32.509 "base_bdevs_list": [ 00:07:32.509 { 00:07:32.509 "name": "BaseBdev1", 00:07:32.509 "uuid": "defb2327-d2e1-4997-808e-708a6d899cd5", 00:07:32.509 "is_configured": true, 00:07:32.509 "data_offset": 2048, 00:07:32.509 "data_size": 63488 00:07:32.509 }, 00:07:32.509 { 00:07:32.509 "name": null, 00:07:32.509 "uuid": "93f7819c-9cd4-4e31-b094-0ce39ef71b54", 00:07:32.509 "is_configured": false, 00:07:32.509 "data_offset": 0, 00:07:32.509 "data_size": 63488 00:07:32.509 }, 00:07:32.509 { 00:07:32.509 "name": "BaseBdev3", 00:07:32.509 "uuid": "1e8f9c9f-316a-4ab4-a75f-726bb6740431", 00:07:32.509 "is_configured": true, 00:07:32.509 "data_offset": 2048, 00:07:32.509 "data_size": 63488 00:07:32.509 } 00:07:32.509 ] 00:07:32.509 }' 00:07:32.509 23:41:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:32.509 23:41:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:32.774 23:41:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:32.774 23:41:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:32.774 23:41:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:32.774 23:41:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:07:32.774 23:41:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:32.774 23:41:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:07:32.774 23:41:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:07:32.774 23:41:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:32.774 23:41:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:32.774 [2024-11-26 23:41:20.872049] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:07:32.774 23:41:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:32.774 23:41:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:07:32.774 23:41:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:32.774 23:41:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:32.774 23:41:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:32.774 23:41:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:32.774 23:41:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:07:32.774 23:41:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:32.774 23:41:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:32.774 23:41:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:32.774 23:41:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:32.774 23:41:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:32.774 23:41:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:32.774 23:41:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:32.774 23:41:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:32.774 23:41:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:33.034 23:41:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:33.034 "name": "Existed_Raid", 00:07:33.034 "uuid": "6e346487-78f6-4ca3-8331-530c0e45e621", 00:07:33.034 "strip_size_kb": 64, 00:07:33.034 "state": "configuring", 00:07:33.034 "raid_level": "raid0", 00:07:33.034 "superblock": true, 00:07:33.034 "num_base_bdevs": 3, 00:07:33.034 "num_base_bdevs_discovered": 1, 00:07:33.034 "num_base_bdevs_operational": 3, 00:07:33.034 "base_bdevs_list": [ 00:07:33.034 { 00:07:33.034 "name": "BaseBdev1", 00:07:33.034 "uuid": "defb2327-d2e1-4997-808e-708a6d899cd5", 00:07:33.034 "is_configured": true, 00:07:33.034 "data_offset": 2048, 00:07:33.034 "data_size": 63488 00:07:33.034 }, 00:07:33.034 { 00:07:33.034 "name": null, 00:07:33.034 "uuid": "93f7819c-9cd4-4e31-b094-0ce39ef71b54", 00:07:33.034 "is_configured": false, 00:07:33.034 "data_offset": 0, 00:07:33.034 "data_size": 63488 00:07:33.034 }, 00:07:33.034 { 00:07:33.034 "name": null, 00:07:33.034 "uuid": "1e8f9c9f-316a-4ab4-a75f-726bb6740431", 00:07:33.034 "is_configured": false, 00:07:33.034 "data_offset": 0, 00:07:33.034 "data_size": 63488 00:07:33.034 } 00:07:33.034 ] 00:07:33.034 }' 00:07:33.034 23:41:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:33.034 23:41:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:33.304 23:41:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:33.305 23:41:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:33.305 23:41:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:07:33.305 23:41:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:33.305 23:41:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:33.305 23:41:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:07:33.305 23:41:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:07:33.305 23:41:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:33.305 23:41:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:33.305 [2024-11-26 23:41:21.339259] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:07:33.305 23:41:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:33.305 23:41:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:07:33.305 23:41:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:33.305 23:41:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:33.305 23:41:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:33.305 23:41:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:33.305 23:41:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:07:33.305 23:41:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:33.305 23:41:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:33.305 23:41:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:33.305 23:41:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:33.305 23:41:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:33.305 23:41:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:33.305 23:41:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:33.305 23:41:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:33.305 23:41:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:33.305 23:41:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:33.305 "name": "Existed_Raid", 00:07:33.305 "uuid": "6e346487-78f6-4ca3-8331-530c0e45e621", 00:07:33.305 "strip_size_kb": 64, 00:07:33.305 "state": "configuring", 00:07:33.305 "raid_level": "raid0", 00:07:33.305 "superblock": true, 00:07:33.305 "num_base_bdevs": 3, 00:07:33.305 "num_base_bdevs_discovered": 2, 00:07:33.305 "num_base_bdevs_operational": 3, 00:07:33.305 "base_bdevs_list": [ 00:07:33.305 { 00:07:33.305 "name": "BaseBdev1", 00:07:33.305 "uuid": "defb2327-d2e1-4997-808e-708a6d899cd5", 00:07:33.305 "is_configured": true, 00:07:33.305 "data_offset": 2048, 00:07:33.305 "data_size": 63488 00:07:33.305 }, 00:07:33.305 { 00:07:33.305 "name": null, 00:07:33.305 "uuid": "93f7819c-9cd4-4e31-b094-0ce39ef71b54", 00:07:33.306 "is_configured": false, 00:07:33.306 "data_offset": 0, 00:07:33.306 "data_size": 63488 00:07:33.306 }, 00:07:33.306 { 00:07:33.306 "name": "BaseBdev3", 00:07:33.306 "uuid": "1e8f9c9f-316a-4ab4-a75f-726bb6740431", 00:07:33.306 "is_configured": true, 00:07:33.306 "data_offset": 2048, 00:07:33.306 "data_size": 63488 00:07:33.306 } 00:07:33.306 ] 00:07:33.306 }' 00:07:33.306 23:41:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:33.306 23:41:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:33.898 23:41:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:07:33.898 23:41:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:33.898 23:41:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:33.898 23:41:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:33.898 23:41:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:33.898 23:41:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:07:33.898 23:41:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:07:33.898 23:41:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:33.898 23:41:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:33.898 [2024-11-26 23:41:21.854417] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:07:33.898 23:41:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:33.898 23:41:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:07:33.898 23:41:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:33.898 23:41:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:33.898 23:41:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:33.898 23:41:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:33.898 23:41:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:07:33.898 23:41:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:33.898 23:41:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:33.898 23:41:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:33.898 23:41:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:33.898 23:41:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:33.898 23:41:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:33.898 23:41:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:33.898 23:41:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:33.898 23:41:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:33.898 23:41:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:33.898 "name": "Existed_Raid", 00:07:33.898 "uuid": "6e346487-78f6-4ca3-8331-530c0e45e621", 00:07:33.898 "strip_size_kb": 64, 00:07:33.898 "state": "configuring", 00:07:33.898 "raid_level": "raid0", 00:07:33.898 "superblock": true, 00:07:33.898 "num_base_bdevs": 3, 00:07:33.898 "num_base_bdevs_discovered": 1, 00:07:33.898 "num_base_bdevs_operational": 3, 00:07:33.898 "base_bdevs_list": [ 00:07:33.898 { 00:07:33.898 "name": null, 00:07:33.899 "uuid": "defb2327-d2e1-4997-808e-708a6d899cd5", 00:07:33.899 "is_configured": false, 00:07:33.899 "data_offset": 0, 00:07:33.899 "data_size": 63488 00:07:33.899 }, 00:07:33.899 { 00:07:33.899 "name": null, 00:07:33.899 "uuid": "93f7819c-9cd4-4e31-b094-0ce39ef71b54", 00:07:33.899 "is_configured": false, 00:07:33.899 "data_offset": 0, 00:07:33.899 "data_size": 63488 00:07:33.899 }, 00:07:33.899 { 00:07:33.899 "name": "BaseBdev3", 00:07:33.899 "uuid": "1e8f9c9f-316a-4ab4-a75f-726bb6740431", 00:07:33.899 "is_configured": true, 00:07:33.899 "data_offset": 2048, 00:07:33.899 "data_size": 63488 00:07:33.899 } 00:07:33.899 ] 00:07:33.899 }' 00:07:33.899 23:41:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:33.899 23:41:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:34.467 23:41:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:07:34.467 23:41:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:34.467 23:41:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:34.467 23:41:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:34.467 23:41:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:34.467 23:41:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:07:34.467 23:41:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:07:34.467 23:41:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:34.467 23:41:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:34.467 [2024-11-26 23:41:22.343969] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:07:34.467 23:41:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:34.467 23:41:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:07:34.467 23:41:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:34.467 23:41:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:34.467 23:41:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:34.467 23:41:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:34.467 23:41:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:07:34.467 23:41:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:34.467 23:41:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:34.467 23:41:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:34.467 23:41:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:34.467 23:41:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:34.467 23:41:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:34.467 23:41:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:34.467 23:41:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:34.467 23:41:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:34.468 23:41:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:34.468 "name": "Existed_Raid", 00:07:34.468 "uuid": "6e346487-78f6-4ca3-8331-530c0e45e621", 00:07:34.468 "strip_size_kb": 64, 00:07:34.468 "state": "configuring", 00:07:34.468 "raid_level": "raid0", 00:07:34.468 "superblock": true, 00:07:34.468 "num_base_bdevs": 3, 00:07:34.468 "num_base_bdevs_discovered": 2, 00:07:34.468 "num_base_bdevs_operational": 3, 00:07:34.468 "base_bdevs_list": [ 00:07:34.468 { 00:07:34.468 "name": null, 00:07:34.468 "uuid": "defb2327-d2e1-4997-808e-708a6d899cd5", 00:07:34.468 "is_configured": false, 00:07:34.468 "data_offset": 0, 00:07:34.468 "data_size": 63488 00:07:34.468 }, 00:07:34.468 { 00:07:34.468 "name": "BaseBdev2", 00:07:34.468 "uuid": "93f7819c-9cd4-4e31-b094-0ce39ef71b54", 00:07:34.468 "is_configured": true, 00:07:34.468 "data_offset": 2048, 00:07:34.468 "data_size": 63488 00:07:34.468 }, 00:07:34.468 { 00:07:34.468 "name": "BaseBdev3", 00:07:34.468 "uuid": "1e8f9c9f-316a-4ab4-a75f-726bb6740431", 00:07:34.468 "is_configured": true, 00:07:34.468 "data_offset": 2048, 00:07:34.468 "data_size": 63488 00:07:34.468 } 00:07:34.468 ] 00:07:34.468 }' 00:07:34.468 23:41:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:34.468 23:41:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:34.728 23:41:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:07:34.728 23:41:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:34.728 23:41:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:34.728 23:41:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:34.728 23:41:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:34.728 23:41:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:07:34.728 23:41:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:07:34.728 23:41:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:34.728 23:41:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:34.728 23:41:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:34.986 23:41:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:34.986 23:41:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u defb2327-d2e1-4997-808e-708a6d899cd5 00:07:34.986 23:41:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:34.986 23:41:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:34.986 NewBaseBdev 00:07:34.986 [2024-11-26 23:41:22.901836] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:07:34.986 [2024-11-26 23:41:22.901994] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:07:34.986 [2024-11-26 23:41:22.902010] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:07:34.986 [2024-11-26 23:41:22.902264] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002870 00:07:34.986 [2024-11-26 23:41:22.902393] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:07:34.986 [2024-11-26 23:41:22.902403] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001c80 00:07:34.986 [2024-11-26 23:41:22.902506] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:34.986 23:41:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:34.986 23:41:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:07:34.986 23:41:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:07:34.987 23:41:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:07:34.987 23:41:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:07:34.987 23:41:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:07:34.987 23:41:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:07:34.987 23:41:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:07:34.987 23:41:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:34.987 23:41:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:34.987 23:41:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:34.987 23:41:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:07:34.987 23:41:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:34.987 23:41:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:34.987 [ 00:07:34.987 { 00:07:34.987 "name": "NewBaseBdev", 00:07:34.987 "aliases": [ 00:07:34.987 "defb2327-d2e1-4997-808e-708a6d899cd5" 00:07:34.987 ], 00:07:34.987 "product_name": "Malloc disk", 00:07:34.987 "block_size": 512, 00:07:34.987 "num_blocks": 65536, 00:07:34.987 "uuid": "defb2327-d2e1-4997-808e-708a6d899cd5", 00:07:34.987 "assigned_rate_limits": { 00:07:34.987 "rw_ios_per_sec": 0, 00:07:34.987 "rw_mbytes_per_sec": 0, 00:07:34.987 "r_mbytes_per_sec": 0, 00:07:34.987 "w_mbytes_per_sec": 0 00:07:34.987 }, 00:07:34.987 "claimed": true, 00:07:34.987 "claim_type": "exclusive_write", 00:07:34.987 "zoned": false, 00:07:34.987 "supported_io_types": { 00:07:34.987 "read": true, 00:07:34.987 "write": true, 00:07:34.987 "unmap": true, 00:07:34.987 "flush": true, 00:07:34.987 "reset": true, 00:07:34.987 "nvme_admin": false, 00:07:34.987 "nvme_io": false, 00:07:34.987 "nvme_io_md": false, 00:07:34.987 "write_zeroes": true, 00:07:34.987 "zcopy": true, 00:07:34.987 "get_zone_info": false, 00:07:34.987 "zone_management": false, 00:07:34.987 "zone_append": false, 00:07:34.987 "compare": false, 00:07:34.987 "compare_and_write": false, 00:07:34.987 "abort": true, 00:07:34.987 "seek_hole": false, 00:07:34.987 "seek_data": false, 00:07:34.987 "copy": true, 00:07:34.987 "nvme_iov_md": false 00:07:34.987 }, 00:07:34.987 "memory_domains": [ 00:07:34.987 { 00:07:34.987 "dma_device_id": "system", 00:07:34.987 "dma_device_type": 1 00:07:34.987 }, 00:07:34.987 { 00:07:34.987 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:34.987 "dma_device_type": 2 00:07:34.987 } 00:07:34.987 ], 00:07:34.987 "driver_specific": {} 00:07:34.987 } 00:07:34.987 ] 00:07:34.987 23:41:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:34.987 23:41:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:07:34.987 23:41:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid0 64 3 00:07:34.987 23:41:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:34.987 23:41:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:34.987 23:41:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:34.987 23:41:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:34.987 23:41:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:07:34.987 23:41:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:34.987 23:41:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:34.987 23:41:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:34.987 23:41:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:34.987 23:41:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:34.987 23:41:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:34.987 23:41:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:34.987 23:41:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:34.987 23:41:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:34.987 23:41:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:34.987 "name": "Existed_Raid", 00:07:34.987 "uuid": "6e346487-78f6-4ca3-8331-530c0e45e621", 00:07:34.987 "strip_size_kb": 64, 00:07:34.987 "state": "online", 00:07:34.987 "raid_level": "raid0", 00:07:34.987 "superblock": true, 00:07:34.987 "num_base_bdevs": 3, 00:07:34.987 "num_base_bdevs_discovered": 3, 00:07:34.987 "num_base_bdevs_operational": 3, 00:07:34.987 "base_bdevs_list": [ 00:07:34.987 { 00:07:34.987 "name": "NewBaseBdev", 00:07:34.987 "uuid": "defb2327-d2e1-4997-808e-708a6d899cd5", 00:07:34.987 "is_configured": true, 00:07:34.987 "data_offset": 2048, 00:07:34.987 "data_size": 63488 00:07:34.987 }, 00:07:34.987 { 00:07:34.987 "name": "BaseBdev2", 00:07:34.987 "uuid": "93f7819c-9cd4-4e31-b094-0ce39ef71b54", 00:07:34.987 "is_configured": true, 00:07:34.987 "data_offset": 2048, 00:07:34.987 "data_size": 63488 00:07:34.987 }, 00:07:34.987 { 00:07:34.987 "name": "BaseBdev3", 00:07:34.987 "uuid": "1e8f9c9f-316a-4ab4-a75f-726bb6740431", 00:07:34.987 "is_configured": true, 00:07:34.987 "data_offset": 2048, 00:07:34.987 "data_size": 63488 00:07:34.987 } 00:07:34.987 ] 00:07:34.987 }' 00:07:34.987 23:41:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:34.987 23:41:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:35.244 23:41:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:07:35.244 23:41:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:07:35.244 23:41:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:07:35.244 23:41:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:07:35.244 23:41:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:07:35.245 23:41:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:07:35.245 23:41:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:07:35.245 23:41:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:35.245 23:41:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:35.245 23:41:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:07:35.245 [2024-11-26 23:41:23.357457] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:35.504 23:41:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:35.504 23:41:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:07:35.504 "name": "Existed_Raid", 00:07:35.504 "aliases": [ 00:07:35.504 "6e346487-78f6-4ca3-8331-530c0e45e621" 00:07:35.504 ], 00:07:35.504 "product_name": "Raid Volume", 00:07:35.504 "block_size": 512, 00:07:35.504 "num_blocks": 190464, 00:07:35.504 "uuid": "6e346487-78f6-4ca3-8331-530c0e45e621", 00:07:35.504 "assigned_rate_limits": { 00:07:35.504 "rw_ios_per_sec": 0, 00:07:35.504 "rw_mbytes_per_sec": 0, 00:07:35.504 "r_mbytes_per_sec": 0, 00:07:35.504 "w_mbytes_per_sec": 0 00:07:35.504 }, 00:07:35.504 "claimed": false, 00:07:35.504 "zoned": false, 00:07:35.504 "supported_io_types": { 00:07:35.504 "read": true, 00:07:35.504 "write": true, 00:07:35.504 "unmap": true, 00:07:35.504 "flush": true, 00:07:35.504 "reset": true, 00:07:35.504 "nvme_admin": false, 00:07:35.504 "nvme_io": false, 00:07:35.504 "nvme_io_md": false, 00:07:35.504 "write_zeroes": true, 00:07:35.504 "zcopy": false, 00:07:35.504 "get_zone_info": false, 00:07:35.504 "zone_management": false, 00:07:35.504 "zone_append": false, 00:07:35.504 "compare": false, 00:07:35.504 "compare_and_write": false, 00:07:35.504 "abort": false, 00:07:35.504 "seek_hole": false, 00:07:35.504 "seek_data": false, 00:07:35.504 "copy": false, 00:07:35.504 "nvme_iov_md": false 00:07:35.504 }, 00:07:35.504 "memory_domains": [ 00:07:35.504 { 00:07:35.504 "dma_device_id": "system", 00:07:35.504 "dma_device_type": 1 00:07:35.504 }, 00:07:35.504 { 00:07:35.504 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:35.504 "dma_device_type": 2 00:07:35.504 }, 00:07:35.504 { 00:07:35.504 "dma_device_id": "system", 00:07:35.504 "dma_device_type": 1 00:07:35.504 }, 00:07:35.504 { 00:07:35.504 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:35.504 "dma_device_type": 2 00:07:35.504 }, 00:07:35.504 { 00:07:35.504 "dma_device_id": "system", 00:07:35.504 "dma_device_type": 1 00:07:35.504 }, 00:07:35.504 { 00:07:35.504 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:35.504 "dma_device_type": 2 00:07:35.504 } 00:07:35.504 ], 00:07:35.504 "driver_specific": { 00:07:35.504 "raid": { 00:07:35.504 "uuid": "6e346487-78f6-4ca3-8331-530c0e45e621", 00:07:35.504 "strip_size_kb": 64, 00:07:35.504 "state": "online", 00:07:35.504 "raid_level": "raid0", 00:07:35.504 "superblock": true, 00:07:35.504 "num_base_bdevs": 3, 00:07:35.504 "num_base_bdevs_discovered": 3, 00:07:35.504 "num_base_bdevs_operational": 3, 00:07:35.504 "base_bdevs_list": [ 00:07:35.504 { 00:07:35.504 "name": "NewBaseBdev", 00:07:35.504 "uuid": "defb2327-d2e1-4997-808e-708a6d899cd5", 00:07:35.504 "is_configured": true, 00:07:35.504 "data_offset": 2048, 00:07:35.504 "data_size": 63488 00:07:35.504 }, 00:07:35.504 { 00:07:35.504 "name": "BaseBdev2", 00:07:35.504 "uuid": "93f7819c-9cd4-4e31-b094-0ce39ef71b54", 00:07:35.504 "is_configured": true, 00:07:35.504 "data_offset": 2048, 00:07:35.504 "data_size": 63488 00:07:35.504 }, 00:07:35.504 { 00:07:35.504 "name": "BaseBdev3", 00:07:35.504 "uuid": "1e8f9c9f-316a-4ab4-a75f-726bb6740431", 00:07:35.504 "is_configured": true, 00:07:35.504 "data_offset": 2048, 00:07:35.504 "data_size": 63488 00:07:35.504 } 00:07:35.504 ] 00:07:35.504 } 00:07:35.504 } 00:07:35.504 }' 00:07:35.504 23:41:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:07:35.504 23:41:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:07:35.504 BaseBdev2 00:07:35.504 BaseBdev3' 00:07:35.504 23:41:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:35.504 23:41:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:07:35.504 23:41:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:35.504 23:41:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:07:35.504 23:41:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:35.504 23:41:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:35.504 23:41:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:35.504 23:41:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:35.504 23:41:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:35.504 23:41:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:35.504 23:41:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:35.504 23:41:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:07:35.504 23:41:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:35.504 23:41:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:35.504 23:41:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:35.504 23:41:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:35.504 23:41:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:35.504 23:41:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:35.504 23:41:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:35.504 23:41:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:07:35.504 23:41:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:35.504 23:41:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:35.504 23:41:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:35.504 23:41:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:35.764 23:41:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:35.764 23:41:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:35.764 23:41:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:07:35.765 23:41:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:35.765 23:41:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:35.765 [2024-11-26 23:41:23.640644] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:07:35.765 [2024-11-26 23:41:23.640675] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:35.765 [2024-11-26 23:41:23.640756] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:35.765 [2024-11-26 23:41:23.640809] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:35.765 [2024-11-26 23:41:23.640827] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name Existed_Raid, state offline 00:07:35.765 23:41:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:35.765 23:41:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 75353 00:07:35.765 23:41:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # '[' -z 75353 ']' 00:07:35.765 23:41:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@958 -- # kill -0 75353 00:07:35.765 23:41:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # uname 00:07:35.765 23:41:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:07:35.765 23:41:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 75353 00:07:35.765 23:41:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:07:35.765 23:41:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:07:35.765 killing process with pid 75353 00:07:35.765 23:41:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@972 -- # echo 'killing process with pid 75353' 00:07:35.765 23:41:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@973 -- # kill 75353 00:07:35.765 [2024-11-26 23:41:23.681521] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:35.765 23:41:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@978 -- # wait 75353 00:07:35.765 [2024-11-26 23:41:23.712224] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:36.024 23:41:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:07:36.024 00:07:36.024 real 0m8.903s 00:07:36.024 user 0m15.294s 00:07:36.024 sys 0m1.749s 00:07:36.024 23:41:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1130 -- # xtrace_disable 00:07:36.024 23:41:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:36.024 ************************************ 00:07:36.024 END TEST raid_state_function_test_sb 00:07:36.024 ************************************ 00:07:36.024 23:41:23 bdev_raid -- bdev/bdev_raid.sh@970 -- # run_test raid_superblock_test raid_superblock_test raid0 3 00:07:36.024 23:41:23 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:07:36.024 23:41:23 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:36.024 23:41:23 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:36.024 ************************************ 00:07:36.024 START TEST raid_superblock_test 00:07:36.024 ************************************ 00:07:36.024 23:41:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1129 -- # raid_superblock_test raid0 3 00:07:36.024 23:41:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=raid0 00:07:36.024 23:41:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=3 00:07:36.024 23:41:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:07:36.024 23:41:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:07:36.024 23:41:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:07:36.024 23:41:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:07:36.024 23:41:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:07:36.024 23:41:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:07:36.024 23:41:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:07:36.024 23:41:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:07:36.024 23:41:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:07:36.024 23:41:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:07:36.024 23:41:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:07:36.024 23:41:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' raid0 '!=' raid1 ']' 00:07:36.024 23:41:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@405 -- # strip_size=64 00:07:36.024 23:41:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@406 -- # strip_size_create_arg='-z 64' 00:07:36.024 23:41:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=75957 00:07:36.024 23:41:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:07:36.024 23:41:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 75957 00:07:36.024 23:41:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@835 -- # '[' -z 75957 ']' 00:07:36.024 23:41:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:36.024 23:41:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:36.024 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:36.024 23:41:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:36.024 23:41:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:36.024 23:41:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:36.024 [2024-11-26 23:41:24.081056] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:07:36.024 [2024-11-26 23:41:24.081195] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid75957 ] 00:07:36.283 [2024-11-26 23:41:24.234716] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:36.283 [2024-11-26 23:41:24.260065] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:36.283 [2024-11-26 23:41:24.301512] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:36.283 [2024-11-26 23:41:24.301554] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:36.853 23:41:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:36.853 23:41:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@868 -- # return 0 00:07:36.853 23:41:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:07:36.853 23:41:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:07:36.853 23:41:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:07:36.853 23:41:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:07:36.853 23:41:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:07:36.853 23:41:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:07:36.853 23:41:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:07:36.853 23:41:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:07:36.853 23:41:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:07:36.853 23:41:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:36.853 23:41:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:36.853 malloc1 00:07:36.853 23:41:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:36.853 23:41:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:07:36.853 23:41:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:36.853 23:41:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:36.853 [2024-11-26 23:41:24.928421] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:07:36.853 [2024-11-26 23:41:24.928504] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:36.853 [2024-11-26 23:41:24.928524] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:07:36.853 [2024-11-26 23:41:24.928539] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:36.853 [2024-11-26 23:41:24.930648] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:36.853 [2024-11-26 23:41:24.930689] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:07:36.853 pt1 00:07:36.853 23:41:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:36.853 23:41:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:07:36.853 23:41:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:07:36.853 23:41:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:07:36.853 23:41:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:07:36.853 23:41:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:07:36.853 23:41:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:07:36.853 23:41:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:07:36.853 23:41:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:07:36.853 23:41:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:07:36.853 23:41:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:36.853 23:41:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:36.853 malloc2 00:07:36.853 23:41:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:36.853 23:41:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:07:36.853 23:41:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:36.853 23:41:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:36.853 [2024-11-26 23:41:24.956780] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:07:36.853 [2024-11-26 23:41:24.956838] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:36.853 [2024-11-26 23:41:24.956856] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:07:36.853 [2024-11-26 23:41:24.956866] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:36.853 [2024-11-26 23:41:24.958990] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:36.853 [2024-11-26 23:41:24.959028] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:07:36.853 pt2 00:07:36.853 23:41:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:36.853 23:41:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:07:36.853 23:41:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:07:36.853 23:41:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc3 00:07:36.853 23:41:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt3 00:07:36.853 23:41:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000003 00:07:36.853 23:41:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:07:36.853 23:41:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:07:36.853 23:41:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:07:36.853 23:41:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc3 00:07:36.853 23:41:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:36.853 23:41:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:36.853 malloc3 00:07:36.853 23:41:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:36.853 23:41:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:07:36.853 23:41:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:36.853 23:41:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:37.113 [2024-11-26 23:41:24.985010] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:07:37.113 [2024-11-26 23:41:24.985077] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:37.113 [2024-11-26 23:41:24.985094] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:07:37.113 [2024-11-26 23:41:24.985104] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:37.113 [2024-11-26 23:41:24.987165] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:37.113 [2024-11-26 23:41:24.987203] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:07:37.113 pt3 00:07:37.113 23:41:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:37.113 23:41:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:07:37.113 23:41:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:07:37.113 23:41:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''pt1 pt2 pt3'\''' -n raid_bdev1 -s 00:07:37.113 23:41:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:37.113 23:41:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:37.113 [2024-11-26 23:41:24.997050] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:07:37.113 [2024-11-26 23:41:24.998840] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:07:37.113 [2024-11-26 23:41:24.998899] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:07:37.113 [2024-11-26 23:41:24.999047] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:07:37.113 [2024-11-26 23:41:24.999066] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:07:37.113 [2024-11-26 23:41:24.999304] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:07:37.113 [2024-11-26 23:41:24.999438] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:07:37.113 [2024-11-26 23:41:24.999452] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:07:37.113 [2024-11-26 23:41:24.999562] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:37.113 23:41:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:37.113 23:41:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 3 00:07:37.113 23:41:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:37.113 23:41:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:37.113 23:41:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:37.113 23:41:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:37.113 23:41:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:07:37.113 23:41:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:37.113 23:41:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:37.113 23:41:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:37.113 23:41:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:37.113 23:41:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:37.113 23:41:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:37.113 23:41:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:37.113 23:41:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:37.113 23:41:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:37.113 23:41:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:37.113 "name": "raid_bdev1", 00:07:37.113 "uuid": "cbb0c36d-9b9b-4515-9d9d-545a0f139f8c", 00:07:37.113 "strip_size_kb": 64, 00:07:37.113 "state": "online", 00:07:37.113 "raid_level": "raid0", 00:07:37.113 "superblock": true, 00:07:37.113 "num_base_bdevs": 3, 00:07:37.113 "num_base_bdevs_discovered": 3, 00:07:37.113 "num_base_bdevs_operational": 3, 00:07:37.113 "base_bdevs_list": [ 00:07:37.113 { 00:07:37.113 "name": "pt1", 00:07:37.113 "uuid": "00000000-0000-0000-0000-000000000001", 00:07:37.113 "is_configured": true, 00:07:37.113 "data_offset": 2048, 00:07:37.113 "data_size": 63488 00:07:37.113 }, 00:07:37.113 { 00:07:37.113 "name": "pt2", 00:07:37.113 "uuid": "00000000-0000-0000-0000-000000000002", 00:07:37.113 "is_configured": true, 00:07:37.113 "data_offset": 2048, 00:07:37.113 "data_size": 63488 00:07:37.113 }, 00:07:37.113 { 00:07:37.113 "name": "pt3", 00:07:37.113 "uuid": "00000000-0000-0000-0000-000000000003", 00:07:37.113 "is_configured": true, 00:07:37.113 "data_offset": 2048, 00:07:37.113 "data_size": 63488 00:07:37.113 } 00:07:37.113 ] 00:07:37.113 }' 00:07:37.113 23:41:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:37.113 23:41:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:37.373 23:41:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:07:37.373 23:41:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:07:37.373 23:41:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:07:37.373 23:41:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:07:37.373 23:41:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:07:37.373 23:41:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:07:37.373 23:41:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:07:37.373 23:41:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:37.373 23:41:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:37.373 23:41:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:07:37.373 [2024-11-26 23:41:25.420578] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:37.373 23:41:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:37.373 23:41:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:07:37.373 "name": "raid_bdev1", 00:07:37.373 "aliases": [ 00:07:37.373 "cbb0c36d-9b9b-4515-9d9d-545a0f139f8c" 00:07:37.373 ], 00:07:37.373 "product_name": "Raid Volume", 00:07:37.373 "block_size": 512, 00:07:37.373 "num_blocks": 190464, 00:07:37.373 "uuid": "cbb0c36d-9b9b-4515-9d9d-545a0f139f8c", 00:07:37.373 "assigned_rate_limits": { 00:07:37.373 "rw_ios_per_sec": 0, 00:07:37.373 "rw_mbytes_per_sec": 0, 00:07:37.373 "r_mbytes_per_sec": 0, 00:07:37.373 "w_mbytes_per_sec": 0 00:07:37.373 }, 00:07:37.373 "claimed": false, 00:07:37.373 "zoned": false, 00:07:37.373 "supported_io_types": { 00:07:37.373 "read": true, 00:07:37.373 "write": true, 00:07:37.373 "unmap": true, 00:07:37.373 "flush": true, 00:07:37.373 "reset": true, 00:07:37.373 "nvme_admin": false, 00:07:37.374 "nvme_io": false, 00:07:37.374 "nvme_io_md": false, 00:07:37.374 "write_zeroes": true, 00:07:37.374 "zcopy": false, 00:07:37.374 "get_zone_info": false, 00:07:37.374 "zone_management": false, 00:07:37.374 "zone_append": false, 00:07:37.374 "compare": false, 00:07:37.374 "compare_and_write": false, 00:07:37.374 "abort": false, 00:07:37.374 "seek_hole": false, 00:07:37.374 "seek_data": false, 00:07:37.374 "copy": false, 00:07:37.374 "nvme_iov_md": false 00:07:37.374 }, 00:07:37.374 "memory_domains": [ 00:07:37.374 { 00:07:37.374 "dma_device_id": "system", 00:07:37.374 "dma_device_type": 1 00:07:37.374 }, 00:07:37.374 { 00:07:37.374 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:37.374 "dma_device_type": 2 00:07:37.374 }, 00:07:37.374 { 00:07:37.374 "dma_device_id": "system", 00:07:37.374 "dma_device_type": 1 00:07:37.374 }, 00:07:37.374 { 00:07:37.374 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:37.374 "dma_device_type": 2 00:07:37.374 }, 00:07:37.374 { 00:07:37.374 "dma_device_id": "system", 00:07:37.374 "dma_device_type": 1 00:07:37.374 }, 00:07:37.374 { 00:07:37.374 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:37.374 "dma_device_type": 2 00:07:37.374 } 00:07:37.374 ], 00:07:37.374 "driver_specific": { 00:07:37.374 "raid": { 00:07:37.374 "uuid": "cbb0c36d-9b9b-4515-9d9d-545a0f139f8c", 00:07:37.374 "strip_size_kb": 64, 00:07:37.374 "state": "online", 00:07:37.374 "raid_level": "raid0", 00:07:37.374 "superblock": true, 00:07:37.374 "num_base_bdevs": 3, 00:07:37.374 "num_base_bdevs_discovered": 3, 00:07:37.374 "num_base_bdevs_operational": 3, 00:07:37.374 "base_bdevs_list": [ 00:07:37.374 { 00:07:37.374 "name": "pt1", 00:07:37.374 "uuid": "00000000-0000-0000-0000-000000000001", 00:07:37.374 "is_configured": true, 00:07:37.374 "data_offset": 2048, 00:07:37.374 "data_size": 63488 00:07:37.374 }, 00:07:37.374 { 00:07:37.374 "name": "pt2", 00:07:37.374 "uuid": "00000000-0000-0000-0000-000000000002", 00:07:37.374 "is_configured": true, 00:07:37.374 "data_offset": 2048, 00:07:37.374 "data_size": 63488 00:07:37.374 }, 00:07:37.374 { 00:07:37.374 "name": "pt3", 00:07:37.374 "uuid": "00000000-0000-0000-0000-000000000003", 00:07:37.374 "is_configured": true, 00:07:37.374 "data_offset": 2048, 00:07:37.374 "data_size": 63488 00:07:37.374 } 00:07:37.374 ] 00:07:37.374 } 00:07:37.374 } 00:07:37.374 }' 00:07:37.374 23:41:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:07:37.637 23:41:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:07:37.637 pt2 00:07:37.637 pt3' 00:07:37.637 23:41:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:37.637 23:41:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:07:37.637 23:41:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:37.637 23:41:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:37.637 23:41:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:07:37.637 23:41:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:37.637 23:41:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:37.637 23:41:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:37.637 23:41:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:37.637 23:41:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:37.637 23:41:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:37.637 23:41:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:07:37.637 23:41:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:37.637 23:41:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:37.637 23:41:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:37.637 23:41:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:37.637 23:41:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:37.637 23:41:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:37.637 23:41:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:37.637 23:41:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:37.637 23:41:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:07:37.637 23:41:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:37.637 23:41:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:37.637 23:41:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:37.637 23:41:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:37.637 23:41:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:37.637 23:41:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:07:37.637 23:41:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:07:37.638 23:41:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:37.638 23:41:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:37.638 [2024-11-26 23:41:25.668076] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:37.638 23:41:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:37.638 23:41:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=cbb0c36d-9b9b-4515-9d9d-545a0f139f8c 00:07:37.638 23:41:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z cbb0c36d-9b9b-4515-9d9d-545a0f139f8c ']' 00:07:37.638 23:41:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:07:37.638 23:41:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:37.638 23:41:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:37.638 [2024-11-26 23:41:25.711749] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:07:37.638 [2024-11-26 23:41:25.711783] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:37.638 [2024-11-26 23:41:25.711859] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:37.638 [2024-11-26 23:41:25.711914] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:37.638 [2024-11-26 23:41:25.711924] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:07:37.638 23:41:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:37.638 23:41:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:37.638 23:41:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:37.638 23:41:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:07:37.638 23:41:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:37.638 23:41:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:37.908 23:41:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:07:37.908 23:41:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:07:37.908 23:41:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:07:37.908 23:41:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:07:37.908 23:41:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:37.908 23:41:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:37.908 23:41:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:37.908 23:41:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:07:37.908 23:41:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:07:37.908 23:41:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:37.908 23:41:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:37.908 23:41:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:37.908 23:41:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:07:37.908 23:41:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt3 00:07:37.908 23:41:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:37.908 23:41:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:37.908 23:41:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:37.908 23:41:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:07:37.908 23:41:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:37.909 23:41:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:37.909 23:41:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:07:37.909 23:41:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:37.909 23:41:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:07:37.909 23:41:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:07:37.909 23:41:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@652 -- # local es=0 00:07:37.909 23:41:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:07:37.909 23:41:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:07:37.909 23:41:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:07:37.909 23:41:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:07:37.909 23:41:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:07:37.909 23:41:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:07:37.909 23:41:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:37.909 23:41:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:37.909 [2024-11-26 23:41:25.863542] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:07:37.909 [2024-11-26 23:41:25.865373] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:07:37.909 [2024-11-26 23:41:25.865434] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc3 is claimed 00:07:37.909 [2024-11-26 23:41:25.865480] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:07:37.909 [2024-11-26 23:41:25.865518] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:07:37.909 [2024-11-26 23:41:25.865551] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc3 00:07:37.909 [2024-11-26 23:41:25.865563] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:07:37.909 [2024-11-26 23:41:25.865573] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state configuring 00:07:37.909 request: 00:07:37.909 { 00:07:37.909 "name": "raid_bdev1", 00:07:37.909 "raid_level": "raid0", 00:07:37.909 "base_bdevs": [ 00:07:37.909 "malloc1", 00:07:37.909 "malloc2", 00:07:37.909 "malloc3" 00:07:37.909 ], 00:07:37.909 "strip_size_kb": 64, 00:07:37.909 "superblock": false, 00:07:37.909 "method": "bdev_raid_create", 00:07:37.909 "req_id": 1 00:07:37.909 } 00:07:37.909 Got JSON-RPC error response 00:07:37.909 response: 00:07:37.909 { 00:07:37.909 "code": -17, 00:07:37.909 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:07:37.909 } 00:07:37.909 23:41:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:07:37.909 23:41:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # es=1 00:07:37.909 23:41:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:07:37.909 23:41:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:07:37.909 23:41:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:07:37.909 23:41:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:37.909 23:41:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:37.909 23:41:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:37.909 23:41:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:07:37.909 23:41:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:37.909 23:41:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:07:37.909 23:41:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:07:37.909 23:41:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:07:37.909 23:41:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:37.909 23:41:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:37.909 [2024-11-26 23:41:25.927398] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:07:37.909 [2024-11-26 23:41:25.927449] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:37.909 [2024-11-26 23:41:25.927464] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:07:37.909 [2024-11-26 23:41:25.927474] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:37.909 [2024-11-26 23:41:25.929546] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:37.909 [2024-11-26 23:41:25.929580] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:07:37.909 [2024-11-26 23:41:25.929658] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:07:37.909 [2024-11-26 23:41:25.929713] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:07:37.909 pt1 00:07:37.909 23:41:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:37.909 23:41:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid0 64 3 00:07:37.909 23:41:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:37.909 23:41:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:37.909 23:41:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:37.909 23:41:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:37.909 23:41:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:07:37.909 23:41:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:37.909 23:41:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:37.909 23:41:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:37.909 23:41:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:37.909 23:41:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:37.909 23:41:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:37.909 23:41:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:37.909 23:41:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:37.909 23:41:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:37.909 23:41:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:37.909 "name": "raid_bdev1", 00:07:37.909 "uuid": "cbb0c36d-9b9b-4515-9d9d-545a0f139f8c", 00:07:37.909 "strip_size_kb": 64, 00:07:37.909 "state": "configuring", 00:07:37.909 "raid_level": "raid0", 00:07:37.909 "superblock": true, 00:07:37.909 "num_base_bdevs": 3, 00:07:37.909 "num_base_bdevs_discovered": 1, 00:07:37.909 "num_base_bdevs_operational": 3, 00:07:37.909 "base_bdevs_list": [ 00:07:37.909 { 00:07:37.909 "name": "pt1", 00:07:37.909 "uuid": "00000000-0000-0000-0000-000000000001", 00:07:37.909 "is_configured": true, 00:07:37.909 "data_offset": 2048, 00:07:37.909 "data_size": 63488 00:07:37.909 }, 00:07:37.909 { 00:07:37.909 "name": null, 00:07:37.909 "uuid": "00000000-0000-0000-0000-000000000002", 00:07:37.909 "is_configured": false, 00:07:37.909 "data_offset": 2048, 00:07:37.909 "data_size": 63488 00:07:37.909 }, 00:07:37.909 { 00:07:37.909 "name": null, 00:07:37.909 "uuid": "00000000-0000-0000-0000-000000000003", 00:07:37.909 "is_configured": false, 00:07:37.909 "data_offset": 2048, 00:07:37.909 "data_size": 63488 00:07:37.909 } 00:07:37.909 ] 00:07:37.909 }' 00:07:37.909 23:41:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:37.909 23:41:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:38.478 23:41:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 3 -gt 2 ']' 00:07:38.478 23:41:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@472 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:07:38.478 23:41:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:38.478 23:41:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:38.478 [2024-11-26 23:41:26.322741] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:07:38.478 [2024-11-26 23:41:26.322819] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:38.478 [2024-11-26 23:41:26.322843] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:07:38.478 [2024-11-26 23:41:26.322857] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:38.478 [2024-11-26 23:41:26.323271] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:38.478 [2024-11-26 23:41:26.323300] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:07:38.478 [2024-11-26 23:41:26.323390] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:07:38.478 [2024-11-26 23:41:26.323414] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:07:38.478 pt2 00:07:38.478 23:41:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:38.478 23:41:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@473 -- # rpc_cmd bdev_passthru_delete pt2 00:07:38.478 23:41:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:38.478 23:41:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:38.478 [2024-11-26 23:41:26.330715] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt2 00:07:38.478 23:41:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:38.478 23:41:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@474 -- # verify_raid_bdev_state raid_bdev1 configuring raid0 64 3 00:07:38.478 23:41:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:38.478 23:41:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:38.478 23:41:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:38.478 23:41:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:38.478 23:41:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:07:38.478 23:41:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:38.478 23:41:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:38.478 23:41:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:38.478 23:41:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:38.478 23:41:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:38.478 23:41:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:38.478 23:41:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:38.478 23:41:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:38.478 23:41:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:38.478 23:41:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:38.478 "name": "raid_bdev1", 00:07:38.478 "uuid": "cbb0c36d-9b9b-4515-9d9d-545a0f139f8c", 00:07:38.478 "strip_size_kb": 64, 00:07:38.478 "state": "configuring", 00:07:38.478 "raid_level": "raid0", 00:07:38.478 "superblock": true, 00:07:38.478 "num_base_bdevs": 3, 00:07:38.478 "num_base_bdevs_discovered": 1, 00:07:38.478 "num_base_bdevs_operational": 3, 00:07:38.478 "base_bdevs_list": [ 00:07:38.478 { 00:07:38.478 "name": "pt1", 00:07:38.478 "uuid": "00000000-0000-0000-0000-000000000001", 00:07:38.478 "is_configured": true, 00:07:38.478 "data_offset": 2048, 00:07:38.478 "data_size": 63488 00:07:38.478 }, 00:07:38.478 { 00:07:38.478 "name": null, 00:07:38.478 "uuid": "00000000-0000-0000-0000-000000000002", 00:07:38.478 "is_configured": false, 00:07:38.478 "data_offset": 0, 00:07:38.478 "data_size": 63488 00:07:38.478 }, 00:07:38.478 { 00:07:38.478 "name": null, 00:07:38.478 "uuid": "00000000-0000-0000-0000-000000000003", 00:07:38.478 "is_configured": false, 00:07:38.478 "data_offset": 2048, 00:07:38.478 "data_size": 63488 00:07:38.478 } 00:07:38.478 ] 00:07:38.478 }' 00:07:38.478 23:41:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:38.478 23:41:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:38.738 23:41:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:07:38.738 23:41:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:07:38.738 23:41:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:07:38.738 23:41:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:38.738 23:41:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:38.738 [2024-11-26 23:41:26.746059] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:07:38.738 [2024-11-26 23:41:26.746152] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:38.738 [2024-11-26 23:41:26.746173] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:07:38.738 [2024-11-26 23:41:26.746182] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:38.738 [2024-11-26 23:41:26.746608] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:38.738 [2024-11-26 23:41:26.746633] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:07:38.738 [2024-11-26 23:41:26.746710] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:07:38.738 [2024-11-26 23:41:26.746731] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:07:38.738 pt2 00:07:38.738 23:41:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:38.738 23:41:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:07:38.738 23:41:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:07:38.738 23:41:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:07:38.738 23:41:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:38.738 23:41:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:38.738 [2024-11-26 23:41:26.754031] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:07:38.738 [2024-11-26 23:41:26.754085] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:38.738 [2024-11-26 23:41:26.754104] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:07:38.738 [2024-11-26 23:41:26.754112] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:38.738 [2024-11-26 23:41:26.754477] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:38.738 [2024-11-26 23:41:26.754497] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:07:38.738 [2024-11-26 23:41:26.754558] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:07:38.738 [2024-11-26 23:41:26.754576] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:07:38.738 [2024-11-26 23:41:26.754671] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:07:38.738 [2024-11-26 23:41:26.754679] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:07:38.738 [2024-11-26 23:41:26.754904] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:07:38.738 [2024-11-26 23:41:26.755024] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:07:38.738 [2024-11-26 23:41:26.755038] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:07:38.738 [2024-11-26 23:41:26.755138] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:38.738 pt3 00:07:38.738 23:41:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:38.738 23:41:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:07:38.738 23:41:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:07:38.738 23:41:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 3 00:07:38.738 23:41:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:38.738 23:41:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:38.738 23:41:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:38.738 23:41:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:38.738 23:41:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:07:38.738 23:41:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:38.738 23:41:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:38.738 23:41:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:38.738 23:41:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:38.738 23:41:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:38.738 23:41:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:38.738 23:41:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:38.738 23:41:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:38.738 23:41:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:38.738 23:41:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:38.738 "name": "raid_bdev1", 00:07:38.738 "uuid": "cbb0c36d-9b9b-4515-9d9d-545a0f139f8c", 00:07:38.738 "strip_size_kb": 64, 00:07:38.738 "state": "online", 00:07:38.738 "raid_level": "raid0", 00:07:38.738 "superblock": true, 00:07:38.738 "num_base_bdevs": 3, 00:07:38.738 "num_base_bdevs_discovered": 3, 00:07:38.738 "num_base_bdevs_operational": 3, 00:07:38.738 "base_bdevs_list": [ 00:07:38.739 { 00:07:38.739 "name": "pt1", 00:07:38.739 "uuid": "00000000-0000-0000-0000-000000000001", 00:07:38.739 "is_configured": true, 00:07:38.739 "data_offset": 2048, 00:07:38.739 "data_size": 63488 00:07:38.739 }, 00:07:38.739 { 00:07:38.739 "name": "pt2", 00:07:38.739 "uuid": "00000000-0000-0000-0000-000000000002", 00:07:38.739 "is_configured": true, 00:07:38.739 "data_offset": 2048, 00:07:38.739 "data_size": 63488 00:07:38.739 }, 00:07:38.739 { 00:07:38.739 "name": "pt3", 00:07:38.739 "uuid": "00000000-0000-0000-0000-000000000003", 00:07:38.739 "is_configured": true, 00:07:38.739 "data_offset": 2048, 00:07:38.739 "data_size": 63488 00:07:38.739 } 00:07:38.739 ] 00:07:38.739 }' 00:07:38.739 23:41:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:38.739 23:41:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:39.308 23:41:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:07:39.308 23:41:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:07:39.308 23:41:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:07:39.308 23:41:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:07:39.308 23:41:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:07:39.308 23:41:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:07:39.308 23:41:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:07:39.308 23:41:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:07:39.308 23:41:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:39.308 23:41:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:39.308 [2024-11-26 23:41:27.213545] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:39.308 23:41:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:39.308 23:41:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:07:39.308 "name": "raid_bdev1", 00:07:39.308 "aliases": [ 00:07:39.308 "cbb0c36d-9b9b-4515-9d9d-545a0f139f8c" 00:07:39.308 ], 00:07:39.308 "product_name": "Raid Volume", 00:07:39.308 "block_size": 512, 00:07:39.308 "num_blocks": 190464, 00:07:39.308 "uuid": "cbb0c36d-9b9b-4515-9d9d-545a0f139f8c", 00:07:39.308 "assigned_rate_limits": { 00:07:39.308 "rw_ios_per_sec": 0, 00:07:39.308 "rw_mbytes_per_sec": 0, 00:07:39.308 "r_mbytes_per_sec": 0, 00:07:39.308 "w_mbytes_per_sec": 0 00:07:39.308 }, 00:07:39.308 "claimed": false, 00:07:39.308 "zoned": false, 00:07:39.308 "supported_io_types": { 00:07:39.308 "read": true, 00:07:39.308 "write": true, 00:07:39.308 "unmap": true, 00:07:39.308 "flush": true, 00:07:39.308 "reset": true, 00:07:39.308 "nvme_admin": false, 00:07:39.308 "nvme_io": false, 00:07:39.308 "nvme_io_md": false, 00:07:39.308 "write_zeroes": true, 00:07:39.308 "zcopy": false, 00:07:39.308 "get_zone_info": false, 00:07:39.308 "zone_management": false, 00:07:39.308 "zone_append": false, 00:07:39.308 "compare": false, 00:07:39.308 "compare_and_write": false, 00:07:39.308 "abort": false, 00:07:39.308 "seek_hole": false, 00:07:39.308 "seek_data": false, 00:07:39.308 "copy": false, 00:07:39.308 "nvme_iov_md": false 00:07:39.308 }, 00:07:39.308 "memory_domains": [ 00:07:39.308 { 00:07:39.308 "dma_device_id": "system", 00:07:39.308 "dma_device_type": 1 00:07:39.308 }, 00:07:39.308 { 00:07:39.308 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:39.308 "dma_device_type": 2 00:07:39.308 }, 00:07:39.308 { 00:07:39.308 "dma_device_id": "system", 00:07:39.308 "dma_device_type": 1 00:07:39.308 }, 00:07:39.308 { 00:07:39.308 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:39.308 "dma_device_type": 2 00:07:39.308 }, 00:07:39.308 { 00:07:39.308 "dma_device_id": "system", 00:07:39.308 "dma_device_type": 1 00:07:39.308 }, 00:07:39.308 { 00:07:39.308 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:39.308 "dma_device_type": 2 00:07:39.308 } 00:07:39.308 ], 00:07:39.308 "driver_specific": { 00:07:39.308 "raid": { 00:07:39.308 "uuid": "cbb0c36d-9b9b-4515-9d9d-545a0f139f8c", 00:07:39.308 "strip_size_kb": 64, 00:07:39.308 "state": "online", 00:07:39.308 "raid_level": "raid0", 00:07:39.308 "superblock": true, 00:07:39.308 "num_base_bdevs": 3, 00:07:39.308 "num_base_bdevs_discovered": 3, 00:07:39.308 "num_base_bdevs_operational": 3, 00:07:39.308 "base_bdevs_list": [ 00:07:39.308 { 00:07:39.308 "name": "pt1", 00:07:39.308 "uuid": "00000000-0000-0000-0000-000000000001", 00:07:39.308 "is_configured": true, 00:07:39.308 "data_offset": 2048, 00:07:39.308 "data_size": 63488 00:07:39.308 }, 00:07:39.308 { 00:07:39.308 "name": "pt2", 00:07:39.308 "uuid": "00000000-0000-0000-0000-000000000002", 00:07:39.308 "is_configured": true, 00:07:39.308 "data_offset": 2048, 00:07:39.308 "data_size": 63488 00:07:39.308 }, 00:07:39.308 { 00:07:39.308 "name": "pt3", 00:07:39.308 "uuid": "00000000-0000-0000-0000-000000000003", 00:07:39.308 "is_configured": true, 00:07:39.308 "data_offset": 2048, 00:07:39.308 "data_size": 63488 00:07:39.308 } 00:07:39.308 ] 00:07:39.308 } 00:07:39.308 } 00:07:39.308 }' 00:07:39.308 23:41:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:07:39.308 23:41:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:07:39.308 pt2 00:07:39.308 pt3' 00:07:39.308 23:41:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:39.308 23:41:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:07:39.308 23:41:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:39.308 23:41:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:07:39.308 23:41:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:39.308 23:41:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:39.308 23:41:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:39.308 23:41:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:39.308 23:41:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:39.308 23:41:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:39.308 23:41:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:39.308 23:41:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:07:39.308 23:41:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:39.308 23:41:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:39.308 23:41:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:39.308 23:41:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:39.308 23:41:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:39.308 23:41:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:39.308 23:41:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:39.308 23:41:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:39.308 23:41:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:07:39.309 23:41:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:39.309 23:41:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:39.568 23:41:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:39.568 23:41:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:39.568 23:41:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:39.568 23:41:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:07:39.568 23:41:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:07:39.568 23:41:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:39.568 23:41:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:39.568 [2024-11-26 23:41:27.453079] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:39.568 23:41:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:39.568 23:41:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' cbb0c36d-9b9b-4515-9d9d-545a0f139f8c '!=' cbb0c36d-9b9b-4515-9d9d-545a0f139f8c ']' 00:07:39.568 23:41:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy raid0 00:07:39.568 23:41:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:07:39.568 23:41:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # return 1 00:07:39.568 23:41:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 75957 00:07:39.568 23:41:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # '[' -z 75957 ']' 00:07:39.568 23:41:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@958 -- # kill -0 75957 00:07:39.568 23:41:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # uname 00:07:39.568 23:41:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:07:39.568 23:41:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 75957 00:07:39.569 23:41:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:07:39.569 23:41:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:07:39.569 killing process with pid 75957 00:07:39.569 23:41:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 75957' 00:07:39.569 23:41:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@973 -- # kill 75957 00:07:39.569 [2024-11-26 23:41:27.523656] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:39.569 [2024-11-26 23:41:27.523746] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:39.569 [2024-11-26 23:41:27.523810] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:39.569 [2024-11-26 23:41:27.523820] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:07:39.569 23:41:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@978 -- # wait 75957 00:07:39.569 [2024-11-26 23:41:27.555995] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:39.828 23:41:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:07:39.828 00:07:39.828 real 0m3.762s 00:07:39.828 user 0m5.957s 00:07:39.828 sys 0m0.794s 00:07:39.828 23:41:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:07:39.828 23:41:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:39.828 ************************************ 00:07:39.828 END TEST raid_superblock_test 00:07:39.828 ************************************ 00:07:39.828 23:41:27 bdev_raid -- bdev/bdev_raid.sh@971 -- # run_test raid_read_error_test raid_io_error_test raid0 3 read 00:07:39.828 23:41:27 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:07:39.828 23:41:27 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:39.828 23:41:27 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:39.828 ************************************ 00:07:39.828 START TEST raid_read_error_test 00:07:39.828 ************************************ 00:07:39.828 23:41:27 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test raid0 3 read 00:07:39.828 23:41:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid0 00:07:39.828 23:41:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=3 00:07:39.828 23:41:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=read 00:07:39.828 23:41:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:07:39.828 23:41:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:07:39.828 23:41:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:07:39.828 23:41:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:07:39.828 23:41:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:07:39.828 23:41:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:07:39.828 23:41:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:07:39.828 23:41:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:07:39.828 23:41:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:07:39.828 23:41:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:07:39.828 23:41:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:07:39.828 23:41:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:07:39.828 23:41:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:07:39.828 23:41:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:07:39.828 23:41:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:07:39.828 23:41:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:07:39.828 23:41:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:07:39.828 23:41:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:07:39.828 23:41:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid0 '!=' raid1 ']' 00:07:39.828 23:41:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:07:39.828 23:41:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:07:39.828 23:41:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:07:39.828 23:41:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.R9tuaIElPc 00:07:39.828 23:41:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=76199 00:07:39.828 23:41:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:07:39.828 23:41:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 76199 00:07:39.828 23:41:27 bdev_raid.raid_read_error_test -- common/autotest_common.sh@835 -- # '[' -z 76199 ']' 00:07:39.828 23:41:27 bdev_raid.raid_read_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:39.828 23:41:27 bdev_raid.raid_read_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:39.828 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:39.828 23:41:27 bdev_raid.raid_read_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:39.828 23:41:27 bdev_raid.raid_read_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:39.828 23:41:27 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:39.828 [2024-11-26 23:41:27.931657] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:07:39.828 [2024-11-26 23:41:27.931784] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid76199 ] 00:07:40.086 [2024-11-26 23:41:28.086872] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:40.086 [2024-11-26 23:41:28.112073] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:40.086 [2024-11-26 23:41:28.153930] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:40.086 [2024-11-26 23:41:28.153982] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:40.654 23:41:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:40.654 23:41:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@868 -- # return 0 00:07:40.654 23:41:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:07:40.654 23:41:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:07:40.654 23:41:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:40.654 23:41:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:40.654 BaseBdev1_malloc 00:07:40.654 23:41:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:40.654 23:41:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:07:40.654 23:41:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:40.654 23:41:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:40.654 true 00:07:40.654 23:41:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:40.654 23:41:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:07:40.654 23:41:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:40.654 23:41:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:40.914 [2024-11-26 23:41:28.784335] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:07:40.914 [2024-11-26 23:41:28.784417] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:40.914 [2024-11-26 23:41:28.784437] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006980 00:07:40.914 [2024-11-26 23:41:28.784445] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:40.914 [2024-11-26 23:41:28.786506] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:40.914 [2024-11-26 23:41:28.786545] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:07:40.914 BaseBdev1 00:07:40.914 23:41:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:40.914 23:41:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:07:40.914 23:41:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:07:40.914 23:41:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:40.914 23:41:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:40.914 BaseBdev2_malloc 00:07:40.914 23:41:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:40.914 23:41:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:07:40.914 23:41:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:40.914 23:41:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:40.914 true 00:07:40.914 23:41:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:40.914 23:41:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:07:40.914 23:41:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:40.914 23:41:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:40.914 [2024-11-26 23:41:28.824751] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:07:40.914 [2024-11-26 23:41:28.824792] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:40.915 [2024-11-26 23:41:28.824809] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:07:40.915 [2024-11-26 23:41:28.824826] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:40.915 [2024-11-26 23:41:28.826872] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:40.915 [2024-11-26 23:41:28.826911] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:07:40.915 BaseBdev2 00:07:40.915 23:41:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:40.915 23:41:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:07:40.915 23:41:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:07:40.915 23:41:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:40.915 23:41:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:40.915 BaseBdev3_malloc 00:07:40.915 23:41:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:40.915 23:41:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:07:40.915 23:41:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:40.915 23:41:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:40.915 true 00:07:40.915 23:41:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:40.915 23:41:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:07:40.915 23:41:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:40.915 23:41:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:40.915 [2024-11-26 23:41:28.869074] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:07:40.915 [2024-11-26 23:41:28.869123] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:40.915 [2024-11-26 23:41:28.869141] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008780 00:07:40.915 [2024-11-26 23:41:28.869149] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:40.915 [2024-11-26 23:41:28.871138] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:40.915 [2024-11-26 23:41:28.871175] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:07:40.915 BaseBdev3 00:07:40.915 23:41:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:40.915 23:41:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n raid_bdev1 -s 00:07:40.915 23:41:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:40.915 23:41:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:40.915 [2024-11-26 23:41:28.881104] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:40.915 [2024-11-26 23:41:28.882868] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:07:40.915 [2024-11-26 23:41:28.882941] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:07:40.915 [2024-11-26 23:41:28.883102] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:07:40.915 [2024-11-26 23:41:28.883125] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:07:40.915 [2024-11-26 23:41:28.883377] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002bb0 00:07:40.915 [2024-11-26 23:41:28.883518] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:07:40.915 [2024-11-26 23:41:28.883537] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001c80 00:07:40.915 [2024-11-26 23:41:28.883660] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:40.915 23:41:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:40.915 23:41:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 3 00:07:40.915 23:41:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:40.915 23:41:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:40.915 23:41:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:40.915 23:41:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:40.915 23:41:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:07:40.915 23:41:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:40.915 23:41:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:40.915 23:41:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:40.915 23:41:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:40.915 23:41:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:40.915 23:41:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:40.915 23:41:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:40.915 23:41:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:40.915 23:41:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:40.915 23:41:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:40.915 "name": "raid_bdev1", 00:07:40.915 "uuid": "065f6cc0-589e-4b27-8e53-95fbcd487cf3", 00:07:40.915 "strip_size_kb": 64, 00:07:40.915 "state": "online", 00:07:40.915 "raid_level": "raid0", 00:07:40.915 "superblock": true, 00:07:40.915 "num_base_bdevs": 3, 00:07:40.915 "num_base_bdevs_discovered": 3, 00:07:40.915 "num_base_bdevs_operational": 3, 00:07:40.915 "base_bdevs_list": [ 00:07:40.915 { 00:07:40.915 "name": "BaseBdev1", 00:07:40.915 "uuid": "d0080658-d61e-552b-aa53-b1d87c39fd8d", 00:07:40.915 "is_configured": true, 00:07:40.915 "data_offset": 2048, 00:07:40.915 "data_size": 63488 00:07:40.915 }, 00:07:40.915 { 00:07:40.915 "name": "BaseBdev2", 00:07:40.915 "uuid": "3aee1ba0-62a9-5ee9-9834-8fee9e5f0283", 00:07:40.915 "is_configured": true, 00:07:40.915 "data_offset": 2048, 00:07:40.915 "data_size": 63488 00:07:40.915 }, 00:07:40.915 { 00:07:40.915 "name": "BaseBdev3", 00:07:40.915 "uuid": "1ceaa9e3-00c6-5a9c-9c6d-233dc83f4e9c", 00:07:40.915 "is_configured": true, 00:07:40.915 "data_offset": 2048, 00:07:40.915 "data_size": 63488 00:07:40.915 } 00:07:40.915 ] 00:07:40.915 }' 00:07:40.915 23:41:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:40.915 23:41:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:41.180 23:41:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:07:41.180 23:41:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:07:41.438 [2024-11-26 23:41:29.396621] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002d50 00:07:42.378 23:41:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:07:42.378 23:41:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:42.378 23:41:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:42.378 23:41:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:42.378 23:41:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:07:42.378 23:41:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid0 = \r\a\i\d\1 ]] 00:07:42.378 23:41:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=3 00:07:42.378 23:41:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 3 00:07:42.378 23:41:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:42.378 23:41:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:42.378 23:41:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:42.378 23:41:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:42.378 23:41:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:07:42.378 23:41:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:42.378 23:41:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:42.378 23:41:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:42.378 23:41:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:42.378 23:41:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:42.378 23:41:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:42.378 23:41:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:42.378 23:41:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:42.378 23:41:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:42.378 23:41:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:42.378 "name": "raid_bdev1", 00:07:42.378 "uuid": "065f6cc0-589e-4b27-8e53-95fbcd487cf3", 00:07:42.378 "strip_size_kb": 64, 00:07:42.378 "state": "online", 00:07:42.378 "raid_level": "raid0", 00:07:42.378 "superblock": true, 00:07:42.378 "num_base_bdevs": 3, 00:07:42.378 "num_base_bdevs_discovered": 3, 00:07:42.378 "num_base_bdevs_operational": 3, 00:07:42.378 "base_bdevs_list": [ 00:07:42.378 { 00:07:42.378 "name": "BaseBdev1", 00:07:42.378 "uuid": "d0080658-d61e-552b-aa53-b1d87c39fd8d", 00:07:42.378 "is_configured": true, 00:07:42.378 "data_offset": 2048, 00:07:42.378 "data_size": 63488 00:07:42.378 }, 00:07:42.378 { 00:07:42.378 "name": "BaseBdev2", 00:07:42.378 "uuid": "3aee1ba0-62a9-5ee9-9834-8fee9e5f0283", 00:07:42.378 "is_configured": true, 00:07:42.378 "data_offset": 2048, 00:07:42.378 "data_size": 63488 00:07:42.378 }, 00:07:42.378 { 00:07:42.378 "name": "BaseBdev3", 00:07:42.378 "uuid": "1ceaa9e3-00c6-5a9c-9c6d-233dc83f4e9c", 00:07:42.378 "is_configured": true, 00:07:42.378 "data_offset": 2048, 00:07:42.378 "data_size": 63488 00:07:42.378 } 00:07:42.378 ] 00:07:42.378 }' 00:07:42.378 23:41:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:42.378 23:41:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:42.638 23:41:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:07:42.638 23:41:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:42.638 23:41:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:42.638 [2024-11-26 23:41:30.748480] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:07:42.638 [2024-11-26 23:41:30.748520] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:42.638 [2024-11-26 23:41:30.751061] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:42.638 [2024-11-26 23:41:30.751119] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:42.638 [2024-11-26 23:41:30.751155] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:42.638 [2024-11-26 23:41:30.751166] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name raid_bdev1, state offline 00:07:42.638 { 00:07:42.638 "results": [ 00:07:42.638 { 00:07:42.638 "job": "raid_bdev1", 00:07:42.638 "core_mask": "0x1", 00:07:42.638 "workload": "randrw", 00:07:42.638 "percentage": 50, 00:07:42.638 "status": "finished", 00:07:42.638 "queue_depth": 1, 00:07:42.638 "io_size": 131072, 00:07:42.638 "runtime": 1.352636, 00:07:42.638 "iops": 17014.924931762867, 00:07:42.638 "mibps": 2126.8656164703584, 00:07:42.638 "io_failed": 1, 00:07:42.638 "io_timeout": 0, 00:07:42.638 "avg_latency_us": 81.12494911457075, 00:07:42.638 "min_latency_us": 19.227947598253277, 00:07:42.638 "max_latency_us": 1452.380786026201 00:07:42.638 } 00:07:42.638 ], 00:07:42.638 "core_count": 1 00:07:42.638 } 00:07:42.638 23:41:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:42.638 23:41:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 76199 00:07:42.638 23:41:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # '[' -z 76199 ']' 00:07:42.638 23:41:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@958 -- # kill -0 76199 00:07:42.638 23:41:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # uname 00:07:42.638 23:41:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:07:42.638 23:41:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 76199 00:07:42.897 23:41:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:07:42.897 23:41:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:07:42.897 killing process with pid 76199 00:07:42.897 23:41:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 76199' 00:07:42.897 23:41:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@973 -- # kill 76199 00:07:42.897 [2024-11-26 23:41:30.796000] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:42.897 23:41:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@978 -- # wait 76199 00:07:42.897 [2024-11-26 23:41:30.820177] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:42.897 23:41:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.R9tuaIElPc 00:07:42.897 23:41:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:07:42.897 23:41:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:07:43.157 23:41:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.74 00:07:43.157 23:41:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid0 00:07:43.157 23:41:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:07:43.157 23:41:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:07:43.157 23:41:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.74 != \0\.\0\0 ]] 00:07:43.157 00:07:43.157 real 0m3.200s 00:07:43.157 user 0m4.051s 00:07:43.157 sys 0m0.511s 00:07:43.157 23:41:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:07:43.157 23:41:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:43.157 ************************************ 00:07:43.157 END TEST raid_read_error_test 00:07:43.157 ************************************ 00:07:43.157 23:41:31 bdev_raid -- bdev/bdev_raid.sh@972 -- # run_test raid_write_error_test raid_io_error_test raid0 3 write 00:07:43.157 23:41:31 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:07:43.157 23:41:31 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:43.157 23:41:31 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:43.157 ************************************ 00:07:43.157 START TEST raid_write_error_test 00:07:43.157 ************************************ 00:07:43.157 23:41:31 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test raid0 3 write 00:07:43.157 23:41:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid0 00:07:43.157 23:41:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=3 00:07:43.157 23:41:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=write 00:07:43.157 23:41:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:07:43.157 23:41:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:07:43.157 23:41:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:07:43.157 23:41:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:07:43.157 23:41:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:07:43.157 23:41:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:07:43.157 23:41:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:07:43.157 23:41:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:07:43.157 23:41:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:07:43.157 23:41:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:07:43.157 23:41:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:07:43.157 23:41:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:07:43.157 23:41:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:07:43.157 23:41:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:07:43.157 23:41:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:07:43.157 23:41:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:07:43.157 23:41:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:07:43.157 23:41:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:07:43.157 23:41:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid0 '!=' raid1 ']' 00:07:43.157 23:41:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:07:43.157 23:41:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:07:43.157 23:41:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:07:43.157 23:41:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.5XGoyvfwGD 00:07:43.157 23:41:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=76328 00:07:43.157 23:41:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:07:43.157 23:41:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 76328 00:07:43.157 23:41:31 bdev_raid.raid_write_error_test -- common/autotest_common.sh@835 -- # '[' -z 76328 ']' 00:07:43.157 23:41:31 bdev_raid.raid_write_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:43.157 23:41:31 bdev_raid.raid_write_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:43.157 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:43.157 23:41:31 bdev_raid.raid_write_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:43.157 23:41:31 bdev_raid.raid_write_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:43.157 23:41:31 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:43.157 [2024-11-26 23:41:31.197777] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:07:43.157 [2024-11-26 23:41:31.197900] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid76328 ] 00:07:43.416 [2024-11-26 23:41:31.353623] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:43.416 [2024-11-26 23:41:31.377611] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:43.416 [2024-11-26 23:41:31.418974] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:43.416 [2024-11-26 23:41:31.419016] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:43.998 23:41:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:43.998 23:41:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@868 -- # return 0 00:07:43.998 23:41:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:07:43.998 23:41:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:07:43.998 23:41:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:43.998 23:41:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:43.998 BaseBdev1_malloc 00:07:43.998 23:41:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:43.998 23:41:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:07:43.998 23:41:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:43.998 23:41:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:43.998 true 00:07:43.998 23:41:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:43.998 23:41:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:07:43.998 23:41:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:43.998 23:41:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:43.998 [2024-11-26 23:41:32.061724] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:07:43.998 [2024-11-26 23:41:32.061793] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:43.998 [2024-11-26 23:41:32.061823] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006980 00:07:43.998 [2024-11-26 23:41:32.061834] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:43.998 [2024-11-26 23:41:32.063935] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:43.998 [2024-11-26 23:41:32.063965] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:07:43.998 BaseBdev1 00:07:43.998 23:41:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:43.998 23:41:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:07:43.998 23:41:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:07:43.998 23:41:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:43.998 23:41:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:43.998 BaseBdev2_malloc 00:07:43.998 23:41:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:43.998 23:41:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:07:43.998 23:41:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:43.998 23:41:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:43.998 true 00:07:43.998 23:41:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:43.998 23:41:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:07:43.998 23:41:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:43.998 23:41:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:43.998 [2024-11-26 23:41:32.102144] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:07:43.998 [2024-11-26 23:41:32.102189] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:43.998 [2024-11-26 23:41:32.102205] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:07:43.998 [2024-11-26 23:41:32.102221] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:43.998 [2024-11-26 23:41:32.104245] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:43.998 [2024-11-26 23:41:32.104280] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:07:43.998 BaseBdev2 00:07:43.998 23:41:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:43.998 23:41:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:07:43.998 23:41:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:07:43.998 23:41:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:43.998 23:41:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:43.998 BaseBdev3_malloc 00:07:43.998 23:41:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:43.998 23:41:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:07:43.998 23:41:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:43.998 23:41:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:44.257 true 00:07:44.257 23:41:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:44.257 23:41:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:07:44.257 23:41:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:44.257 23:41:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:44.257 [2024-11-26 23:41:32.142626] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:07:44.257 [2024-11-26 23:41:32.142673] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:44.257 [2024-11-26 23:41:32.142707] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008780 00:07:44.257 [2024-11-26 23:41:32.142716] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:44.257 [2024-11-26 23:41:32.144739] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:44.257 [2024-11-26 23:41:32.144772] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:07:44.257 BaseBdev3 00:07:44.257 23:41:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:44.257 23:41:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n raid_bdev1 -s 00:07:44.257 23:41:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:44.257 23:41:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:44.257 [2024-11-26 23:41:32.154651] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:44.257 [2024-11-26 23:41:32.156551] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:07:44.257 [2024-11-26 23:41:32.156626] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:07:44.257 [2024-11-26 23:41:32.156807] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:07:44.257 [2024-11-26 23:41:32.156822] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:07:44.257 [2024-11-26 23:41:32.157084] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002bb0 00:07:44.257 [2024-11-26 23:41:32.157244] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:07:44.257 [2024-11-26 23:41:32.157262] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001c80 00:07:44.257 [2024-11-26 23:41:32.157414] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:44.257 23:41:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:44.257 23:41:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 3 00:07:44.257 23:41:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:44.257 23:41:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:44.257 23:41:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:44.257 23:41:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:44.257 23:41:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:07:44.257 23:41:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:44.257 23:41:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:44.257 23:41:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:44.257 23:41:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:44.257 23:41:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:44.257 23:41:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:44.258 23:41:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:44.258 23:41:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:44.258 23:41:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:44.258 23:41:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:44.258 "name": "raid_bdev1", 00:07:44.258 "uuid": "f2ab960d-9332-4ca7-bd47-c43b33b05cd8", 00:07:44.258 "strip_size_kb": 64, 00:07:44.258 "state": "online", 00:07:44.258 "raid_level": "raid0", 00:07:44.258 "superblock": true, 00:07:44.258 "num_base_bdevs": 3, 00:07:44.258 "num_base_bdevs_discovered": 3, 00:07:44.258 "num_base_bdevs_operational": 3, 00:07:44.258 "base_bdevs_list": [ 00:07:44.258 { 00:07:44.258 "name": "BaseBdev1", 00:07:44.258 "uuid": "78b97c3e-bb60-58e2-9f91-86ba201bbda5", 00:07:44.258 "is_configured": true, 00:07:44.258 "data_offset": 2048, 00:07:44.258 "data_size": 63488 00:07:44.258 }, 00:07:44.258 { 00:07:44.258 "name": "BaseBdev2", 00:07:44.258 "uuid": "bfdd91a4-0f02-51b8-84ce-0db867875f5e", 00:07:44.258 "is_configured": true, 00:07:44.258 "data_offset": 2048, 00:07:44.258 "data_size": 63488 00:07:44.258 }, 00:07:44.258 { 00:07:44.258 "name": "BaseBdev3", 00:07:44.258 "uuid": "ab0c9836-d109-5992-8121-9f614e2ff796", 00:07:44.258 "is_configured": true, 00:07:44.258 "data_offset": 2048, 00:07:44.258 "data_size": 63488 00:07:44.258 } 00:07:44.258 ] 00:07:44.258 }' 00:07:44.258 23:41:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:44.258 23:41:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:44.518 23:41:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:07:44.518 23:41:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:07:44.778 [2024-11-26 23:41:32.682168] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002d50 00:07:45.716 23:41:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:07:45.716 23:41:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:45.716 23:41:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:45.716 23:41:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:45.716 23:41:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:07:45.716 23:41:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid0 = \r\a\i\d\1 ]] 00:07:45.716 23:41:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=3 00:07:45.716 23:41:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 3 00:07:45.716 23:41:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:45.716 23:41:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:45.716 23:41:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:45.716 23:41:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:45.717 23:41:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:07:45.717 23:41:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:45.717 23:41:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:45.717 23:41:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:45.717 23:41:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:45.717 23:41:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:45.717 23:41:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:45.717 23:41:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:45.717 23:41:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:45.717 23:41:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:45.717 23:41:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:45.717 "name": "raid_bdev1", 00:07:45.717 "uuid": "f2ab960d-9332-4ca7-bd47-c43b33b05cd8", 00:07:45.717 "strip_size_kb": 64, 00:07:45.717 "state": "online", 00:07:45.717 "raid_level": "raid0", 00:07:45.717 "superblock": true, 00:07:45.717 "num_base_bdevs": 3, 00:07:45.717 "num_base_bdevs_discovered": 3, 00:07:45.717 "num_base_bdevs_operational": 3, 00:07:45.717 "base_bdevs_list": [ 00:07:45.717 { 00:07:45.717 "name": "BaseBdev1", 00:07:45.717 "uuid": "78b97c3e-bb60-58e2-9f91-86ba201bbda5", 00:07:45.717 "is_configured": true, 00:07:45.717 "data_offset": 2048, 00:07:45.717 "data_size": 63488 00:07:45.717 }, 00:07:45.717 { 00:07:45.717 "name": "BaseBdev2", 00:07:45.717 "uuid": "bfdd91a4-0f02-51b8-84ce-0db867875f5e", 00:07:45.717 "is_configured": true, 00:07:45.717 "data_offset": 2048, 00:07:45.717 "data_size": 63488 00:07:45.717 }, 00:07:45.717 { 00:07:45.717 "name": "BaseBdev3", 00:07:45.717 "uuid": "ab0c9836-d109-5992-8121-9f614e2ff796", 00:07:45.717 "is_configured": true, 00:07:45.717 "data_offset": 2048, 00:07:45.717 "data_size": 63488 00:07:45.717 } 00:07:45.717 ] 00:07:45.717 }' 00:07:45.717 23:41:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:45.717 23:41:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:45.976 23:41:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:07:45.976 23:41:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:45.976 23:41:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:45.976 [2024-11-26 23:41:34.053686] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:07:45.976 [2024-11-26 23:41:34.053731] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:45.976 [2024-11-26 23:41:34.056320] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:45.976 [2024-11-26 23:41:34.056385] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:45.976 [2024-11-26 23:41:34.056422] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:45.976 [2024-11-26 23:41:34.056434] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name raid_bdev1, state offline 00:07:45.976 { 00:07:45.976 "results": [ 00:07:45.976 { 00:07:45.976 "job": "raid_bdev1", 00:07:45.976 "core_mask": "0x1", 00:07:45.976 "workload": "randrw", 00:07:45.976 "percentage": 50, 00:07:45.976 "status": "finished", 00:07:45.976 "queue_depth": 1, 00:07:45.976 "io_size": 131072, 00:07:45.976 "runtime": 1.372493, 00:07:45.976 "iops": 16801.542885828927, 00:07:45.976 "mibps": 2100.192860728616, 00:07:45.976 "io_failed": 1, 00:07:45.976 "io_timeout": 0, 00:07:45.976 "avg_latency_us": 82.13004242213881, 00:07:45.976 "min_latency_us": 25.152838427947597, 00:07:45.976 "max_latency_us": 1323.598253275109 00:07:45.976 } 00:07:45.976 ], 00:07:45.976 "core_count": 1 00:07:45.976 } 00:07:45.976 23:41:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:45.976 23:41:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 76328 00:07:45.976 23:41:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # '[' -z 76328 ']' 00:07:45.976 23:41:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@958 -- # kill -0 76328 00:07:45.976 23:41:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # uname 00:07:45.976 23:41:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:07:45.976 23:41:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 76328 00:07:45.976 23:41:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:07:45.976 23:41:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:07:45.976 killing process with pid 76328 00:07:45.976 23:41:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 76328' 00:07:45.976 23:41:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@973 -- # kill 76328 00:07:45.976 [2024-11-26 23:41:34.102311] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:45.976 23:41:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@978 -- # wait 76328 00:07:46.236 [2024-11-26 23:41:34.128085] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:46.236 23:41:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.5XGoyvfwGD 00:07:46.236 23:41:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:07:46.236 23:41:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:07:46.236 23:41:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.73 00:07:46.236 23:41:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid0 00:07:46.236 23:41:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:07:46.236 23:41:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:07:46.236 23:41:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.73 != \0\.\0\0 ]] 00:07:46.236 00:07:46.236 real 0m3.228s 00:07:46.236 user 0m4.110s 00:07:46.236 sys 0m0.517s 00:07:46.236 23:41:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:07:46.236 23:41:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:46.236 ************************************ 00:07:46.236 END TEST raid_write_error_test 00:07:46.236 ************************************ 00:07:46.496 23:41:34 bdev_raid -- bdev/bdev_raid.sh@967 -- # for level in raid0 concat raid1 00:07:46.496 23:41:34 bdev_raid -- bdev/bdev_raid.sh@968 -- # run_test raid_state_function_test raid_state_function_test concat 3 false 00:07:46.496 23:41:34 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:07:46.496 23:41:34 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:46.496 23:41:34 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:46.496 ************************************ 00:07:46.496 START TEST raid_state_function_test 00:07:46.496 ************************************ 00:07:46.496 23:41:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1129 -- # raid_state_function_test concat 3 false 00:07:46.496 23:41:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=concat 00:07:46.496 23:41:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=3 00:07:46.496 23:41:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:07:46.496 23:41:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:07:46.496 23:41:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:07:46.497 23:41:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:46.497 23:41:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:07:46.497 23:41:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:07:46.497 23:41:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:46.497 23:41:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:07:46.497 23:41:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:07:46.497 23:41:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:46.497 23:41:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:07:46.497 23:41:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:07:46.497 23:41:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:46.497 23:41:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:07:46.497 23:41:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:07:46.497 23:41:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:07:46.497 23:41:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:07:46.497 23:41:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:07:46.497 23:41:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:07:46.497 23:41:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' concat '!=' raid1 ']' 00:07:46.497 23:41:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:07:46.497 23:41:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:07:46.497 23:41:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:07:46.497 23:41:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:07:46.497 23:41:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=76455 00:07:46.497 23:41:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:07:46.497 Process raid pid: 76455 00:07:46.497 23:41:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 76455' 00:07:46.497 23:41:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 76455 00:07:46.497 23:41:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@835 -- # '[' -z 76455 ']' 00:07:46.497 23:41:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:46.497 23:41:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:46.497 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:46.497 23:41:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:46.497 23:41:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:46.497 23:41:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:46.497 [2024-11-26 23:41:34.496890] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:07:46.497 [2024-11-26 23:41:34.497007] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:07:46.757 [2024-11-26 23:41:34.651632] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:46.757 [2024-11-26 23:41:34.678039] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:46.757 [2024-11-26 23:41:34.719718] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:46.757 [2024-11-26 23:41:34.719761] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:47.326 23:41:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:47.326 23:41:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@868 -- # return 0 00:07:47.326 23:41:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:07:47.326 23:41:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:47.326 23:41:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:47.326 [2024-11-26 23:41:35.342148] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:07:47.326 [2024-11-26 23:41:35.342202] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:07:47.326 [2024-11-26 23:41:35.342212] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:47.326 [2024-11-26 23:41:35.342221] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:47.326 [2024-11-26 23:41:35.342227] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:07:47.326 [2024-11-26 23:41:35.342328] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:07:47.326 23:41:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:47.326 23:41:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:07:47.326 23:41:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:47.326 23:41:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:47.326 23:41:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:47.326 23:41:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:47.326 23:41:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:07:47.326 23:41:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:47.326 23:41:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:47.326 23:41:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:47.326 23:41:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:47.326 23:41:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:47.326 23:41:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:47.326 23:41:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:47.326 23:41:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:47.326 23:41:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:47.326 23:41:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:47.326 "name": "Existed_Raid", 00:07:47.326 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:47.326 "strip_size_kb": 64, 00:07:47.326 "state": "configuring", 00:07:47.326 "raid_level": "concat", 00:07:47.326 "superblock": false, 00:07:47.326 "num_base_bdevs": 3, 00:07:47.326 "num_base_bdevs_discovered": 0, 00:07:47.326 "num_base_bdevs_operational": 3, 00:07:47.326 "base_bdevs_list": [ 00:07:47.326 { 00:07:47.326 "name": "BaseBdev1", 00:07:47.326 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:47.326 "is_configured": false, 00:07:47.326 "data_offset": 0, 00:07:47.326 "data_size": 0 00:07:47.326 }, 00:07:47.326 { 00:07:47.326 "name": "BaseBdev2", 00:07:47.326 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:47.326 "is_configured": false, 00:07:47.326 "data_offset": 0, 00:07:47.326 "data_size": 0 00:07:47.326 }, 00:07:47.326 { 00:07:47.326 "name": "BaseBdev3", 00:07:47.326 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:47.326 "is_configured": false, 00:07:47.326 "data_offset": 0, 00:07:47.326 "data_size": 0 00:07:47.326 } 00:07:47.326 ] 00:07:47.326 }' 00:07:47.326 23:41:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:47.326 23:41:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:47.897 23:41:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:07:47.897 23:41:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:47.897 23:41:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:47.897 [2024-11-26 23:41:35.781289] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:07:47.897 [2024-11-26 23:41:35.781440] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:07:47.897 23:41:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:47.897 23:41:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:07:47.897 23:41:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:47.897 23:41:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:47.897 [2024-11-26 23:41:35.793290] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:07:47.897 [2024-11-26 23:41:35.793384] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:07:47.897 [2024-11-26 23:41:35.793413] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:47.897 [2024-11-26 23:41:35.793436] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:47.897 [2024-11-26 23:41:35.793453] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:07:47.897 [2024-11-26 23:41:35.793473] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:07:47.897 23:41:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:47.897 23:41:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:07:47.897 23:41:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:47.897 23:41:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:47.897 BaseBdev1 00:07:47.897 [2024-11-26 23:41:35.813853] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:47.897 23:41:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:47.897 23:41:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:07:47.897 23:41:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:07:47.897 23:41:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:07:47.897 23:41:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:07:47.897 23:41:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:07:47.897 23:41:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:07:47.897 23:41:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:07:47.897 23:41:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:47.897 23:41:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:47.897 23:41:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:47.897 23:41:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:07:47.897 23:41:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:47.897 23:41:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:47.897 [ 00:07:47.897 { 00:07:47.897 "name": "BaseBdev1", 00:07:47.897 "aliases": [ 00:07:47.897 "340746cb-b5de-4b2e-9700-163faa7f0c1e" 00:07:47.897 ], 00:07:47.897 "product_name": "Malloc disk", 00:07:47.897 "block_size": 512, 00:07:47.897 "num_blocks": 65536, 00:07:47.897 "uuid": "340746cb-b5de-4b2e-9700-163faa7f0c1e", 00:07:47.897 "assigned_rate_limits": { 00:07:47.897 "rw_ios_per_sec": 0, 00:07:47.897 "rw_mbytes_per_sec": 0, 00:07:47.897 "r_mbytes_per_sec": 0, 00:07:47.897 "w_mbytes_per_sec": 0 00:07:47.897 }, 00:07:47.897 "claimed": true, 00:07:47.897 "claim_type": "exclusive_write", 00:07:47.897 "zoned": false, 00:07:47.897 "supported_io_types": { 00:07:47.897 "read": true, 00:07:47.897 "write": true, 00:07:47.897 "unmap": true, 00:07:47.897 "flush": true, 00:07:47.897 "reset": true, 00:07:47.897 "nvme_admin": false, 00:07:47.897 "nvme_io": false, 00:07:47.897 "nvme_io_md": false, 00:07:47.897 "write_zeroes": true, 00:07:47.897 "zcopy": true, 00:07:47.897 "get_zone_info": false, 00:07:47.897 "zone_management": false, 00:07:47.897 "zone_append": false, 00:07:47.897 "compare": false, 00:07:47.897 "compare_and_write": false, 00:07:47.897 "abort": true, 00:07:47.897 "seek_hole": false, 00:07:47.897 "seek_data": false, 00:07:47.897 "copy": true, 00:07:47.897 "nvme_iov_md": false 00:07:47.897 }, 00:07:47.897 "memory_domains": [ 00:07:47.897 { 00:07:47.897 "dma_device_id": "system", 00:07:47.897 "dma_device_type": 1 00:07:47.897 }, 00:07:47.897 { 00:07:47.897 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:47.897 "dma_device_type": 2 00:07:47.897 } 00:07:47.897 ], 00:07:47.897 "driver_specific": {} 00:07:47.897 } 00:07:47.897 ] 00:07:47.897 23:41:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:47.897 23:41:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:07:47.897 23:41:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:07:47.897 23:41:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:47.897 23:41:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:47.897 23:41:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:47.897 23:41:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:47.897 23:41:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:07:47.897 23:41:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:47.897 23:41:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:47.897 23:41:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:47.897 23:41:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:47.897 23:41:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:47.898 23:41:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:47.898 23:41:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:47.898 23:41:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:47.898 23:41:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:47.898 23:41:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:47.898 "name": "Existed_Raid", 00:07:47.898 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:47.898 "strip_size_kb": 64, 00:07:47.898 "state": "configuring", 00:07:47.898 "raid_level": "concat", 00:07:47.898 "superblock": false, 00:07:47.898 "num_base_bdevs": 3, 00:07:47.898 "num_base_bdevs_discovered": 1, 00:07:47.898 "num_base_bdevs_operational": 3, 00:07:47.898 "base_bdevs_list": [ 00:07:47.898 { 00:07:47.898 "name": "BaseBdev1", 00:07:47.898 "uuid": "340746cb-b5de-4b2e-9700-163faa7f0c1e", 00:07:47.898 "is_configured": true, 00:07:47.898 "data_offset": 0, 00:07:47.898 "data_size": 65536 00:07:47.898 }, 00:07:47.898 { 00:07:47.898 "name": "BaseBdev2", 00:07:47.898 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:47.898 "is_configured": false, 00:07:47.898 "data_offset": 0, 00:07:47.898 "data_size": 0 00:07:47.898 }, 00:07:47.898 { 00:07:47.898 "name": "BaseBdev3", 00:07:47.898 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:47.898 "is_configured": false, 00:07:47.898 "data_offset": 0, 00:07:47.898 "data_size": 0 00:07:47.898 } 00:07:47.898 ] 00:07:47.898 }' 00:07:47.898 23:41:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:47.898 23:41:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:48.158 23:41:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:07:48.158 23:41:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:48.158 23:41:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:48.158 [2024-11-26 23:41:36.265125] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:07:48.158 [2024-11-26 23:41:36.265260] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:07:48.158 23:41:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:48.158 23:41:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:07:48.158 23:41:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:48.158 23:41:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:48.158 [2024-11-26 23:41:36.277137] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:48.158 [2024-11-26 23:41:36.278980] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:48.158 [2024-11-26 23:41:36.279025] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:48.158 [2024-11-26 23:41:36.279035] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:07:48.158 [2024-11-26 23:41:36.279047] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:07:48.158 23:41:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:48.158 23:41:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:07:48.158 23:41:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:07:48.158 23:41:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:07:48.158 23:41:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:48.158 23:41:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:48.158 23:41:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:48.158 23:41:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:48.158 23:41:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:07:48.158 23:41:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:48.158 23:41:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:48.158 23:41:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:48.158 23:41:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:48.418 23:41:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:48.418 23:41:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:48.418 23:41:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:48.418 23:41:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:48.418 23:41:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:48.418 23:41:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:48.418 "name": "Existed_Raid", 00:07:48.418 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:48.418 "strip_size_kb": 64, 00:07:48.418 "state": "configuring", 00:07:48.418 "raid_level": "concat", 00:07:48.418 "superblock": false, 00:07:48.418 "num_base_bdevs": 3, 00:07:48.418 "num_base_bdevs_discovered": 1, 00:07:48.418 "num_base_bdevs_operational": 3, 00:07:48.418 "base_bdevs_list": [ 00:07:48.418 { 00:07:48.418 "name": "BaseBdev1", 00:07:48.418 "uuid": "340746cb-b5de-4b2e-9700-163faa7f0c1e", 00:07:48.418 "is_configured": true, 00:07:48.418 "data_offset": 0, 00:07:48.418 "data_size": 65536 00:07:48.418 }, 00:07:48.418 { 00:07:48.418 "name": "BaseBdev2", 00:07:48.418 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:48.418 "is_configured": false, 00:07:48.418 "data_offset": 0, 00:07:48.418 "data_size": 0 00:07:48.418 }, 00:07:48.418 { 00:07:48.418 "name": "BaseBdev3", 00:07:48.418 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:48.418 "is_configured": false, 00:07:48.418 "data_offset": 0, 00:07:48.418 "data_size": 0 00:07:48.418 } 00:07:48.418 ] 00:07:48.418 }' 00:07:48.418 23:41:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:48.418 23:41:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:48.678 23:41:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:07:48.678 23:41:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:48.678 23:41:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:48.678 [2024-11-26 23:41:36.775071] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:07:48.678 BaseBdev2 00:07:48.678 23:41:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:48.678 23:41:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:07:48.678 23:41:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:07:48.678 23:41:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:07:48.678 23:41:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:07:48.678 23:41:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:07:48.678 23:41:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:07:48.678 23:41:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:07:48.678 23:41:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:48.678 23:41:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:48.679 23:41:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:48.679 23:41:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:07:48.679 23:41:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:48.679 23:41:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:48.679 [ 00:07:48.679 { 00:07:48.679 "name": "BaseBdev2", 00:07:48.679 "aliases": [ 00:07:48.679 "ddeba4fb-de63-4fb4-b122-26a7c87c296a" 00:07:48.679 ], 00:07:48.679 "product_name": "Malloc disk", 00:07:48.679 "block_size": 512, 00:07:48.679 "num_blocks": 65536, 00:07:48.679 "uuid": "ddeba4fb-de63-4fb4-b122-26a7c87c296a", 00:07:48.679 "assigned_rate_limits": { 00:07:48.679 "rw_ios_per_sec": 0, 00:07:48.679 "rw_mbytes_per_sec": 0, 00:07:48.679 "r_mbytes_per_sec": 0, 00:07:48.679 "w_mbytes_per_sec": 0 00:07:48.679 }, 00:07:48.679 "claimed": true, 00:07:48.679 "claim_type": "exclusive_write", 00:07:48.679 "zoned": false, 00:07:48.679 "supported_io_types": { 00:07:48.679 "read": true, 00:07:48.679 "write": true, 00:07:48.679 "unmap": true, 00:07:48.679 "flush": true, 00:07:48.679 "reset": true, 00:07:48.679 "nvme_admin": false, 00:07:48.679 "nvme_io": false, 00:07:48.679 "nvme_io_md": false, 00:07:48.679 "write_zeroes": true, 00:07:48.679 "zcopy": true, 00:07:48.679 "get_zone_info": false, 00:07:48.679 "zone_management": false, 00:07:48.679 "zone_append": false, 00:07:48.679 "compare": false, 00:07:48.679 "compare_and_write": false, 00:07:48.937 "abort": true, 00:07:48.937 "seek_hole": false, 00:07:48.937 "seek_data": false, 00:07:48.937 "copy": true, 00:07:48.937 "nvme_iov_md": false 00:07:48.937 }, 00:07:48.937 "memory_domains": [ 00:07:48.937 { 00:07:48.937 "dma_device_id": "system", 00:07:48.937 "dma_device_type": 1 00:07:48.937 }, 00:07:48.937 { 00:07:48.937 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:48.937 "dma_device_type": 2 00:07:48.937 } 00:07:48.937 ], 00:07:48.937 "driver_specific": {} 00:07:48.937 } 00:07:48.937 ] 00:07:48.937 23:41:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:48.937 23:41:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:07:48.937 23:41:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:07:48.937 23:41:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:07:48.937 23:41:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:07:48.937 23:41:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:48.937 23:41:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:48.937 23:41:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:48.937 23:41:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:48.937 23:41:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:07:48.938 23:41:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:48.938 23:41:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:48.938 23:41:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:48.938 23:41:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:48.938 23:41:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:48.938 23:41:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:48.938 23:41:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:48.938 23:41:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:48.938 23:41:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:48.938 23:41:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:48.938 "name": "Existed_Raid", 00:07:48.938 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:48.938 "strip_size_kb": 64, 00:07:48.938 "state": "configuring", 00:07:48.938 "raid_level": "concat", 00:07:48.938 "superblock": false, 00:07:48.938 "num_base_bdevs": 3, 00:07:48.938 "num_base_bdevs_discovered": 2, 00:07:48.938 "num_base_bdevs_operational": 3, 00:07:48.938 "base_bdevs_list": [ 00:07:48.938 { 00:07:48.938 "name": "BaseBdev1", 00:07:48.938 "uuid": "340746cb-b5de-4b2e-9700-163faa7f0c1e", 00:07:48.938 "is_configured": true, 00:07:48.938 "data_offset": 0, 00:07:48.938 "data_size": 65536 00:07:48.938 }, 00:07:48.938 { 00:07:48.938 "name": "BaseBdev2", 00:07:48.938 "uuid": "ddeba4fb-de63-4fb4-b122-26a7c87c296a", 00:07:48.938 "is_configured": true, 00:07:48.938 "data_offset": 0, 00:07:48.938 "data_size": 65536 00:07:48.938 }, 00:07:48.938 { 00:07:48.938 "name": "BaseBdev3", 00:07:48.938 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:48.938 "is_configured": false, 00:07:48.938 "data_offset": 0, 00:07:48.938 "data_size": 0 00:07:48.938 } 00:07:48.938 ] 00:07:48.938 }' 00:07:48.938 23:41:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:48.938 23:41:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:49.197 23:41:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:07:49.197 23:41:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:49.197 23:41:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:49.197 [2024-11-26 23:41:37.268307] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:07:49.197 [2024-11-26 23:41:37.268440] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:07:49.197 [2024-11-26 23:41:37.268475] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 196608, blocklen 512 00:07:49.197 [2024-11-26 23:41:37.269328] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:07:49.197 [2024-11-26 23:41:37.269791] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:07:49.197 [2024-11-26 23:41:37.269827] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:07:49.197 [2024-11-26 23:41:37.270485] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:49.197 BaseBdev3 00:07:49.197 23:41:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:49.197 23:41:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:07:49.197 23:41:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:07:49.197 23:41:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:07:49.197 23:41:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:07:49.197 23:41:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:07:49.197 23:41:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:07:49.197 23:41:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:07:49.197 23:41:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:49.197 23:41:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:49.197 23:41:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:49.197 23:41:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:07:49.197 23:41:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:49.197 23:41:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:49.197 [ 00:07:49.197 { 00:07:49.197 "name": "BaseBdev3", 00:07:49.197 "aliases": [ 00:07:49.197 "eeda4268-d136-41d1-b00a-b778e5746d5a" 00:07:49.197 ], 00:07:49.197 "product_name": "Malloc disk", 00:07:49.197 "block_size": 512, 00:07:49.197 "num_blocks": 65536, 00:07:49.197 "uuid": "eeda4268-d136-41d1-b00a-b778e5746d5a", 00:07:49.197 "assigned_rate_limits": { 00:07:49.197 "rw_ios_per_sec": 0, 00:07:49.197 "rw_mbytes_per_sec": 0, 00:07:49.197 "r_mbytes_per_sec": 0, 00:07:49.197 "w_mbytes_per_sec": 0 00:07:49.197 }, 00:07:49.197 "claimed": true, 00:07:49.197 "claim_type": "exclusive_write", 00:07:49.197 "zoned": false, 00:07:49.197 "supported_io_types": { 00:07:49.197 "read": true, 00:07:49.197 "write": true, 00:07:49.197 "unmap": true, 00:07:49.197 "flush": true, 00:07:49.197 "reset": true, 00:07:49.197 "nvme_admin": false, 00:07:49.197 "nvme_io": false, 00:07:49.197 "nvme_io_md": false, 00:07:49.197 "write_zeroes": true, 00:07:49.197 "zcopy": true, 00:07:49.197 "get_zone_info": false, 00:07:49.197 "zone_management": false, 00:07:49.197 "zone_append": false, 00:07:49.197 "compare": false, 00:07:49.197 "compare_and_write": false, 00:07:49.197 "abort": true, 00:07:49.197 "seek_hole": false, 00:07:49.197 "seek_data": false, 00:07:49.197 "copy": true, 00:07:49.197 "nvme_iov_md": false 00:07:49.197 }, 00:07:49.197 "memory_domains": [ 00:07:49.197 { 00:07:49.197 "dma_device_id": "system", 00:07:49.198 "dma_device_type": 1 00:07:49.198 }, 00:07:49.198 { 00:07:49.198 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:49.198 "dma_device_type": 2 00:07:49.198 } 00:07:49.198 ], 00:07:49.198 "driver_specific": {} 00:07:49.198 } 00:07:49.198 ] 00:07:49.198 23:41:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:49.198 23:41:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:07:49.198 23:41:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:07:49.198 23:41:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:07:49.198 23:41:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online concat 64 3 00:07:49.198 23:41:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:49.198 23:41:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:49.198 23:41:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:49.198 23:41:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:49.198 23:41:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:07:49.198 23:41:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:49.198 23:41:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:49.198 23:41:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:49.198 23:41:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:49.198 23:41:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:49.198 23:41:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:49.198 23:41:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:49.198 23:41:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:49.460 23:41:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:49.460 23:41:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:49.460 "name": "Existed_Raid", 00:07:49.460 "uuid": "532d5fee-8bb2-456e-aeab-aefcfbe0278b", 00:07:49.460 "strip_size_kb": 64, 00:07:49.460 "state": "online", 00:07:49.460 "raid_level": "concat", 00:07:49.460 "superblock": false, 00:07:49.460 "num_base_bdevs": 3, 00:07:49.460 "num_base_bdevs_discovered": 3, 00:07:49.460 "num_base_bdevs_operational": 3, 00:07:49.460 "base_bdevs_list": [ 00:07:49.460 { 00:07:49.460 "name": "BaseBdev1", 00:07:49.460 "uuid": "340746cb-b5de-4b2e-9700-163faa7f0c1e", 00:07:49.460 "is_configured": true, 00:07:49.460 "data_offset": 0, 00:07:49.460 "data_size": 65536 00:07:49.460 }, 00:07:49.460 { 00:07:49.460 "name": "BaseBdev2", 00:07:49.460 "uuid": "ddeba4fb-de63-4fb4-b122-26a7c87c296a", 00:07:49.460 "is_configured": true, 00:07:49.460 "data_offset": 0, 00:07:49.460 "data_size": 65536 00:07:49.460 }, 00:07:49.460 { 00:07:49.460 "name": "BaseBdev3", 00:07:49.460 "uuid": "eeda4268-d136-41d1-b00a-b778e5746d5a", 00:07:49.460 "is_configured": true, 00:07:49.460 "data_offset": 0, 00:07:49.460 "data_size": 65536 00:07:49.460 } 00:07:49.460 ] 00:07:49.460 }' 00:07:49.460 23:41:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:49.460 23:41:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:49.721 23:41:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:07:49.721 23:41:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:07:49.721 23:41:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:07:49.721 23:41:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:07:49.721 23:41:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:07:49.721 23:41:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:07:49.721 23:41:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:07:49.721 23:41:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:49.721 23:41:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:49.721 23:41:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:07:49.721 [2024-11-26 23:41:37.779786] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:49.721 23:41:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:49.721 23:41:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:07:49.721 "name": "Existed_Raid", 00:07:49.721 "aliases": [ 00:07:49.721 "532d5fee-8bb2-456e-aeab-aefcfbe0278b" 00:07:49.721 ], 00:07:49.721 "product_name": "Raid Volume", 00:07:49.721 "block_size": 512, 00:07:49.721 "num_blocks": 196608, 00:07:49.721 "uuid": "532d5fee-8bb2-456e-aeab-aefcfbe0278b", 00:07:49.721 "assigned_rate_limits": { 00:07:49.721 "rw_ios_per_sec": 0, 00:07:49.721 "rw_mbytes_per_sec": 0, 00:07:49.721 "r_mbytes_per_sec": 0, 00:07:49.721 "w_mbytes_per_sec": 0 00:07:49.721 }, 00:07:49.721 "claimed": false, 00:07:49.721 "zoned": false, 00:07:49.721 "supported_io_types": { 00:07:49.721 "read": true, 00:07:49.721 "write": true, 00:07:49.721 "unmap": true, 00:07:49.721 "flush": true, 00:07:49.721 "reset": true, 00:07:49.721 "nvme_admin": false, 00:07:49.721 "nvme_io": false, 00:07:49.721 "nvme_io_md": false, 00:07:49.721 "write_zeroes": true, 00:07:49.721 "zcopy": false, 00:07:49.721 "get_zone_info": false, 00:07:49.721 "zone_management": false, 00:07:49.721 "zone_append": false, 00:07:49.721 "compare": false, 00:07:49.721 "compare_and_write": false, 00:07:49.721 "abort": false, 00:07:49.721 "seek_hole": false, 00:07:49.721 "seek_data": false, 00:07:49.721 "copy": false, 00:07:49.721 "nvme_iov_md": false 00:07:49.721 }, 00:07:49.721 "memory_domains": [ 00:07:49.721 { 00:07:49.721 "dma_device_id": "system", 00:07:49.721 "dma_device_type": 1 00:07:49.721 }, 00:07:49.721 { 00:07:49.721 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:49.721 "dma_device_type": 2 00:07:49.721 }, 00:07:49.721 { 00:07:49.721 "dma_device_id": "system", 00:07:49.721 "dma_device_type": 1 00:07:49.721 }, 00:07:49.721 { 00:07:49.721 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:49.721 "dma_device_type": 2 00:07:49.721 }, 00:07:49.721 { 00:07:49.721 "dma_device_id": "system", 00:07:49.721 "dma_device_type": 1 00:07:49.721 }, 00:07:49.721 { 00:07:49.721 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:49.721 "dma_device_type": 2 00:07:49.721 } 00:07:49.721 ], 00:07:49.721 "driver_specific": { 00:07:49.721 "raid": { 00:07:49.721 "uuid": "532d5fee-8bb2-456e-aeab-aefcfbe0278b", 00:07:49.721 "strip_size_kb": 64, 00:07:49.721 "state": "online", 00:07:49.721 "raid_level": "concat", 00:07:49.721 "superblock": false, 00:07:49.721 "num_base_bdevs": 3, 00:07:49.721 "num_base_bdevs_discovered": 3, 00:07:49.721 "num_base_bdevs_operational": 3, 00:07:49.721 "base_bdevs_list": [ 00:07:49.721 { 00:07:49.721 "name": "BaseBdev1", 00:07:49.721 "uuid": "340746cb-b5de-4b2e-9700-163faa7f0c1e", 00:07:49.721 "is_configured": true, 00:07:49.721 "data_offset": 0, 00:07:49.721 "data_size": 65536 00:07:49.721 }, 00:07:49.721 { 00:07:49.721 "name": "BaseBdev2", 00:07:49.721 "uuid": "ddeba4fb-de63-4fb4-b122-26a7c87c296a", 00:07:49.721 "is_configured": true, 00:07:49.721 "data_offset": 0, 00:07:49.721 "data_size": 65536 00:07:49.721 }, 00:07:49.721 { 00:07:49.721 "name": "BaseBdev3", 00:07:49.721 "uuid": "eeda4268-d136-41d1-b00a-b778e5746d5a", 00:07:49.721 "is_configured": true, 00:07:49.721 "data_offset": 0, 00:07:49.721 "data_size": 65536 00:07:49.721 } 00:07:49.721 ] 00:07:49.721 } 00:07:49.721 } 00:07:49.721 }' 00:07:49.721 23:41:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:07:49.982 23:41:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:07:49.982 BaseBdev2 00:07:49.982 BaseBdev3' 00:07:49.982 23:41:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:49.982 23:41:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:07:49.982 23:41:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:49.982 23:41:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:07:49.982 23:41:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:49.982 23:41:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:49.982 23:41:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:49.982 23:41:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:49.982 23:41:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:49.982 23:41:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:49.982 23:41:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:49.982 23:41:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:49.982 23:41:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:07:49.982 23:41:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:49.982 23:41:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:49.982 23:41:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:49.982 23:41:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:49.982 23:41:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:49.982 23:41:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:49.982 23:41:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:07:49.982 23:41:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:49.982 23:41:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:49.982 23:41:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:49.982 23:41:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:49.982 23:41:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:49.982 23:41:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:49.982 23:41:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:07:49.982 23:41:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:49.982 23:41:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:49.982 [2024-11-26 23:41:38.082932] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:07:49.982 [2024-11-26 23:41:38.082969] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:49.982 [2024-11-26 23:41:38.083030] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:49.982 23:41:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:49.982 23:41:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:07:49.982 23:41:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy concat 00:07:49.982 23:41:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:07:49.982 23:41:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # return 1 00:07:49.982 23:41:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:07:49.982 23:41:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline concat 64 2 00:07:49.982 23:41:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:49.982 23:41:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:07:49.982 23:41:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:49.982 23:41:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:49.982 23:41:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:49.982 23:41:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:49.982 23:41:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:49.982 23:41:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:49.982 23:41:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:49.982 23:41:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:49.982 23:41:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:49.982 23:41:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:49.982 23:41:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:50.242 23:41:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:50.242 23:41:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:50.242 "name": "Existed_Raid", 00:07:50.242 "uuid": "532d5fee-8bb2-456e-aeab-aefcfbe0278b", 00:07:50.242 "strip_size_kb": 64, 00:07:50.242 "state": "offline", 00:07:50.242 "raid_level": "concat", 00:07:50.242 "superblock": false, 00:07:50.242 "num_base_bdevs": 3, 00:07:50.242 "num_base_bdevs_discovered": 2, 00:07:50.242 "num_base_bdevs_operational": 2, 00:07:50.242 "base_bdevs_list": [ 00:07:50.242 { 00:07:50.242 "name": null, 00:07:50.242 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:50.242 "is_configured": false, 00:07:50.242 "data_offset": 0, 00:07:50.242 "data_size": 65536 00:07:50.242 }, 00:07:50.242 { 00:07:50.242 "name": "BaseBdev2", 00:07:50.242 "uuid": "ddeba4fb-de63-4fb4-b122-26a7c87c296a", 00:07:50.242 "is_configured": true, 00:07:50.242 "data_offset": 0, 00:07:50.242 "data_size": 65536 00:07:50.242 }, 00:07:50.242 { 00:07:50.242 "name": "BaseBdev3", 00:07:50.242 "uuid": "eeda4268-d136-41d1-b00a-b778e5746d5a", 00:07:50.242 "is_configured": true, 00:07:50.242 "data_offset": 0, 00:07:50.242 "data_size": 65536 00:07:50.242 } 00:07:50.242 ] 00:07:50.242 }' 00:07:50.242 23:41:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:50.242 23:41:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:50.502 23:41:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:07:50.502 23:41:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:07:50.502 23:41:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:50.502 23:41:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:50.502 23:41:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:07:50.502 23:41:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:50.502 23:41:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:50.502 23:41:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:07:50.502 23:41:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:07:50.502 23:41:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:07:50.502 23:41:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:50.502 23:41:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:50.502 [2024-11-26 23:41:38.589401] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:07:50.502 23:41:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:50.502 23:41:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:07:50.502 23:41:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:07:50.502 23:41:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:07:50.502 23:41:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:50.502 23:41:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:50.502 23:41:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:50.502 23:41:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:50.763 23:41:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:07:50.763 23:41:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:07:50.763 23:41:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:07:50.763 23:41:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:50.763 23:41:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:50.763 [2024-11-26 23:41:38.644572] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:07:50.763 [2024-11-26 23:41:38.644677] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:07:50.763 23:41:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:50.763 23:41:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:07:50.763 23:41:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:07:50.763 23:41:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:50.763 23:41:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:07:50.763 23:41:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:50.763 23:41:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:50.763 23:41:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:50.763 23:41:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:07:50.763 23:41:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:07:50.763 23:41:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 3 -gt 2 ']' 00:07:50.763 23:41:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:07:50.763 23:41:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:07:50.763 23:41:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:07:50.763 23:41:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:50.763 23:41:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:50.763 BaseBdev2 00:07:50.763 23:41:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:50.763 23:41:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:07:50.763 23:41:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:07:50.763 23:41:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:07:50.763 23:41:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:07:50.763 23:41:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:07:50.763 23:41:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:07:50.763 23:41:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:07:50.763 23:41:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:50.763 23:41:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:50.763 23:41:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:50.763 23:41:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:07:50.764 23:41:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:50.764 23:41:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:50.764 [ 00:07:50.764 { 00:07:50.764 "name": "BaseBdev2", 00:07:50.764 "aliases": [ 00:07:50.764 "018898b5-4779-4a25-9adf-ebba1b825c36" 00:07:50.764 ], 00:07:50.764 "product_name": "Malloc disk", 00:07:50.764 "block_size": 512, 00:07:50.764 "num_blocks": 65536, 00:07:50.764 "uuid": "018898b5-4779-4a25-9adf-ebba1b825c36", 00:07:50.764 "assigned_rate_limits": { 00:07:50.764 "rw_ios_per_sec": 0, 00:07:50.764 "rw_mbytes_per_sec": 0, 00:07:50.764 "r_mbytes_per_sec": 0, 00:07:50.764 "w_mbytes_per_sec": 0 00:07:50.764 }, 00:07:50.764 "claimed": false, 00:07:50.764 "zoned": false, 00:07:50.764 "supported_io_types": { 00:07:50.764 "read": true, 00:07:50.764 "write": true, 00:07:50.764 "unmap": true, 00:07:50.764 "flush": true, 00:07:50.764 "reset": true, 00:07:50.764 "nvme_admin": false, 00:07:50.764 "nvme_io": false, 00:07:50.764 "nvme_io_md": false, 00:07:50.764 "write_zeroes": true, 00:07:50.764 "zcopy": true, 00:07:50.764 "get_zone_info": false, 00:07:50.764 "zone_management": false, 00:07:50.764 "zone_append": false, 00:07:50.764 "compare": false, 00:07:50.764 "compare_and_write": false, 00:07:50.764 "abort": true, 00:07:50.764 "seek_hole": false, 00:07:50.764 "seek_data": false, 00:07:50.764 "copy": true, 00:07:50.764 "nvme_iov_md": false 00:07:50.764 }, 00:07:50.764 "memory_domains": [ 00:07:50.764 { 00:07:50.764 "dma_device_id": "system", 00:07:50.764 "dma_device_type": 1 00:07:50.764 }, 00:07:50.764 { 00:07:50.764 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:50.764 "dma_device_type": 2 00:07:50.764 } 00:07:50.764 ], 00:07:50.764 "driver_specific": {} 00:07:50.764 } 00:07:50.764 ] 00:07:50.764 23:41:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:50.764 23:41:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:07:50.764 23:41:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:07:50.764 23:41:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:07:50.764 23:41:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:07:50.764 23:41:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:50.764 23:41:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:50.764 BaseBdev3 00:07:50.764 23:41:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:50.764 23:41:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:07:50.764 23:41:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:07:50.764 23:41:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:07:50.764 23:41:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:07:50.764 23:41:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:07:50.764 23:41:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:07:50.764 23:41:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:07:50.764 23:41:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:50.764 23:41:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:50.764 23:41:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:50.764 23:41:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:07:50.764 23:41:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:50.764 23:41:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:50.764 [ 00:07:50.764 { 00:07:50.764 "name": "BaseBdev3", 00:07:50.764 "aliases": [ 00:07:50.764 "b8c0b66f-8730-44ce-8a0b-310849b5753d" 00:07:50.764 ], 00:07:50.764 "product_name": "Malloc disk", 00:07:50.764 "block_size": 512, 00:07:50.764 "num_blocks": 65536, 00:07:50.764 "uuid": "b8c0b66f-8730-44ce-8a0b-310849b5753d", 00:07:50.764 "assigned_rate_limits": { 00:07:50.764 "rw_ios_per_sec": 0, 00:07:50.764 "rw_mbytes_per_sec": 0, 00:07:50.764 "r_mbytes_per_sec": 0, 00:07:50.764 "w_mbytes_per_sec": 0 00:07:50.764 }, 00:07:50.764 "claimed": false, 00:07:50.764 "zoned": false, 00:07:50.764 "supported_io_types": { 00:07:50.764 "read": true, 00:07:50.764 "write": true, 00:07:50.764 "unmap": true, 00:07:50.764 "flush": true, 00:07:50.764 "reset": true, 00:07:50.764 "nvme_admin": false, 00:07:50.764 "nvme_io": false, 00:07:50.764 "nvme_io_md": false, 00:07:50.764 "write_zeroes": true, 00:07:50.764 "zcopy": true, 00:07:50.764 "get_zone_info": false, 00:07:50.764 "zone_management": false, 00:07:50.764 "zone_append": false, 00:07:50.764 "compare": false, 00:07:50.764 "compare_and_write": false, 00:07:50.764 "abort": true, 00:07:50.764 "seek_hole": false, 00:07:50.764 "seek_data": false, 00:07:50.764 "copy": true, 00:07:50.764 "nvme_iov_md": false 00:07:50.764 }, 00:07:50.764 "memory_domains": [ 00:07:50.764 { 00:07:50.764 "dma_device_id": "system", 00:07:50.764 "dma_device_type": 1 00:07:50.764 }, 00:07:50.764 { 00:07:50.764 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:50.764 "dma_device_type": 2 00:07:50.764 } 00:07:50.764 ], 00:07:50.764 "driver_specific": {} 00:07:50.764 } 00:07:50.764 ] 00:07:50.764 23:41:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:50.764 23:41:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:07:50.764 23:41:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:07:50.764 23:41:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:07:50.764 23:41:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:07:50.764 23:41:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:50.764 23:41:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:50.764 [2024-11-26 23:41:38.816744] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:07:50.764 [2024-11-26 23:41:38.816889] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:07:50.764 [2024-11-26 23:41:38.816920] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:07:50.764 [2024-11-26 23:41:38.818861] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:07:50.764 23:41:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:50.764 23:41:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:07:50.764 23:41:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:50.764 23:41:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:50.765 23:41:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:50.765 23:41:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:50.765 23:41:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:07:50.765 23:41:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:50.765 23:41:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:50.765 23:41:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:50.765 23:41:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:50.765 23:41:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:50.765 23:41:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:50.765 23:41:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:50.765 23:41:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:50.765 23:41:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:50.765 23:41:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:50.765 "name": "Existed_Raid", 00:07:50.765 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:50.765 "strip_size_kb": 64, 00:07:50.765 "state": "configuring", 00:07:50.765 "raid_level": "concat", 00:07:50.765 "superblock": false, 00:07:50.765 "num_base_bdevs": 3, 00:07:50.765 "num_base_bdevs_discovered": 2, 00:07:50.765 "num_base_bdevs_operational": 3, 00:07:50.765 "base_bdevs_list": [ 00:07:50.765 { 00:07:50.765 "name": "BaseBdev1", 00:07:50.765 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:50.765 "is_configured": false, 00:07:50.765 "data_offset": 0, 00:07:50.765 "data_size": 0 00:07:50.765 }, 00:07:50.765 { 00:07:50.765 "name": "BaseBdev2", 00:07:50.765 "uuid": "018898b5-4779-4a25-9adf-ebba1b825c36", 00:07:50.765 "is_configured": true, 00:07:50.765 "data_offset": 0, 00:07:50.765 "data_size": 65536 00:07:50.765 }, 00:07:50.765 { 00:07:50.765 "name": "BaseBdev3", 00:07:50.765 "uuid": "b8c0b66f-8730-44ce-8a0b-310849b5753d", 00:07:50.765 "is_configured": true, 00:07:50.765 "data_offset": 0, 00:07:50.765 "data_size": 65536 00:07:50.765 } 00:07:50.765 ] 00:07:50.765 }' 00:07:50.765 23:41:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:50.765 23:41:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:51.334 23:41:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:07:51.334 23:41:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:51.334 23:41:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:51.334 [2024-11-26 23:41:39.255971] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:07:51.334 23:41:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:51.334 23:41:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:07:51.334 23:41:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:51.334 23:41:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:51.334 23:41:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:51.334 23:41:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:51.334 23:41:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:07:51.334 23:41:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:51.334 23:41:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:51.334 23:41:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:51.334 23:41:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:51.334 23:41:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:51.334 23:41:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:51.334 23:41:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:51.334 23:41:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:51.334 23:41:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:51.334 23:41:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:51.334 "name": "Existed_Raid", 00:07:51.334 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:51.334 "strip_size_kb": 64, 00:07:51.334 "state": "configuring", 00:07:51.334 "raid_level": "concat", 00:07:51.334 "superblock": false, 00:07:51.334 "num_base_bdevs": 3, 00:07:51.334 "num_base_bdevs_discovered": 1, 00:07:51.334 "num_base_bdevs_operational": 3, 00:07:51.334 "base_bdevs_list": [ 00:07:51.334 { 00:07:51.334 "name": "BaseBdev1", 00:07:51.334 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:51.334 "is_configured": false, 00:07:51.334 "data_offset": 0, 00:07:51.334 "data_size": 0 00:07:51.334 }, 00:07:51.334 { 00:07:51.334 "name": null, 00:07:51.334 "uuid": "018898b5-4779-4a25-9adf-ebba1b825c36", 00:07:51.334 "is_configured": false, 00:07:51.334 "data_offset": 0, 00:07:51.334 "data_size": 65536 00:07:51.334 }, 00:07:51.334 { 00:07:51.334 "name": "BaseBdev3", 00:07:51.334 "uuid": "b8c0b66f-8730-44ce-8a0b-310849b5753d", 00:07:51.334 "is_configured": true, 00:07:51.334 "data_offset": 0, 00:07:51.334 "data_size": 65536 00:07:51.334 } 00:07:51.334 ] 00:07:51.334 }' 00:07:51.334 23:41:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:51.335 23:41:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:51.594 23:41:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:07:51.594 23:41:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:51.594 23:41:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:51.594 23:41:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:51.594 23:41:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:51.594 23:41:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:07:51.594 23:41:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:07:51.594 23:41:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:51.594 23:41:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:51.594 [2024-11-26 23:41:39.710081] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:51.594 BaseBdev1 00:07:51.594 23:41:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:51.594 23:41:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:07:51.594 23:41:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:07:51.594 23:41:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:07:51.594 23:41:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:07:51.594 23:41:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:07:51.594 23:41:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:07:51.594 23:41:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:07:51.594 23:41:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:51.594 23:41:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:51.854 23:41:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:51.854 23:41:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:07:51.854 23:41:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:51.854 23:41:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:51.854 [ 00:07:51.854 { 00:07:51.854 "name": "BaseBdev1", 00:07:51.854 "aliases": [ 00:07:51.854 "6f329dca-a6f1-4e40-b2d7-f81979a92d57" 00:07:51.854 ], 00:07:51.854 "product_name": "Malloc disk", 00:07:51.854 "block_size": 512, 00:07:51.854 "num_blocks": 65536, 00:07:51.854 "uuid": "6f329dca-a6f1-4e40-b2d7-f81979a92d57", 00:07:51.854 "assigned_rate_limits": { 00:07:51.854 "rw_ios_per_sec": 0, 00:07:51.854 "rw_mbytes_per_sec": 0, 00:07:51.854 "r_mbytes_per_sec": 0, 00:07:51.854 "w_mbytes_per_sec": 0 00:07:51.854 }, 00:07:51.854 "claimed": true, 00:07:51.854 "claim_type": "exclusive_write", 00:07:51.854 "zoned": false, 00:07:51.854 "supported_io_types": { 00:07:51.854 "read": true, 00:07:51.854 "write": true, 00:07:51.854 "unmap": true, 00:07:51.854 "flush": true, 00:07:51.854 "reset": true, 00:07:51.854 "nvme_admin": false, 00:07:51.854 "nvme_io": false, 00:07:51.854 "nvme_io_md": false, 00:07:51.854 "write_zeroes": true, 00:07:51.854 "zcopy": true, 00:07:51.854 "get_zone_info": false, 00:07:51.854 "zone_management": false, 00:07:51.854 "zone_append": false, 00:07:51.854 "compare": false, 00:07:51.854 "compare_and_write": false, 00:07:51.854 "abort": true, 00:07:51.854 "seek_hole": false, 00:07:51.854 "seek_data": false, 00:07:51.854 "copy": true, 00:07:51.854 "nvme_iov_md": false 00:07:51.854 }, 00:07:51.854 "memory_domains": [ 00:07:51.854 { 00:07:51.854 "dma_device_id": "system", 00:07:51.854 "dma_device_type": 1 00:07:51.854 }, 00:07:51.854 { 00:07:51.854 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:51.854 "dma_device_type": 2 00:07:51.854 } 00:07:51.854 ], 00:07:51.854 "driver_specific": {} 00:07:51.854 } 00:07:51.854 ] 00:07:51.854 23:41:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:51.854 23:41:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:07:51.854 23:41:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:07:51.854 23:41:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:51.854 23:41:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:51.854 23:41:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:51.854 23:41:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:51.854 23:41:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:07:51.854 23:41:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:51.854 23:41:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:51.854 23:41:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:51.854 23:41:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:51.854 23:41:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:51.854 23:41:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:51.854 23:41:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:51.854 23:41:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:51.854 23:41:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:51.854 23:41:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:51.854 "name": "Existed_Raid", 00:07:51.854 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:51.854 "strip_size_kb": 64, 00:07:51.854 "state": "configuring", 00:07:51.854 "raid_level": "concat", 00:07:51.854 "superblock": false, 00:07:51.854 "num_base_bdevs": 3, 00:07:51.854 "num_base_bdevs_discovered": 2, 00:07:51.854 "num_base_bdevs_operational": 3, 00:07:51.854 "base_bdevs_list": [ 00:07:51.854 { 00:07:51.854 "name": "BaseBdev1", 00:07:51.854 "uuid": "6f329dca-a6f1-4e40-b2d7-f81979a92d57", 00:07:51.854 "is_configured": true, 00:07:51.854 "data_offset": 0, 00:07:51.854 "data_size": 65536 00:07:51.854 }, 00:07:51.854 { 00:07:51.854 "name": null, 00:07:51.854 "uuid": "018898b5-4779-4a25-9adf-ebba1b825c36", 00:07:51.854 "is_configured": false, 00:07:51.854 "data_offset": 0, 00:07:51.854 "data_size": 65536 00:07:51.854 }, 00:07:51.854 { 00:07:51.854 "name": "BaseBdev3", 00:07:51.854 "uuid": "b8c0b66f-8730-44ce-8a0b-310849b5753d", 00:07:51.854 "is_configured": true, 00:07:51.854 "data_offset": 0, 00:07:51.854 "data_size": 65536 00:07:51.854 } 00:07:51.854 ] 00:07:51.854 }' 00:07:51.854 23:41:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:51.854 23:41:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:52.114 23:41:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:07:52.114 23:41:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:52.114 23:41:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:52.114 23:41:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:52.114 23:41:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:52.114 23:41:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:07:52.114 23:41:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:07:52.114 23:41:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:52.114 23:41:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:52.114 [2024-11-26 23:41:40.185376] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:07:52.114 23:41:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:52.114 23:41:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:07:52.114 23:41:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:52.114 23:41:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:52.114 23:41:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:52.114 23:41:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:52.114 23:41:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:07:52.114 23:41:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:52.114 23:41:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:52.114 23:41:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:52.114 23:41:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:52.114 23:41:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:52.114 23:41:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:52.114 23:41:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:52.114 23:41:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:52.114 23:41:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:52.114 23:41:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:52.114 "name": "Existed_Raid", 00:07:52.114 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:52.114 "strip_size_kb": 64, 00:07:52.114 "state": "configuring", 00:07:52.114 "raid_level": "concat", 00:07:52.114 "superblock": false, 00:07:52.114 "num_base_bdevs": 3, 00:07:52.114 "num_base_bdevs_discovered": 1, 00:07:52.114 "num_base_bdevs_operational": 3, 00:07:52.114 "base_bdevs_list": [ 00:07:52.114 { 00:07:52.114 "name": "BaseBdev1", 00:07:52.114 "uuid": "6f329dca-a6f1-4e40-b2d7-f81979a92d57", 00:07:52.114 "is_configured": true, 00:07:52.114 "data_offset": 0, 00:07:52.114 "data_size": 65536 00:07:52.114 }, 00:07:52.114 { 00:07:52.114 "name": null, 00:07:52.114 "uuid": "018898b5-4779-4a25-9adf-ebba1b825c36", 00:07:52.114 "is_configured": false, 00:07:52.114 "data_offset": 0, 00:07:52.114 "data_size": 65536 00:07:52.114 }, 00:07:52.114 { 00:07:52.114 "name": null, 00:07:52.114 "uuid": "b8c0b66f-8730-44ce-8a0b-310849b5753d", 00:07:52.114 "is_configured": false, 00:07:52.114 "data_offset": 0, 00:07:52.114 "data_size": 65536 00:07:52.114 } 00:07:52.114 ] 00:07:52.114 }' 00:07:52.114 23:41:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:52.114 23:41:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:52.704 23:41:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:52.704 23:41:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:07:52.704 23:41:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:52.704 23:41:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:52.704 23:41:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:52.704 23:41:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:07:52.704 23:41:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:07:52.704 23:41:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:52.704 23:41:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:52.704 [2024-11-26 23:41:40.664563] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:07:52.704 23:41:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:52.704 23:41:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:07:52.704 23:41:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:52.704 23:41:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:52.704 23:41:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:52.704 23:41:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:52.704 23:41:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:07:52.704 23:41:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:52.704 23:41:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:52.704 23:41:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:52.704 23:41:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:52.704 23:41:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:52.704 23:41:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:52.705 23:41:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:52.705 23:41:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:52.705 23:41:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:52.705 23:41:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:52.705 "name": "Existed_Raid", 00:07:52.705 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:52.705 "strip_size_kb": 64, 00:07:52.705 "state": "configuring", 00:07:52.705 "raid_level": "concat", 00:07:52.705 "superblock": false, 00:07:52.705 "num_base_bdevs": 3, 00:07:52.705 "num_base_bdevs_discovered": 2, 00:07:52.705 "num_base_bdevs_operational": 3, 00:07:52.705 "base_bdevs_list": [ 00:07:52.705 { 00:07:52.705 "name": "BaseBdev1", 00:07:52.705 "uuid": "6f329dca-a6f1-4e40-b2d7-f81979a92d57", 00:07:52.705 "is_configured": true, 00:07:52.705 "data_offset": 0, 00:07:52.705 "data_size": 65536 00:07:52.705 }, 00:07:52.705 { 00:07:52.705 "name": null, 00:07:52.705 "uuid": "018898b5-4779-4a25-9adf-ebba1b825c36", 00:07:52.705 "is_configured": false, 00:07:52.705 "data_offset": 0, 00:07:52.705 "data_size": 65536 00:07:52.705 }, 00:07:52.705 { 00:07:52.705 "name": "BaseBdev3", 00:07:52.705 "uuid": "b8c0b66f-8730-44ce-8a0b-310849b5753d", 00:07:52.705 "is_configured": true, 00:07:52.705 "data_offset": 0, 00:07:52.705 "data_size": 65536 00:07:52.705 } 00:07:52.705 ] 00:07:52.705 }' 00:07:52.705 23:41:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:52.705 23:41:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:53.275 23:41:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:07:53.275 23:41:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:53.275 23:41:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:53.275 23:41:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:53.275 23:41:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:53.275 23:41:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:07:53.275 23:41:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:07:53.275 23:41:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:53.275 23:41:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:53.275 [2024-11-26 23:41:41.175697] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:07:53.275 23:41:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:53.275 23:41:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:07:53.275 23:41:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:53.275 23:41:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:53.275 23:41:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:53.275 23:41:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:53.275 23:41:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:07:53.275 23:41:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:53.275 23:41:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:53.275 23:41:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:53.275 23:41:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:53.275 23:41:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:53.275 23:41:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:53.275 23:41:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:53.275 23:41:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:53.275 23:41:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:53.275 23:41:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:53.275 "name": "Existed_Raid", 00:07:53.275 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:53.275 "strip_size_kb": 64, 00:07:53.275 "state": "configuring", 00:07:53.275 "raid_level": "concat", 00:07:53.275 "superblock": false, 00:07:53.275 "num_base_bdevs": 3, 00:07:53.275 "num_base_bdevs_discovered": 1, 00:07:53.275 "num_base_bdevs_operational": 3, 00:07:53.275 "base_bdevs_list": [ 00:07:53.275 { 00:07:53.275 "name": null, 00:07:53.275 "uuid": "6f329dca-a6f1-4e40-b2d7-f81979a92d57", 00:07:53.275 "is_configured": false, 00:07:53.275 "data_offset": 0, 00:07:53.275 "data_size": 65536 00:07:53.275 }, 00:07:53.275 { 00:07:53.275 "name": null, 00:07:53.275 "uuid": "018898b5-4779-4a25-9adf-ebba1b825c36", 00:07:53.275 "is_configured": false, 00:07:53.275 "data_offset": 0, 00:07:53.275 "data_size": 65536 00:07:53.275 }, 00:07:53.275 { 00:07:53.275 "name": "BaseBdev3", 00:07:53.275 "uuid": "b8c0b66f-8730-44ce-8a0b-310849b5753d", 00:07:53.275 "is_configured": true, 00:07:53.275 "data_offset": 0, 00:07:53.275 "data_size": 65536 00:07:53.275 } 00:07:53.275 ] 00:07:53.275 }' 00:07:53.275 23:41:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:53.275 23:41:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:53.536 23:41:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:07:53.536 23:41:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:53.536 23:41:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:53.536 23:41:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:53.796 23:41:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:53.796 23:41:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:07:53.796 23:41:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:07:53.796 23:41:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:53.796 23:41:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:53.796 [2024-11-26 23:41:41.685245] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:07:53.796 23:41:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:53.796 23:41:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:07:53.796 23:41:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:53.796 23:41:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:53.796 23:41:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:53.796 23:41:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:53.796 23:41:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:07:53.796 23:41:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:53.796 23:41:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:53.796 23:41:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:53.796 23:41:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:53.796 23:41:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:53.796 23:41:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:53.796 23:41:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:53.796 23:41:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:53.796 23:41:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:53.796 23:41:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:53.796 "name": "Existed_Raid", 00:07:53.796 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:53.796 "strip_size_kb": 64, 00:07:53.796 "state": "configuring", 00:07:53.796 "raid_level": "concat", 00:07:53.796 "superblock": false, 00:07:53.796 "num_base_bdevs": 3, 00:07:53.796 "num_base_bdevs_discovered": 2, 00:07:53.796 "num_base_bdevs_operational": 3, 00:07:53.796 "base_bdevs_list": [ 00:07:53.796 { 00:07:53.796 "name": null, 00:07:53.796 "uuid": "6f329dca-a6f1-4e40-b2d7-f81979a92d57", 00:07:53.796 "is_configured": false, 00:07:53.796 "data_offset": 0, 00:07:53.796 "data_size": 65536 00:07:53.796 }, 00:07:53.796 { 00:07:53.796 "name": "BaseBdev2", 00:07:53.796 "uuid": "018898b5-4779-4a25-9adf-ebba1b825c36", 00:07:53.796 "is_configured": true, 00:07:53.796 "data_offset": 0, 00:07:53.796 "data_size": 65536 00:07:53.796 }, 00:07:53.796 { 00:07:53.796 "name": "BaseBdev3", 00:07:53.796 "uuid": "b8c0b66f-8730-44ce-8a0b-310849b5753d", 00:07:53.796 "is_configured": true, 00:07:53.796 "data_offset": 0, 00:07:53.796 "data_size": 65536 00:07:53.796 } 00:07:53.796 ] 00:07:53.796 }' 00:07:53.796 23:41:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:53.796 23:41:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:54.063 23:41:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:07:54.063 23:41:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:54.063 23:41:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:54.063 23:41:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:54.063 23:41:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:54.063 23:41:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:07:54.064 23:41:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:54.064 23:41:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:07:54.064 23:41:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:54.064 23:41:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:54.064 23:41:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:54.064 23:41:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 6f329dca-a6f1-4e40-b2d7-f81979a92d57 00:07:54.064 23:41:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:54.064 23:41:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:54.064 [2024-11-26 23:41:42.175226] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:07:54.064 [2024-11-26 23:41:42.175268] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:07:54.064 [2024-11-26 23:41:42.175277] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 196608, blocklen 512 00:07:54.064 [2024-11-26 23:41:42.175543] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002870 00:07:54.064 [2024-11-26 23:41:42.175660] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:07:54.064 [2024-11-26 23:41:42.175669] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001c80 00:07:54.064 [2024-11-26 23:41:42.175843] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:54.064 NewBaseBdev 00:07:54.064 23:41:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:54.064 23:41:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:07:54.064 23:41:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:07:54.064 23:41:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:07:54.064 23:41:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:07:54.064 23:41:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:07:54.064 23:41:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:07:54.064 23:41:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:07:54.064 23:41:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:54.064 23:41:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:54.335 23:41:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:54.335 23:41:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:07:54.335 23:41:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:54.335 23:41:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:54.335 [ 00:07:54.335 { 00:07:54.335 "name": "NewBaseBdev", 00:07:54.335 "aliases": [ 00:07:54.335 "6f329dca-a6f1-4e40-b2d7-f81979a92d57" 00:07:54.335 ], 00:07:54.335 "product_name": "Malloc disk", 00:07:54.335 "block_size": 512, 00:07:54.335 "num_blocks": 65536, 00:07:54.335 "uuid": "6f329dca-a6f1-4e40-b2d7-f81979a92d57", 00:07:54.335 "assigned_rate_limits": { 00:07:54.335 "rw_ios_per_sec": 0, 00:07:54.335 "rw_mbytes_per_sec": 0, 00:07:54.335 "r_mbytes_per_sec": 0, 00:07:54.335 "w_mbytes_per_sec": 0 00:07:54.335 }, 00:07:54.335 "claimed": true, 00:07:54.335 "claim_type": "exclusive_write", 00:07:54.335 "zoned": false, 00:07:54.335 "supported_io_types": { 00:07:54.335 "read": true, 00:07:54.335 "write": true, 00:07:54.335 "unmap": true, 00:07:54.335 "flush": true, 00:07:54.335 "reset": true, 00:07:54.335 "nvme_admin": false, 00:07:54.335 "nvme_io": false, 00:07:54.335 "nvme_io_md": false, 00:07:54.335 "write_zeroes": true, 00:07:54.335 "zcopy": true, 00:07:54.335 "get_zone_info": false, 00:07:54.335 "zone_management": false, 00:07:54.335 "zone_append": false, 00:07:54.335 "compare": false, 00:07:54.335 "compare_and_write": false, 00:07:54.335 "abort": true, 00:07:54.335 "seek_hole": false, 00:07:54.335 "seek_data": false, 00:07:54.335 "copy": true, 00:07:54.335 "nvme_iov_md": false 00:07:54.335 }, 00:07:54.335 "memory_domains": [ 00:07:54.335 { 00:07:54.335 "dma_device_id": "system", 00:07:54.335 "dma_device_type": 1 00:07:54.335 }, 00:07:54.335 { 00:07:54.335 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:54.335 "dma_device_type": 2 00:07:54.335 } 00:07:54.335 ], 00:07:54.335 "driver_specific": {} 00:07:54.335 } 00:07:54.335 ] 00:07:54.335 23:41:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:54.335 23:41:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:07:54.335 23:41:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online concat 64 3 00:07:54.335 23:41:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:54.335 23:41:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:54.335 23:41:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:54.335 23:41:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:54.335 23:41:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:07:54.335 23:41:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:54.335 23:41:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:54.335 23:41:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:54.335 23:41:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:54.335 23:41:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:54.335 23:41:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:54.335 23:41:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:54.335 23:41:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:54.335 23:41:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:54.335 23:41:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:54.335 "name": "Existed_Raid", 00:07:54.335 "uuid": "50d3e29d-cbce-497f-b080-440cb1fdd669", 00:07:54.335 "strip_size_kb": 64, 00:07:54.335 "state": "online", 00:07:54.335 "raid_level": "concat", 00:07:54.335 "superblock": false, 00:07:54.335 "num_base_bdevs": 3, 00:07:54.335 "num_base_bdevs_discovered": 3, 00:07:54.335 "num_base_bdevs_operational": 3, 00:07:54.335 "base_bdevs_list": [ 00:07:54.335 { 00:07:54.335 "name": "NewBaseBdev", 00:07:54.335 "uuid": "6f329dca-a6f1-4e40-b2d7-f81979a92d57", 00:07:54.335 "is_configured": true, 00:07:54.335 "data_offset": 0, 00:07:54.335 "data_size": 65536 00:07:54.335 }, 00:07:54.335 { 00:07:54.335 "name": "BaseBdev2", 00:07:54.335 "uuid": "018898b5-4779-4a25-9adf-ebba1b825c36", 00:07:54.335 "is_configured": true, 00:07:54.335 "data_offset": 0, 00:07:54.335 "data_size": 65536 00:07:54.335 }, 00:07:54.335 { 00:07:54.335 "name": "BaseBdev3", 00:07:54.335 "uuid": "b8c0b66f-8730-44ce-8a0b-310849b5753d", 00:07:54.335 "is_configured": true, 00:07:54.335 "data_offset": 0, 00:07:54.335 "data_size": 65536 00:07:54.335 } 00:07:54.335 ] 00:07:54.335 }' 00:07:54.335 23:41:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:54.335 23:41:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:54.595 23:41:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:07:54.595 23:41:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:07:54.595 23:41:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:07:54.595 23:41:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:07:54.595 23:41:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:07:54.595 23:41:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:07:54.595 23:41:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:07:54.595 23:41:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:07:54.595 23:41:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:54.595 23:41:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:54.595 [2024-11-26 23:41:42.654770] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:54.595 23:41:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:54.596 23:41:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:07:54.596 "name": "Existed_Raid", 00:07:54.596 "aliases": [ 00:07:54.596 "50d3e29d-cbce-497f-b080-440cb1fdd669" 00:07:54.596 ], 00:07:54.596 "product_name": "Raid Volume", 00:07:54.596 "block_size": 512, 00:07:54.596 "num_blocks": 196608, 00:07:54.596 "uuid": "50d3e29d-cbce-497f-b080-440cb1fdd669", 00:07:54.596 "assigned_rate_limits": { 00:07:54.596 "rw_ios_per_sec": 0, 00:07:54.596 "rw_mbytes_per_sec": 0, 00:07:54.596 "r_mbytes_per_sec": 0, 00:07:54.596 "w_mbytes_per_sec": 0 00:07:54.596 }, 00:07:54.596 "claimed": false, 00:07:54.596 "zoned": false, 00:07:54.596 "supported_io_types": { 00:07:54.596 "read": true, 00:07:54.596 "write": true, 00:07:54.596 "unmap": true, 00:07:54.596 "flush": true, 00:07:54.596 "reset": true, 00:07:54.596 "nvme_admin": false, 00:07:54.596 "nvme_io": false, 00:07:54.596 "nvme_io_md": false, 00:07:54.596 "write_zeroes": true, 00:07:54.596 "zcopy": false, 00:07:54.596 "get_zone_info": false, 00:07:54.596 "zone_management": false, 00:07:54.596 "zone_append": false, 00:07:54.596 "compare": false, 00:07:54.596 "compare_and_write": false, 00:07:54.596 "abort": false, 00:07:54.596 "seek_hole": false, 00:07:54.596 "seek_data": false, 00:07:54.596 "copy": false, 00:07:54.596 "nvme_iov_md": false 00:07:54.596 }, 00:07:54.596 "memory_domains": [ 00:07:54.596 { 00:07:54.596 "dma_device_id": "system", 00:07:54.596 "dma_device_type": 1 00:07:54.596 }, 00:07:54.596 { 00:07:54.596 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:54.596 "dma_device_type": 2 00:07:54.596 }, 00:07:54.596 { 00:07:54.596 "dma_device_id": "system", 00:07:54.596 "dma_device_type": 1 00:07:54.596 }, 00:07:54.596 { 00:07:54.596 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:54.596 "dma_device_type": 2 00:07:54.596 }, 00:07:54.596 { 00:07:54.596 "dma_device_id": "system", 00:07:54.596 "dma_device_type": 1 00:07:54.596 }, 00:07:54.596 { 00:07:54.596 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:54.596 "dma_device_type": 2 00:07:54.596 } 00:07:54.596 ], 00:07:54.596 "driver_specific": { 00:07:54.596 "raid": { 00:07:54.596 "uuid": "50d3e29d-cbce-497f-b080-440cb1fdd669", 00:07:54.596 "strip_size_kb": 64, 00:07:54.596 "state": "online", 00:07:54.596 "raid_level": "concat", 00:07:54.596 "superblock": false, 00:07:54.596 "num_base_bdevs": 3, 00:07:54.596 "num_base_bdevs_discovered": 3, 00:07:54.596 "num_base_bdevs_operational": 3, 00:07:54.596 "base_bdevs_list": [ 00:07:54.596 { 00:07:54.596 "name": "NewBaseBdev", 00:07:54.596 "uuid": "6f329dca-a6f1-4e40-b2d7-f81979a92d57", 00:07:54.596 "is_configured": true, 00:07:54.596 "data_offset": 0, 00:07:54.596 "data_size": 65536 00:07:54.596 }, 00:07:54.596 { 00:07:54.596 "name": "BaseBdev2", 00:07:54.596 "uuid": "018898b5-4779-4a25-9adf-ebba1b825c36", 00:07:54.596 "is_configured": true, 00:07:54.596 "data_offset": 0, 00:07:54.596 "data_size": 65536 00:07:54.596 }, 00:07:54.596 { 00:07:54.596 "name": "BaseBdev3", 00:07:54.596 "uuid": "b8c0b66f-8730-44ce-8a0b-310849b5753d", 00:07:54.596 "is_configured": true, 00:07:54.596 "data_offset": 0, 00:07:54.596 "data_size": 65536 00:07:54.596 } 00:07:54.596 ] 00:07:54.596 } 00:07:54.596 } 00:07:54.596 }' 00:07:54.596 23:41:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:07:54.856 23:41:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:07:54.856 BaseBdev2 00:07:54.856 BaseBdev3' 00:07:54.856 23:41:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:54.856 23:41:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:07:54.856 23:41:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:54.856 23:41:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:07:54.856 23:41:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:54.856 23:41:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:54.856 23:41:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:54.856 23:41:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:54.857 23:41:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:54.857 23:41:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:54.857 23:41:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:54.857 23:41:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:54.857 23:41:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:07:54.857 23:41:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:54.857 23:41:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:54.857 23:41:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:54.857 23:41:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:54.857 23:41:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:54.857 23:41:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:54.857 23:41:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:07:54.857 23:41:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:54.857 23:41:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:54.857 23:41:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:54.857 23:41:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:54.857 23:41:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:54.857 23:41:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:54.857 23:41:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:07:54.857 23:41:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:54.857 23:41:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:54.857 [2024-11-26 23:41:42.906098] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:07:54.857 [2024-11-26 23:41:42.906140] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:54.857 [2024-11-26 23:41:42.906210] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:54.857 [2024-11-26 23:41:42.906266] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:54.857 [2024-11-26 23:41:42.906278] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name Existed_Raid, state offline 00:07:54.857 23:41:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:54.857 23:41:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 76455 00:07:54.857 23:41:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # '[' -z 76455 ']' 00:07:54.857 23:41:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@958 -- # kill -0 76455 00:07:54.857 23:41:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # uname 00:07:54.857 23:41:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:07:54.857 23:41:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 76455 00:07:54.857 23:41:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:07:54.857 23:41:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:07:54.857 23:41:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 76455' 00:07:54.857 killing process with pid 76455 00:07:54.857 23:41:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@973 -- # kill 76455 00:07:54.857 [2024-11-26 23:41:42.942220] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:54.857 23:41:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@978 -- # wait 76455 00:07:54.857 [2024-11-26 23:41:42.972669] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:55.117 23:41:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:07:55.117 00:07:55.117 real 0m8.780s 00:07:55.117 user 0m15.044s 00:07:55.117 sys 0m1.727s 00:07:55.117 23:41:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:07:55.117 ************************************ 00:07:55.117 END TEST raid_state_function_test 00:07:55.117 ************************************ 00:07:55.117 23:41:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:55.117 23:41:43 bdev_raid -- bdev/bdev_raid.sh@969 -- # run_test raid_state_function_test_sb raid_state_function_test concat 3 true 00:07:55.117 23:41:43 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:07:55.117 23:41:43 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:55.118 23:41:43 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:55.378 ************************************ 00:07:55.378 START TEST raid_state_function_test_sb 00:07:55.378 ************************************ 00:07:55.378 23:41:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1129 -- # raid_state_function_test concat 3 true 00:07:55.378 23:41:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=concat 00:07:55.378 23:41:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=3 00:07:55.378 23:41:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:07:55.378 23:41:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:07:55.378 23:41:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:07:55.378 23:41:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:55.378 23:41:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:07:55.378 23:41:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:07:55.378 23:41:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:55.378 23:41:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:07:55.378 23:41:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:07:55.378 23:41:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:55.378 23:41:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:07:55.378 23:41:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:07:55.378 23:41:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:55.378 23:41:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:07:55.378 23:41:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:07:55.378 23:41:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:07:55.378 23:41:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:07:55.378 23:41:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:07:55.378 23:41:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:07:55.378 23:41:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' concat '!=' raid1 ']' 00:07:55.378 23:41:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:07:55.378 23:41:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:07:55.378 23:41:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:07:55.378 23:41:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:07:55.378 23:41:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=77060 00:07:55.378 23:41:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:07:55.378 23:41:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 77060' 00:07:55.378 Process raid pid: 77060 00:07:55.378 23:41:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 77060 00:07:55.378 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:55.378 23:41:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@835 -- # '[' -z 77060 ']' 00:07:55.378 23:41:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:55.378 23:41:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:55.378 23:41:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:55.378 23:41:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:55.378 23:41:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:55.378 [2024-11-26 23:41:43.341833] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:07:55.378 [2024-11-26 23:41:43.341946] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:07:55.378 [2024-11-26 23:41:43.494264] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:55.638 [2024-11-26 23:41:43.521996] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:55.638 [2024-11-26 23:41:43.563452] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:55.638 [2024-11-26 23:41:43.563574] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:56.208 23:41:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:56.208 23:41:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@868 -- # return 0 00:07:56.208 23:41:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:07:56.208 23:41:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:56.208 23:41:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:56.208 [2024-11-26 23:41:44.173474] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:07:56.208 [2024-11-26 23:41:44.173542] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:07:56.208 [2024-11-26 23:41:44.173553] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:56.208 [2024-11-26 23:41:44.173562] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:56.208 [2024-11-26 23:41:44.173568] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:07:56.208 [2024-11-26 23:41:44.173578] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:07:56.208 23:41:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:56.208 23:41:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:07:56.208 23:41:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:56.208 23:41:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:56.208 23:41:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:56.208 23:41:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:56.208 23:41:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:07:56.208 23:41:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:56.208 23:41:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:56.208 23:41:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:56.208 23:41:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:56.208 23:41:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:56.208 23:41:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:56.208 23:41:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:56.208 23:41:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:56.208 23:41:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:56.208 23:41:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:56.208 "name": "Existed_Raid", 00:07:56.208 "uuid": "6b8afc26-93c7-4891-86a1-ee78ed3c8de8", 00:07:56.208 "strip_size_kb": 64, 00:07:56.208 "state": "configuring", 00:07:56.208 "raid_level": "concat", 00:07:56.208 "superblock": true, 00:07:56.208 "num_base_bdevs": 3, 00:07:56.208 "num_base_bdevs_discovered": 0, 00:07:56.208 "num_base_bdevs_operational": 3, 00:07:56.208 "base_bdevs_list": [ 00:07:56.208 { 00:07:56.208 "name": "BaseBdev1", 00:07:56.208 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:56.208 "is_configured": false, 00:07:56.208 "data_offset": 0, 00:07:56.208 "data_size": 0 00:07:56.208 }, 00:07:56.208 { 00:07:56.208 "name": "BaseBdev2", 00:07:56.208 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:56.208 "is_configured": false, 00:07:56.208 "data_offset": 0, 00:07:56.208 "data_size": 0 00:07:56.208 }, 00:07:56.208 { 00:07:56.208 "name": "BaseBdev3", 00:07:56.208 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:56.208 "is_configured": false, 00:07:56.208 "data_offset": 0, 00:07:56.208 "data_size": 0 00:07:56.208 } 00:07:56.208 ] 00:07:56.208 }' 00:07:56.208 23:41:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:56.208 23:41:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:56.779 23:41:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:07:56.779 23:41:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:56.779 23:41:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:56.779 [2024-11-26 23:41:44.616603] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:07:56.779 [2024-11-26 23:41:44.616721] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:07:56.779 23:41:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:56.779 23:41:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:07:56.779 23:41:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:56.779 23:41:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:56.779 [2024-11-26 23:41:44.628602] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:07:56.779 [2024-11-26 23:41:44.628682] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:07:56.779 [2024-11-26 23:41:44.628709] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:56.779 [2024-11-26 23:41:44.628731] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:56.779 [2024-11-26 23:41:44.628749] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:07:56.779 [2024-11-26 23:41:44.628769] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:07:56.779 23:41:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:56.779 23:41:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:07:56.779 23:41:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:56.779 23:41:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:56.779 [2024-11-26 23:41:44.649234] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:56.779 BaseBdev1 00:07:56.779 23:41:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:56.779 23:41:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:07:56.779 23:41:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:07:56.779 23:41:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:07:56.779 23:41:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:07:56.779 23:41:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:07:56.779 23:41:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:07:56.779 23:41:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:07:56.779 23:41:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:56.779 23:41:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:56.779 23:41:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:56.779 23:41:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:07:56.779 23:41:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:56.779 23:41:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:56.779 [ 00:07:56.779 { 00:07:56.779 "name": "BaseBdev1", 00:07:56.779 "aliases": [ 00:07:56.779 "157e4f22-acfc-4fb9-9442-e11a63f572e3" 00:07:56.779 ], 00:07:56.779 "product_name": "Malloc disk", 00:07:56.779 "block_size": 512, 00:07:56.779 "num_blocks": 65536, 00:07:56.779 "uuid": "157e4f22-acfc-4fb9-9442-e11a63f572e3", 00:07:56.779 "assigned_rate_limits": { 00:07:56.779 "rw_ios_per_sec": 0, 00:07:56.779 "rw_mbytes_per_sec": 0, 00:07:56.779 "r_mbytes_per_sec": 0, 00:07:56.779 "w_mbytes_per_sec": 0 00:07:56.779 }, 00:07:56.779 "claimed": true, 00:07:56.779 "claim_type": "exclusive_write", 00:07:56.779 "zoned": false, 00:07:56.779 "supported_io_types": { 00:07:56.779 "read": true, 00:07:56.779 "write": true, 00:07:56.779 "unmap": true, 00:07:56.779 "flush": true, 00:07:56.779 "reset": true, 00:07:56.779 "nvme_admin": false, 00:07:56.779 "nvme_io": false, 00:07:56.779 "nvme_io_md": false, 00:07:56.779 "write_zeroes": true, 00:07:56.779 "zcopy": true, 00:07:56.779 "get_zone_info": false, 00:07:56.779 "zone_management": false, 00:07:56.779 "zone_append": false, 00:07:56.779 "compare": false, 00:07:56.779 "compare_and_write": false, 00:07:56.779 "abort": true, 00:07:56.779 "seek_hole": false, 00:07:56.779 "seek_data": false, 00:07:56.779 "copy": true, 00:07:56.779 "nvme_iov_md": false 00:07:56.779 }, 00:07:56.779 "memory_domains": [ 00:07:56.779 { 00:07:56.779 "dma_device_id": "system", 00:07:56.779 "dma_device_type": 1 00:07:56.779 }, 00:07:56.779 { 00:07:56.779 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:56.779 "dma_device_type": 2 00:07:56.779 } 00:07:56.779 ], 00:07:56.779 "driver_specific": {} 00:07:56.779 } 00:07:56.779 ] 00:07:56.779 23:41:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:56.779 23:41:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:07:56.779 23:41:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:07:56.779 23:41:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:56.779 23:41:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:56.779 23:41:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:56.779 23:41:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:56.779 23:41:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:07:56.779 23:41:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:56.779 23:41:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:56.779 23:41:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:56.779 23:41:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:56.779 23:41:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:56.779 23:41:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:56.779 23:41:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:56.779 23:41:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:56.779 23:41:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:56.779 23:41:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:56.779 "name": "Existed_Raid", 00:07:56.779 "uuid": "ca590174-7d2e-427d-81c0-d09f0432f146", 00:07:56.779 "strip_size_kb": 64, 00:07:56.779 "state": "configuring", 00:07:56.779 "raid_level": "concat", 00:07:56.779 "superblock": true, 00:07:56.779 "num_base_bdevs": 3, 00:07:56.779 "num_base_bdevs_discovered": 1, 00:07:56.779 "num_base_bdevs_operational": 3, 00:07:56.779 "base_bdevs_list": [ 00:07:56.779 { 00:07:56.779 "name": "BaseBdev1", 00:07:56.779 "uuid": "157e4f22-acfc-4fb9-9442-e11a63f572e3", 00:07:56.779 "is_configured": true, 00:07:56.779 "data_offset": 2048, 00:07:56.779 "data_size": 63488 00:07:56.779 }, 00:07:56.779 { 00:07:56.780 "name": "BaseBdev2", 00:07:56.780 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:56.780 "is_configured": false, 00:07:56.780 "data_offset": 0, 00:07:56.780 "data_size": 0 00:07:56.780 }, 00:07:56.780 { 00:07:56.780 "name": "BaseBdev3", 00:07:56.780 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:56.780 "is_configured": false, 00:07:56.780 "data_offset": 0, 00:07:56.780 "data_size": 0 00:07:56.780 } 00:07:56.780 ] 00:07:56.780 }' 00:07:56.780 23:41:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:56.780 23:41:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:57.039 23:41:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:07:57.039 23:41:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:57.039 23:41:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:57.039 [2024-11-26 23:41:45.148438] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:07:57.039 [2024-11-26 23:41:45.148546] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:07:57.039 23:41:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:57.039 23:41:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:07:57.039 23:41:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:57.039 23:41:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:57.039 [2024-11-26 23:41:45.160460] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:57.039 [2024-11-26 23:41:45.162424] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:57.039 [2024-11-26 23:41:45.162500] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:57.039 [2024-11-26 23:41:45.162529] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:07:57.039 [2024-11-26 23:41:45.162556] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:07:57.039 23:41:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:57.039 23:41:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:07:57.039 23:41:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:07:57.039 23:41:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:07:57.039 23:41:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:57.039 23:41:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:57.039 23:41:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:57.039 23:41:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:57.299 23:41:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:07:57.299 23:41:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:57.299 23:41:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:57.299 23:41:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:57.299 23:41:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:57.299 23:41:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:57.299 23:41:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:57.299 23:41:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:57.299 23:41:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:57.299 23:41:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:57.299 23:41:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:57.299 "name": "Existed_Raid", 00:07:57.299 "uuid": "3ecac646-309a-4e91-825d-2376439173fa", 00:07:57.299 "strip_size_kb": 64, 00:07:57.299 "state": "configuring", 00:07:57.299 "raid_level": "concat", 00:07:57.299 "superblock": true, 00:07:57.299 "num_base_bdevs": 3, 00:07:57.299 "num_base_bdevs_discovered": 1, 00:07:57.299 "num_base_bdevs_operational": 3, 00:07:57.299 "base_bdevs_list": [ 00:07:57.299 { 00:07:57.299 "name": "BaseBdev1", 00:07:57.299 "uuid": "157e4f22-acfc-4fb9-9442-e11a63f572e3", 00:07:57.299 "is_configured": true, 00:07:57.299 "data_offset": 2048, 00:07:57.299 "data_size": 63488 00:07:57.299 }, 00:07:57.299 { 00:07:57.299 "name": "BaseBdev2", 00:07:57.299 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:57.299 "is_configured": false, 00:07:57.299 "data_offset": 0, 00:07:57.299 "data_size": 0 00:07:57.299 }, 00:07:57.299 { 00:07:57.299 "name": "BaseBdev3", 00:07:57.299 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:57.299 "is_configured": false, 00:07:57.299 "data_offset": 0, 00:07:57.299 "data_size": 0 00:07:57.299 } 00:07:57.299 ] 00:07:57.299 }' 00:07:57.299 23:41:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:57.299 23:41:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:57.559 23:41:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:07:57.559 23:41:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:57.559 23:41:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:57.559 BaseBdev2 00:07:57.559 [2024-11-26 23:41:45.622435] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:07:57.559 23:41:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:57.559 23:41:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:07:57.559 23:41:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:07:57.559 23:41:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:07:57.559 23:41:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:07:57.559 23:41:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:07:57.559 23:41:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:07:57.559 23:41:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:07:57.559 23:41:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:57.559 23:41:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:57.559 23:41:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:57.559 23:41:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:07:57.559 23:41:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:57.559 23:41:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:57.559 [ 00:07:57.559 { 00:07:57.559 "name": "BaseBdev2", 00:07:57.559 "aliases": [ 00:07:57.559 "a622accc-4587-46b3-a71f-97495dcb114e" 00:07:57.559 ], 00:07:57.559 "product_name": "Malloc disk", 00:07:57.559 "block_size": 512, 00:07:57.559 "num_blocks": 65536, 00:07:57.559 "uuid": "a622accc-4587-46b3-a71f-97495dcb114e", 00:07:57.559 "assigned_rate_limits": { 00:07:57.559 "rw_ios_per_sec": 0, 00:07:57.559 "rw_mbytes_per_sec": 0, 00:07:57.559 "r_mbytes_per_sec": 0, 00:07:57.559 "w_mbytes_per_sec": 0 00:07:57.559 }, 00:07:57.559 "claimed": true, 00:07:57.559 "claim_type": "exclusive_write", 00:07:57.559 "zoned": false, 00:07:57.559 "supported_io_types": { 00:07:57.559 "read": true, 00:07:57.559 "write": true, 00:07:57.559 "unmap": true, 00:07:57.559 "flush": true, 00:07:57.560 "reset": true, 00:07:57.560 "nvme_admin": false, 00:07:57.560 "nvme_io": false, 00:07:57.560 "nvme_io_md": false, 00:07:57.560 "write_zeroes": true, 00:07:57.560 "zcopy": true, 00:07:57.560 "get_zone_info": false, 00:07:57.560 "zone_management": false, 00:07:57.560 "zone_append": false, 00:07:57.560 "compare": false, 00:07:57.560 "compare_and_write": false, 00:07:57.560 "abort": true, 00:07:57.560 "seek_hole": false, 00:07:57.560 "seek_data": false, 00:07:57.560 "copy": true, 00:07:57.560 "nvme_iov_md": false 00:07:57.560 }, 00:07:57.560 "memory_domains": [ 00:07:57.560 { 00:07:57.560 "dma_device_id": "system", 00:07:57.560 "dma_device_type": 1 00:07:57.560 }, 00:07:57.560 { 00:07:57.560 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:57.560 "dma_device_type": 2 00:07:57.560 } 00:07:57.560 ], 00:07:57.560 "driver_specific": {} 00:07:57.560 } 00:07:57.560 ] 00:07:57.560 23:41:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:57.560 23:41:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:07:57.560 23:41:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:07:57.560 23:41:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:07:57.560 23:41:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:07:57.560 23:41:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:57.560 23:41:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:57.560 23:41:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:57.560 23:41:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:57.560 23:41:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:07:57.560 23:41:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:57.560 23:41:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:57.560 23:41:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:57.560 23:41:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:57.560 23:41:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:57.560 23:41:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:57.560 23:41:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:57.560 23:41:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:57.560 23:41:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:57.820 23:41:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:57.820 "name": "Existed_Raid", 00:07:57.820 "uuid": "3ecac646-309a-4e91-825d-2376439173fa", 00:07:57.820 "strip_size_kb": 64, 00:07:57.820 "state": "configuring", 00:07:57.820 "raid_level": "concat", 00:07:57.820 "superblock": true, 00:07:57.820 "num_base_bdevs": 3, 00:07:57.820 "num_base_bdevs_discovered": 2, 00:07:57.820 "num_base_bdevs_operational": 3, 00:07:57.820 "base_bdevs_list": [ 00:07:57.820 { 00:07:57.820 "name": "BaseBdev1", 00:07:57.820 "uuid": "157e4f22-acfc-4fb9-9442-e11a63f572e3", 00:07:57.820 "is_configured": true, 00:07:57.820 "data_offset": 2048, 00:07:57.820 "data_size": 63488 00:07:57.820 }, 00:07:57.820 { 00:07:57.820 "name": "BaseBdev2", 00:07:57.820 "uuid": "a622accc-4587-46b3-a71f-97495dcb114e", 00:07:57.820 "is_configured": true, 00:07:57.820 "data_offset": 2048, 00:07:57.820 "data_size": 63488 00:07:57.820 }, 00:07:57.820 { 00:07:57.820 "name": "BaseBdev3", 00:07:57.820 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:57.820 "is_configured": false, 00:07:57.820 "data_offset": 0, 00:07:57.820 "data_size": 0 00:07:57.820 } 00:07:57.820 ] 00:07:57.820 }' 00:07:57.820 23:41:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:57.820 23:41:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:58.080 23:41:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:07:58.080 23:41:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:58.080 23:41:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:58.080 [2024-11-26 23:41:46.132938] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:07:58.080 [2024-11-26 23:41:46.133232] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:07:58.080 [2024-11-26 23:41:46.133299] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:07:58.080 [2024-11-26 23:41:46.133686] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:07:58.080 BaseBdev3 00:07:58.080 [2024-11-26 23:41:46.133891] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:07:58.080 [2024-11-26 23:41:46.133906] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:07:58.080 [2024-11-26 23:41:46.134040] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:58.080 23:41:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:58.080 23:41:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:07:58.080 23:41:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:07:58.081 23:41:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:07:58.081 23:41:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:07:58.081 23:41:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:07:58.081 23:41:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:07:58.081 23:41:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:07:58.081 23:41:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:58.081 23:41:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:58.081 23:41:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:58.081 23:41:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:07:58.081 23:41:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:58.081 23:41:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:58.081 [ 00:07:58.081 { 00:07:58.081 "name": "BaseBdev3", 00:07:58.081 "aliases": [ 00:07:58.081 "73bc6a6f-5a67-425a-a28a-37df54e4c0e0" 00:07:58.081 ], 00:07:58.081 "product_name": "Malloc disk", 00:07:58.081 "block_size": 512, 00:07:58.081 "num_blocks": 65536, 00:07:58.081 "uuid": "73bc6a6f-5a67-425a-a28a-37df54e4c0e0", 00:07:58.081 "assigned_rate_limits": { 00:07:58.081 "rw_ios_per_sec": 0, 00:07:58.081 "rw_mbytes_per_sec": 0, 00:07:58.081 "r_mbytes_per_sec": 0, 00:07:58.081 "w_mbytes_per_sec": 0 00:07:58.081 }, 00:07:58.081 "claimed": true, 00:07:58.081 "claim_type": "exclusive_write", 00:07:58.081 "zoned": false, 00:07:58.081 "supported_io_types": { 00:07:58.081 "read": true, 00:07:58.081 "write": true, 00:07:58.081 "unmap": true, 00:07:58.081 "flush": true, 00:07:58.081 "reset": true, 00:07:58.081 "nvme_admin": false, 00:07:58.081 "nvme_io": false, 00:07:58.081 "nvme_io_md": false, 00:07:58.081 "write_zeroes": true, 00:07:58.081 "zcopy": true, 00:07:58.081 "get_zone_info": false, 00:07:58.081 "zone_management": false, 00:07:58.081 "zone_append": false, 00:07:58.081 "compare": false, 00:07:58.081 "compare_and_write": false, 00:07:58.081 "abort": true, 00:07:58.081 "seek_hole": false, 00:07:58.081 "seek_data": false, 00:07:58.081 "copy": true, 00:07:58.081 "nvme_iov_md": false 00:07:58.081 }, 00:07:58.081 "memory_domains": [ 00:07:58.081 { 00:07:58.081 "dma_device_id": "system", 00:07:58.081 "dma_device_type": 1 00:07:58.081 }, 00:07:58.081 { 00:07:58.081 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:58.081 "dma_device_type": 2 00:07:58.081 } 00:07:58.081 ], 00:07:58.081 "driver_specific": {} 00:07:58.081 } 00:07:58.081 ] 00:07:58.081 23:41:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:58.081 23:41:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:07:58.081 23:41:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:07:58.081 23:41:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:07:58.081 23:41:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online concat 64 3 00:07:58.081 23:41:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:58.081 23:41:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:58.081 23:41:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:58.081 23:41:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:58.081 23:41:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:07:58.081 23:41:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:58.081 23:41:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:58.081 23:41:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:58.081 23:41:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:58.081 23:41:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:58.081 23:41:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:58.081 23:41:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:58.081 23:41:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:58.081 23:41:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:58.339 23:41:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:58.339 "name": "Existed_Raid", 00:07:58.339 "uuid": "3ecac646-309a-4e91-825d-2376439173fa", 00:07:58.339 "strip_size_kb": 64, 00:07:58.339 "state": "online", 00:07:58.339 "raid_level": "concat", 00:07:58.339 "superblock": true, 00:07:58.339 "num_base_bdevs": 3, 00:07:58.339 "num_base_bdevs_discovered": 3, 00:07:58.339 "num_base_bdevs_operational": 3, 00:07:58.339 "base_bdevs_list": [ 00:07:58.339 { 00:07:58.339 "name": "BaseBdev1", 00:07:58.339 "uuid": "157e4f22-acfc-4fb9-9442-e11a63f572e3", 00:07:58.339 "is_configured": true, 00:07:58.339 "data_offset": 2048, 00:07:58.339 "data_size": 63488 00:07:58.339 }, 00:07:58.339 { 00:07:58.339 "name": "BaseBdev2", 00:07:58.339 "uuid": "a622accc-4587-46b3-a71f-97495dcb114e", 00:07:58.339 "is_configured": true, 00:07:58.339 "data_offset": 2048, 00:07:58.339 "data_size": 63488 00:07:58.339 }, 00:07:58.339 { 00:07:58.339 "name": "BaseBdev3", 00:07:58.339 "uuid": "73bc6a6f-5a67-425a-a28a-37df54e4c0e0", 00:07:58.339 "is_configured": true, 00:07:58.339 "data_offset": 2048, 00:07:58.339 "data_size": 63488 00:07:58.339 } 00:07:58.339 ] 00:07:58.339 }' 00:07:58.339 23:41:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:58.339 23:41:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:58.599 23:41:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:07:58.599 23:41:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:07:58.599 23:41:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:07:58.599 23:41:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:07:58.599 23:41:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:07:58.599 23:41:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:07:58.599 23:41:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:07:58.599 23:41:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:07:58.599 23:41:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:58.599 23:41:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:58.599 [2024-11-26 23:41:46.628460] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:58.599 23:41:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:58.599 23:41:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:07:58.599 "name": "Existed_Raid", 00:07:58.599 "aliases": [ 00:07:58.599 "3ecac646-309a-4e91-825d-2376439173fa" 00:07:58.599 ], 00:07:58.599 "product_name": "Raid Volume", 00:07:58.599 "block_size": 512, 00:07:58.599 "num_blocks": 190464, 00:07:58.599 "uuid": "3ecac646-309a-4e91-825d-2376439173fa", 00:07:58.599 "assigned_rate_limits": { 00:07:58.599 "rw_ios_per_sec": 0, 00:07:58.599 "rw_mbytes_per_sec": 0, 00:07:58.599 "r_mbytes_per_sec": 0, 00:07:58.599 "w_mbytes_per_sec": 0 00:07:58.599 }, 00:07:58.599 "claimed": false, 00:07:58.599 "zoned": false, 00:07:58.599 "supported_io_types": { 00:07:58.599 "read": true, 00:07:58.599 "write": true, 00:07:58.599 "unmap": true, 00:07:58.599 "flush": true, 00:07:58.599 "reset": true, 00:07:58.599 "nvme_admin": false, 00:07:58.599 "nvme_io": false, 00:07:58.599 "nvme_io_md": false, 00:07:58.599 "write_zeroes": true, 00:07:58.599 "zcopy": false, 00:07:58.599 "get_zone_info": false, 00:07:58.599 "zone_management": false, 00:07:58.599 "zone_append": false, 00:07:58.599 "compare": false, 00:07:58.599 "compare_and_write": false, 00:07:58.599 "abort": false, 00:07:58.599 "seek_hole": false, 00:07:58.599 "seek_data": false, 00:07:58.599 "copy": false, 00:07:58.599 "nvme_iov_md": false 00:07:58.599 }, 00:07:58.599 "memory_domains": [ 00:07:58.599 { 00:07:58.599 "dma_device_id": "system", 00:07:58.599 "dma_device_type": 1 00:07:58.599 }, 00:07:58.599 { 00:07:58.599 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:58.599 "dma_device_type": 2 00:07:58.599 }, 00:07:58.599 { 00:07:58.599 "dma_device_id": "system", 00:07:58.599 "dma_device_type": 1 00:07:58.599 }, 00:07:58.599 { 00:07:58.599 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:58.599 "dma_device_type": 2 00:07:58.599 }, 00:07:58.599 { 00:07:58.599 "dma_device_id": "system", 00:07:58.599 "dma_device_type": 1 00:07:58.599 }, 00:07:58.599 { 00:07:58.599 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:58.599 "dma_device_type": 2 00:07:58.599 } 00:07:58.599 ], 00:07:58.599 "driver_specific": { 00:07:58.599 "raid": { 00:07:58.599 "uuid": "3ecac646-309a-4e91-825d-2376439173fa", 00:07:58.599 "strip_size_kb": 64, 00:07:58.599 "state": "online", 00:07:58.599 "raid_level": "concat", 00:07:58.599 "superblock": true, 00:07:58.599 "num_base_bdevs": 3, 00:07:58.599 "num_base_bdevs_discovered": 3, 00:07:58.599 "num_base_bdevs_operational": 3, 00:07:58.599 "base_bdevs_list": [ 00:07:58.599 { 00:07:58.599 "name": "BaseBdev1", 00:07:58.599 "uuid": "157e4f22-acfc-4fb9-9442-e11a63f572e3", 00:07:58.599 "is_configured": true, 00:07:58.599 "data_offset": 2048, 00:07:58.599 "data_size": 63488 00:07:58.599 }, 00:07:58.599 { 00:07:58.599 "name": "BaseBdev2", 00:07:58.599 "uuid": "a622accc-4587-46b3-a71f-97495dcb114e", 00:07:58.599 "is_configured": true, 00:07:58.599 "data_offset": 2048, 00:07:58.599 "data_size": 63488 00:07:58.599 }, 00:07:58.599 { 00:07:58.599 "name": "BaseBdev3", 00:07:58.599 "uuid": "73bc6a6f-5a67-425a-a28a-37df54e4c0e0", 00:07:58.599 "is_configured": true, 00:07:58.599 "data_offset": 2048, 00:07:58.599 "data_size": 63488 00:07:58.599 } 00:07:58.599 ] 00:07:58.599 } 00:07:58.599 } 00:07:58.599 }' 00:07:58.599 23:41:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:07:58.599 23:41:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:07:58.599 BaseBdev2 00:07:58.599 BaseBdev3' 00:07:58.599 23:41:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:58.860 23:41:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:07:58.860 23:41:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:58.860 23:41:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:07:58.860 23:41:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:58.860 23:41:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:58.860 23:41:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:58.860 23:41:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:58.860 23:41:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:58.860 23:41:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:58.860 23:41:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:58.860 23:41:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:07:58.860 23:41:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:58.860 23:41:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:58.860 23:41:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:58.860 23:41:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:58.860 23:41:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:58.860 23:41:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:58.860 23:41:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:58.860 23:41:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:07:58.860 23:41:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:58.860 23:41:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:58.860 23:41:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:58.860 23:41:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:58.860 23:41:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:58.860 23:41:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:58.860 23:41:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:07:58.860 23:41:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:58.860 23:41:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:58.860 [2024-11-26 23:41:46.919671] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:07:58.860 [2024-11-26 23:41:46.919746] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:58.860 [2024-11-26 23:41:46.919843] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:58.860 23:41:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:58.860 23:41:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:07:58.860 23:41:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy concat 00:07:58.860 23:41:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:07:58.860 23:41:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # return 1 00:07:58.860 23:41:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:07:58.860 23:41:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline concat 64 2 00:07:58.860 23:41:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:58.860 23:41:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:07:58.860 23:41:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:58.860 23:41:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:58.860 23:41:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:58.860 23:41:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:58.860 23:41:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:58.860 23:41:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:58.860 23:41:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:58.860 23:41:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:58.860 23:41:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:58.860 23:41:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:58.860 23:41:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:58.860 23:41:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:58.860 23:41:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:58.860 "name": "Existed_Raid", 00:07:58.860 "uuid": "3ecac646-309a-4e91-825d-2376439173fa", 00:07:58.860 "strip_size_kb": 64, 00:07:58.860 "state": "offline", 00:07:58.860 "raid_level": "concat", 00:07:58.860 "superblock": true, 00:07:58.860 "num_base_bdevs": 3, 00:07:58.860 "num_base_bdevs_discovered": 2, 00:07:58.860 "num_base_bdevs_operational": 2, 00:07:58.860 "base_bdevs_list": [ 00:07:58.860 { 00:07:58.860 "name": null, 00:07:58.860 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:58.860 "is_configured": false, 00:07:58.860 "data_offset": 0, 00:07:58.860 "data_size": 63488 00:07:58.860 }, 00:07:58.860 { 00:07:58.860 "name": "BaseBdev2", 00:07:58.860 "uuid": "a622accc-4587-46b3-a71f-97495dcb114e", 00:07:58.860 "is_configured": true, 00:07:58.860 "data_offset": 2048, 00:07:58.860 "data_size": 63488 00:07:58.860 }, 00:07:58.860 { 00:07:58.860 "name": "BaseBdev3", 00:07:58.860 "uuid": "73bc6a6f-5a67-425a-a28a-37df54e4c0e0", 00:07:58.860 "is_configured": true, 00:07:58.860 "data_offset": 2048, 00:07:58.860 "data_size": 63488 00:07:58.860 } 00:07:58.860 ] 00:07:58.860 }' 00:07:58.860 23:41:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:58.860 23:41:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:59.430 23:41:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:07:59.430 23:41:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:07:59.430 23:41:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:07:59.430 23:41:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:59.430 23:41:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:59.430 23:41:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:59.430 23:41:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:59.430 23:41:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:07:59.430 23:41:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:07:59.430 23:41:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:07:59.430 23:41:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:59.430 23:41:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:59.430 [2024-11-26 23:41:47.402306] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:07:59.430 23:41:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:59.430 23:41:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:07:59.430 23:41:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:07:59.430 23:41:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:59.430 23:41:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:59.430 23:41:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:59.431 23:41:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:07:59.431 23:41:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:59.431 23:41:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:07:59.431 23:41:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:07:59.431 23:41:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:07:59.431 23:41:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:59.431 23:41:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:59.431 [2024-11-26 23:41:47.473384] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:07:59.431 [2024-11-26 23:41:47.473429] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:07:59.431 23:41:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:59.431 23:41:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:07:59.431 23:41:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:07:59.431 23:41:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:59.431 23:41:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:59.431 23:41:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:59.431 23:41:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:07:59.431 23:41:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:59.431 23:41:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:07:59.431 23:41:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:07:59.431 23:41:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 3 -gt 2 ']' 00:07:59.431 23:41:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:07:59.431 23:41:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:07:59.431 23:41:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:07:59.431 23:41:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:59.431 23:41:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:59.431 BaseBdev2 00:07:59.431 23:41:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:59.431 23:41:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:07:59.431 23:41:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:07:59.431 23:41:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:07:59.431 23:41:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:07:59.431 23:41:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:07:59.431 23:41:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:07:59.431 23:41:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:07:59.431 23:41:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:59.431 23:41:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:59.692 23:41:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:59.692 23:41:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:07:59.692 23:41:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:59.692 23:41:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:59.692 [ 00:07:59.692 { 00:07:59.692 "name": "BaseBdev2", 00:07:59.692 "aliases": [ 00:07:59.692 "a77e9cec-136e-4839-b9c7-5bd5054f9619" 00:07:59.692 ], 00:07:59.692 "product_name": "Malloc disk", 00:07:59.692 "block_size": 512, 00:07:59.692 "num_blocks": 65536, 00:07:59.692 "uuid": "a77e9cec-136e-4839-b9c7-5bd5054f9619", 00:07:59.692 "assigned_rate_limits": { 00:07:59.692 "rw_ios_per_sec": 0, 00:07:59.692 "rw_mbytes_per_sec": 0, 00:07:59.692 "r_mbytes_per_sec": 0, 00:07:59.692 "w_mbytes_per_sec": 0 00:07:59.692 }, 00:07:59.692 "claimed": false, 00:07:59.692 "zoned": false, 00:07:59.692 "supported_io_types": { 00:07:59.692 "read": true, 00:07:59.692 "write": true, 00:07:59.692 "unmap": true, 00:07:59.692 "flush": true, 00:07:59.692 "reset": true, 00:07:59.692 "nvme_admin": false, 00:07:59.692 "nvme_io": false, 00:07:59.692 "nvme_io_md": false, 00:07:59.692 "write_zeroes": true, 00:07:59.692 "zcopy": true, 00:07:59.692 "get_zone_info": false, 00:07:59.692 "zone_management": false, 00:07:59.692 "zone_append": false, 00:07:59.692 "compare": false, 00:07:59.692 "compare_and_write": false, 00:07:59.692 "abort": true, 00:07:59.692 "seek_hole": false, 00:07:59.692 "seek_data": false, 00:07:59.692 "copy": true, 00:07:59.692 "nvme_iov_md": false 00:07:59.692 }, 00:07:59.692 "memory_domains": [ 00:07:59.692 { 00:07:59.692 "dma_device_id": "system", 00:07:59.692 "dma_device_type": 1 00:07:59.692 }, 00:07:59.692 { 00:07:59.692 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:59.692 "dma_device_type": 2 00:07:59.692 } 00:07:59.692 ], 00:07:59.692 "driver_specific": {} 00:07:59.692 } 00:07:59.692 ] 00:07:59.692 23:41:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:59.692 23:41:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:07:59.692 23:41:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:07:59.692 23:41:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:07:59.692 23:41:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:07:59.692 23:41:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:59.692 23:41:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:59.692 BaseBdev3 00:07:59.692 23:41:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:59.692 23:41:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:07:59.692 23:41:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:07:59.692 23:41:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:07:59.692 23:41:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:07:59.692 23:41:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:07:59.692 23:41:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:07:59.692 23:41:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:07:59.692 23:41:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:59.692 23:41:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:59.692 23:41:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:59.692 23:41:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:07:59.692 23:41:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:59.692 23:41:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:59.692 [ 00:07:59.692 { 00:07:59.692 "name": "BaseBdev3", 00:07:59.692 "aliases": [ 00:07:59.692 "1759d92b-10b4-4fd1-86c7-f8de8cc19125" 00:07:59.692 ], 00:07:59.692 "product_name": "Malloc disk", 00:07:59.692 "block_size": 512, 00:07:59.692 "num_blocks": 65536, 00:07:59.692 "uuid": "1759d92b-10b4-4fd1-86c7-f8de8cc19125", 00:07:59.692 "assigned_rate_limits": { 00:07:59.692 "rw_ios_per_sec": 0, 00:07:59.692 "rw_mbytes_per_sec": 0, 00:07:59.692 "r_mbytes_per_sec": 0, 00:07:59.692 "w_mbytes_per_sec": 0 00:07:59.692 }, 00:07:59.692 "claimed": false, 00:07:59.692 "zoned": false, 00:07:59.692 "supported_io_types": { 00:07:59.692 "read": true, 00:07:59.692 "write": true, 00:07:59.692 "unmap": true, 00:07:59.692 "flush": true, 00:07:59.692 "reset": true, 00:07:59.692 "nvme_admin": false, 00:07:59.692 "nvme_io": false, 00:07:59.692 "nvme_io_md": false, 00:07:59.692 "write_zeroes": true, 00:07:59.692 "zcopy": true, 00:07:59.692 "get_zone_info": false, 00:07:59.692 "zone_management": false, 00:07:59.692 "zone_append": false, 00:07:59.692 "compare": false, 00:07:59.692 "compare_and_write": false, 00:07:59.692 "abort": true, 00:07:59.692 "seek_hole": false, 00:07:59.692 "seek_data": false, 00:07:59.692 "copy": true, 00:07:59.692 "nvme_iov_md": false 00:07:59.692 }, 00:07:59.692 "memory_domains": [ 00:07:59.692 { 00:07:59.692 "dma_device_id": "system", 00:07:59.692 "dma_device_type": 1 00:07:59.692 }, 00:07:59.692 { 00:07:59.692 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:59.692 "dma_device_type": 2 00:07:59.692 } 00:07:59.692 ], 00:07:59.692 "driver_specific": {} 00:07:59.692 } 00:07:59.692 ] 00:07:59.692 23:41:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:59.692 23:41:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:07:59.692 23:41:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:07:59.692 23:41:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:07:59.692 23:41:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:07:59.692 23:41:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:59.692 23:41:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:59.692 [2024-11-26 23:41:47.652491] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:07:59.693 [2024-11-26 23:41:47.652533] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:07:59.693 [2024-11-26 23:41:47.652554] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:07:59.693 [2024-11-26 23:41:47.654418] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:07:59.693 23:41:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:59.693 23:41:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:07:59.693 23:41:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:59.693 23:41:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:59.693 23:41:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:59.693 23:41:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:59.693 23:41:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:07:59.693 23:41:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:59.693 23:41:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:59.693 23:41:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:59.693 23:41:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:59.693 23:41:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:59.693 23:41:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:59.693 23:41:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:59.693 23:41:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:59.693 23:41:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:59.693 23:41:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:59.693 "name": "Existed_Raid", 00:07:59.693 "uuid": "900257f7-5c31-4f25-9ca0-9d817eea607f", 00:07:59.693 "strip_size_kb": 64, 00:07:59.693 "state": "configuring", 00:07:59.693 "raid_level": "concat", 00:07:59.693 "superblock": true, 00:07:59.693 "num_base_bdevs": 3, 00:07:59.693 "num_base_bdevs_discovered": 2, 00:07:59.693 "num_base_bdevs_operational": 3, 00:07:59.693 "base_bdevs_list": [ 00:07:59.693 { 00:07:59.693 "name": "BaseBdev1", 00:07:59.693 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:59.693 "is_configured": false, 00:07:59.693 "data_offset": 0, 00:07:59.693 "data_size": 0 00:07:59.693 }, 00:07:59.693 { 00:07:59.693 "name": "BaseBdev2", 00:07:59.693 "uuid": "a77e9cec-136e-4839-b9c7-5bd5054f9619", 00:07:59.693 "is_configured": true, 00:07:59.693 "data_offset": 2048, 00:07:59.693 "data_size": 63488 00:07:59.693 }, 00:07:59.693 { 00:07:59.693 "name": "BaseBdev3", 00:07:59.693 "uuid": "1759d92b-10b4-4fd1-86c7-f8de8cc19125", 00:07:59.693 "is_configured": true, 00:07:59.693 "data_offset": 2048, 00:07:59.693 "data_size": 63488 00:07:59.693 } 00:07:59.693 ] 00:07:59.693 }' 00:07:59.693 23:41:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:59.693 23:41:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:59.953 23:41:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:07:59.953 23:41:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:59.953 23:41:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:59.953 [2024-11-26 23:41:48.071793] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:07:59.953 23:41:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:59.953 23:41:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:07:59.953 23:41:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:59.953 23:41:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:59.953 23:41:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:59.953 23:41:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:59.953 23:41:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:07:59.953 23:41:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:59.953 23:41:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:59.953 23:41:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:59.953 23:41:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:59.953 23:41:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:00.213 23:41:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:00.213 23:41:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:00.213 23:41:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:00.213 23:41:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:00.213 23:41:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:00.213 "name": "Existed_Raid", 00:08:00.213 "uuid": "900257f7-5c31-4f25-9ca0-9d817eea607f", 00:08:00.213 "strip_size_kb": 64, 00:08:00.213 "state": "configuring", 00:08:00.213 "raid_level": "concat", 00:08:00.213 "superblock": true, 00:08:00.213 "num_base_bdevs": 3, 00:08:00.213 "num_base_bdevs_discovered": 1, 00:08:00.213 "num_base_bdevs_operational": 3, 00:08:00.213 "base_bdevs_list": [ 00:08:00.213 { 00:08:00.213 "name": "BaseBdev1", 00:08:00.213 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:00.213 "is_configured": false, 00:08:00.213 "data_offset": 0, 00:08:00.213 "data_size": 0 00:08:00.213 }, 00:08:00.213 { 00:08:00.213 "name": null, 00:08:00.213 "uuid": "a77e9cec-136e-4839-b9c7-5bd5054f9619", 00:08:00.213 "is_configured": false, 00:08:00.213 "data_offset": 0, 00:08:00.213 "data_size": 63488 00:08:00.213 }, 00:08:00.213 { 00:08:00.213 "name": "BaseBdev3", 00:08:00.213 "uuid": "1759d92b-10b4-4fd1-86c7-f8de8cc19125", 00:08:00.214 "is_configured": true, 00:08:00.214 "data_offset": 2048, 00:08:00.214 "data_size": 63488 00:08:00.214 } 00:08:00.214 ] 00:08:00.214 }' 00:08:00.214 23:41:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:00.214 23:41:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:00.473 23:41:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:00.473 23:41:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:00.473 23:41:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:00.473 23:41:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:08:00.473 23:41:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:00.473 23:41:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:08:00.473 23:41:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:08:00.473 23:41:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:00.473 23:41:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:00.473 [2024-11-26 23:41:48.597744] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:00.473 BaseBdev1 00:08:00.473 23:41:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:00.473 23:41:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:08:00.473 23:41:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:08:00.473 23:41:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:00.473 23:41:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:08:00.473 23:41:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:00.473 23:41:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:00.473 23:41:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:00.473 23:41:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:00.473 23:41:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:00.741 23:41:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:00.741 23:41:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:08:00.741 23:41:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:00.741 23:41:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:00.741 [ 00:08:00.741 { 00:08:00.741 "name": "BaseBdev1", 00:08:00.741 "aliases": [ 00:08:00.741 "ab84e5f1-7127-4ff1-b2cb-a5cf28bead2e" 00:08:00.741 ], 00:08:00.741 "product_name": "Malloc disk", 00:08:00.741 "block_size": 512, 00:08:00.741 "num_blocks": 65536, 00:08:00.741 "uuid": "ab84e5f1-7127-4ff1-b2cb-a5cf28bead2e", 00:08:00.741 "assigned_rate_limits": { 00:08:00.741 "rw_ios_per_sec": 0, 00:08:00.741 "rw_mbytes_per_sec": 0, 00:08:00.741 "r_mbytes_per_sec": 0, 00:08:00.741 "w_mbytes_per_sec": 0 00:08:00.741 }, 00:08:00.741 "claimed": true, 00:08:00.741 "claim_type": "exclusive_write", 00:08:00.741 "zoned": false, 00:08:00.741 "supported_io_types": { 00:08:00.741 "read": true, 00:08:00.741 "write": true, 00:08:00.741 "unmap": true, 00:08:00.741 "flush": true, 00:08:00.741 "reset": true, 00:08:00.741 "nvme_admin": false, 00:08:00.741 "nvme_io": false, 00:08:00.741 "nvme_io_md": false, 00:08:00.741 "write_zeroes": true, 00:08:00.741 "zcopy": true, 00:08:00.741 "get_zone_info": false, 00:08:00.741 "zone_management": false, 00:08:00.742 "zone_append": false, 00:08:00.742 "compare": false, 00:08:00.742 "compare_and_write": false, 00:08:00.742 "abort": true, 00:08:00.742 "seek_hole": false, 00:08:00.742 "seek_data": false, 00:08:00.742 "copy": true, 00:08:00.742 "nvme_iov_md": false 00:08:00.742 }, 00:08:00.742 "memory_domains": [ 00:08:00.742 { 00:08:00.742 "dma_device_id": "system", 00:08:00.742 "dma_device_type": 1 00:08:00.742 }, 00:08:00.742 { 00:08:00.742 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:00.742 "dma_device_type": 2 00:08:00.742 } 00:08:00.742 ], 00:08:00.742 "driver_specific": {} 00:08:00.742 } 00:08:00.742 ] 00:08:00.742 23:41:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:00.742 23:41:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:08:00.742 23:41:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:00.742 23:41:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:00.742 23:41:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:00.742 23:41:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:00.742 23:41:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:00.742 23:41:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:00.742 23:41:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:00.742 23:41:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:00.742 23:41:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:00.742 23:41:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:00.742 23:41:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:00.742 23:41:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:00.742 23:41:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:00.742 23:41:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:00.742 23:41:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:00.742 23:41:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:00.742 "name": "Existed_Raid", 00:08:00.742 "uuid": "900257f7-5c31-4f25-9ca0-9d817eea607f", 00:08:00.742 "strip_size_kb": 64, 00:08:00.742 "state": "configuring", 00:08:00.742 "raid_level": "concat", 00:08:00.742 "superblock": true, 00:08:00.742 "num_base_bdevs": 3, 00:08:00.742 "num_base_bdevs_discovered": 2, 00:08:00.742 "num_base_bdevs_operational": 3, 00:08:00.742 "base_bdevs_list": [ 00:08:00.742 { 00:08:00.742 "name": "BaseBdev1", 00:08:00.742 "uuid": "ab84e5f1-7127-4ff1-b2cb-a5cf28bead2e", 00:08:00.742 "is_configured": true, 00:08:00.742 "data_offset": 2048, 00:08:00.742 "data_size": 63488 00:08:00.742 }, 00:08:00.742 { 00:08:00.742 "name": null, 00:08:00.742 "uuid": "a77e9cec-136e-4839-b9c7-5bd5054f9619", 00:08:00.742 "is_configured": false, 00:08:00.742 "data_offset": 0, 00:08:00.742 "data_size": 63488 00:08:00.742 }, 00:08:00.742 { 00:08:00.742 "name": "BaseBdev3", 00:08:00.742 "uuid": "1759d92b-10b4-4fd1-86c7-f8de8cc19125", 00:08:00.742 "is_configured": true, 00:08:00.742 "data_offset": 2048, 00:08:00.742 "data_size": 63488 00:08:00.742 } 00:08:00.742 ] 00:08:00.742 }' 00:08:00.742 23:41:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:00.742 23:41:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:01.002 23:41:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:08:01.002 23:41:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:01.002 23:41:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:01.002 23:41:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:01.002 23:41:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:01.002 23:41:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:08:01.002 23:41:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:08:01.002 23:41:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:01.002 23:41:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:01.002 [2024-11-26 23:41:49.045026] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:08:01.002 23:41:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:01.002 23:41:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:01.002 23:41:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:01.002 23:41:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:01.002 23:41:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:01.002 23:41:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:01.002 23:41:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:01.002 23:41:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:01.002 23:41:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:01.002 23:41:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:01.002 23:41:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:01.002 23:41:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:01.002 23:41:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:01.002 23:41:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:01.002 23:41:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:01.002 23:41:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:01.002 23:41:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:01.002 "name": "Existed_Raid", 00:08:01.002 "uuid": "900257f7-5c31-4f25-9ca0-9d817eea607f", 00:08:01.002 "strip_size_kb": 64, 00:08:01.002 "state": "configuring", 00:08:01.002 "raid_level": "concat", 00:08:01.002 "superblock": true, 00:08:01.002 "num_base_bdevs": 3, 00:08:01.002 "num_base_bdevs_discovered": 1, 00:08:01.002 "num_base_bdevs_operational": 3, 00:08:01.002 "base_bdevs_list": [ 00:08:01.002 { 00:08:01.002 "name": "BaseBdev1", 00:08:01.002 "uuid": "ab84e5f1-7127-4ff1-b2cb-a5cf28bead2e", 00:08:01.002 "is_configured": true, 00:08:01.002 "data_offset": 2048, 00:08:01.002 "data_size": 63488 00:08:01.002 }, 00:08:01.002 { 00:08:01.002 "name": null, 00:08:01.002 "uuid": "a77e9cec-136e-4839-b9c7-5bd5054f9619", 00:08:01.002 "is_configured": false, 00:08:01.002 "data_offset": 0, 00:08:01.002 "data_size": 63488 00:08:01.002 }, 00:08:01.002 { 00:08:01.002 "name": null, 00:08:01.002 "uuid": "1759d92b-10b4-4fd1-86c7-f8de8cc19125", 00:08:01.002 "is_configured": false, 00:08:01.002 "data_offset": 0, 00:08:01.002 "data_size": 63488 00:08:01.002 } 00:08:01.002 ] 00:08:01.002 }' 00:08:01.002 23:41:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:01.002 23:41:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:01.573 23:41:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:01.573 23:41:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:08:01.573 23:41:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:01.573 23:41:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:01.573 23:41:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:01.573 23:41:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:08:01.573 23:41:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:08:01.573 23:41:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:01.573 23:41:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:01.573 [2024-11-26 23:41:49.520255] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:01.573 23:41:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:01.573 23:41:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:01.573 23:41:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:01.573 23:41:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:01.573 23:41:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:01.573 23:41:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:01.573 23:41:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:01.574 23:41:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:01.574 23:41:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:01.574 23:41:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:01.574 23:41:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:01.574 23:41:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:01.574 23:41:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:01.574 23:41:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:01.574 23:41:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:01.574 23:41:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:01.574 23:41:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:01.574 "name": "Existed_Raid", 00:08:01.574 "uuid": "900257f7-5c31-4f25-9ca0-9d817eea607f", 00:08:01.574 "strip_size_kb": 64, 00:08:01.574 "state": "configuring", 00:08:01.574 "raid_level": "concat", 00:08:01.574 "superblock": true, 00:08:01.574 "num_base_bdevs": 3, 00:08:01.574 "num_base_bdevs_discovered": 2, 00:08:01.574 "num_base_bdevs_operational": 3, 00:08:01.574 "base_bdevs_list": [ 00:08:01.574 { 00:08:01.574 "name": "BaseBdev1", 00:08:01.574 "uuid": "ab84e5f1-7127-4ff1-b2cb-a5cf28bead2e", 00:08:01.574 "is_configured": true, 00:08:01.574 "data_offset": 2048, 00:08:01.574 "data_size": 63488 00:08:01.574 }, 00:08:01.574 { 00:08:01.574 "name": null, 00:08:01.574 "uuid": "a77e9cec-136e-4839-b9c7-5bd5054f9619", 00:08:01.574 "is_configured": false, 00:08:01.574 "data_offset": 0, 00:08:01.574 "data_size": 63488 00:08:01.574 }, 00:08:01.574 { 00:08:01.574 "name": "BaseBdev3", 00:08:01.574 "uuid": "1759d92b-10b4-4fd1-86c7-f8de8cc19125", 00:08:01.574 "is_configured": true, 00:08:01.574 "data_offset": 2048, 00:08:01.574 "data_size": 63488 00:08:01.574 } 00:08:01.574 ] 00:08:01.574 }' 00:08:01.574 23:41:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:01.574 23:41:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:01.834 23:41:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:01.834 23:41:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:01.834 23:41:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:08:01.834 23:41:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:01.834 23:41:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:02.094 23:41:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:08:02.094 23:41:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:08:02.094 23:41:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:02.094 23:41:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:02.094 [2024-11-26 23:41:49.983487] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:08:02.094 23:41:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:02.094 23:41:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:02.094 23:41:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:02.094 23:41:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:02.094 23:41:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:02.094 23:41:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:02.094 23:41:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:02.094 23:41:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:02.094 23:41:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:02.094 23:41:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:02.094 23:41:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:02.094 23:41:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:02.094 23:41:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:02.094 23:41:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:02.094 23:41:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:02.094 23:41:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:02.094 23:41:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:02.094 "name": "Existed_Raid", 00:08:02.094 "uuid": "900257f7-5c31-4f25-9ca0-9d817eea607f", 00:08:02.094 "strip_size_kb": 64, 00:08:02.094 "state": "configuring", 00:08:02.094 "raid_level": "concat", 00:08:02.094 "superblock": true, 00:08:02.094 "num_base_bdevs": 3, 00:08:02.094 "num_base_bdevs_discovered": 1, 00:08:02.094 "num_base_bdevs_operational": 3, 00:08:02.094 "base_bdevs_list": [ 00:08:02.094 { 00:08:02.094 "name": null, 00:08:02.094 "uuid": "ab84e5f1-7127-4ff1-b2cb-a5cf28bead2e", 00:08:02.094 "is_configured": false, 00:08:02.094 "data_offset": 0, 00:08:02.094 "data_size": 63488 00:08:02.094 }, 00:08:02.094 { 00:08:02.094 "name": null, 00:08:02.094 "uuid": "a77e9cec-136e-4839-b9c7-5bd5054f9619", 00:08:02.094 "is_configured": false, 00:08:02.094 "data_offset": 0, 00:08:02.094 "data_size": 63488 00:08:02.094 }, 00:08:02.094 { 00:08:02.094 "name": "BaseBdev3", 00:08:02.094 "uuid": "1759d92b-10b4-4fd1-86c7-f8de8cc19125", 00:08:02.094 "is_configured": true, 00:08:02.094 "data_offset": 2048, 00:08:02.094 "data_size": 63488 00:08:02.094 } 00:08:02.094 ] 00:08:02.094 }' 00:08:02.094 23:41:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:02.094 23:41:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:02.354 23:41:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:02.354 23:41:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:02.354 23:41:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:02.354 23:41:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:08:02.354 23:41:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:02.354 23:41:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:08:02.354 23:41:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:08:02.354 23:41:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:02.354 23:41:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:02.354 [2024-11-26 23:41:50.481002] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:02.627 23:41:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:02.627 23:41:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:02.627 23:41:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:02.627 23:41:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:02.627 23:41:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:02.627 23:41:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:02.627 23:41:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:02.627 23:41:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:02.627 23:41:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:02.627 23:41:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:02.627 23:41:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:02.627 23:41:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:02.627 23:41:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:02.627 23:41:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:02.627 23:41:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:02.627 23:41:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:02.627 23:41:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:02.627 "name": "Existed_Raid", 00:08:02.627 "uuid": "900257f7-5c31-4f25-9ca0-9d817eea607f", 00:08:02.627 "strip_size_kb": 64, 00:08:02.627 "state": "configuring", 00:08:02.627 "raid_level": "concat", 00:08:02.627 "superblock": true, 00:08:02.627 "num_base_bdevs": 3, 00:08:02.627 "num_base_bdevs_discovered": 2, 00:08:02.627 "num_base_bdevs_operational": 3, 00:08:02.627 "base_bdevs_list": [ 00:08:02.627 { 00:08:02.627 "name": null, 00:08:02.627 "uuid": "ab84e5f1-7127-4ff1-b2cb-a5cf28bead2e", 00:08:02.627 "is_configured": false, 00:08:02.627 "data_offset": 0, 00:08:02.627 "data_size": 63488 00:08:02.627 }, 00:08:02.627 { 00:08:02.627 "name": "BaseBdev2", 00:08:02.627 "uuid": "a77e9cec-136e-4839-b9c7-5bd5054f9619", 00:08:02.627 "is_configured": true, 00:08:02.627 "data_offset": 2048, 00:08:02.627 "data_size": 63488 00:08:02.627 }, 00:08:02.627 { 00:08:02.627 "name": "BaseBdev3", 00:08:02.627 "uuid": "1759d92b-10b4-4fd1-86c7-f8de8cc19125", 00:08:02.627 "is_configured": true, 00:08:02.627 "data_offset": 2048, 00:08:02.627 "data_size": 63488 00:08:02.627 } 00:08:02.627 ] 00:08:02.627 }' 00:08:02.627 23:41:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:02.627 23:41:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:02.892 23:41:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:08:02.892 23:41:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:02.892 23:41:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:02.892 23:41:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:02.892 23:41:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:02.892 23:41:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:08:02.892 23:41:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:02.892 23:41:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:08:02.892 23:41:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:02.892 23:41:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:02.892 23:41:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:02.892 23:41:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u ab84e5f1-7127-4ff1-b2cb-a5cf28bead2e 00:08:02.892 23:41:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:02.892 23:41:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:02.892 [2024-11-26 23:41:50.991001] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:08:02.892 [2024-11-26 23:41:50.991239] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:08:02.892 [2024-11-26 23:41:50.991292] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:08:02.892 [2024-11-26 23:41:50.991570] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002870 00:08:02.892 NewBaseBdev 00:08:02.892 [2024-11-26 23:41:50.991742] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:08:02.892 [2024-11-26 23:41:50.991790] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001c80 00:08:02.892 [2024-11-26 23:41:50.991942] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:02.892 23:41:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:02.892 23:41:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:08:02.892 23:41:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:08:02.892 23:41:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:02.892 23:41:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:08:02.892 23:41:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:02.892 23:41:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:02.892 23:41:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:02.892 23:41:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:02.892 23:41:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:02.892 23:41:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:02.892 23:41:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:08:02.892 23:41:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:02.892 23:41:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:02.892 [ 00:08:02.892 { 00:08:02.892 "name": "NewBaseBdev", 00:08:02.892 "aliases": [ 00:08:02.892 "ab84e5f1-7127-4ff1-b2cb-a5cf28bead2e" 00:08:02.892 ], 00:08:02.892 "product_name": "Malloc disk", 00:08:02.892 "block_size": 512, 00:08:02.892 "num_blocks": 65536, 00:08:02.892 "uuid": "ab84e5f1-7127-4ff1-b2cb-a5cf28bead2e", 00:08:02.892 "assigned_rate_limits": { 00:08:02.892 "rw_ios_per_sec": 0, 00:08:02.892 "rw_mbytes_per_sec": 0, 00:08:02.892 "r_mbytes_per_sec": 0, 00:08:02.892 "w_mbytes_per_sec": 0 00:08:02.892 }, 00:08:02.892 "claimed": true, 00:08:02.892 "claim_type": "exclusive_write", 00:08:02.892 "zoned": false, 00:08:02.892 "supported_io_types": { 00:08:02.892 "read": true, 00:08:02.892 "write": true, 00:08:02.892 "unmap": true, 00:08:02.892 "flush": true, 00:08:02.892 "reset": true, 00:08:02.892 "nvme_admin": false, 00:08:03.153 "nvme_io": false, 00:08:03.153 "nvme_io_md": false, 00:08:03.153 "write_zeroes": true, 00:08:03.153 "zcopy": true, 00:08:03.153 "get_zone_info": false, 00:08:03.153 "zone_management": false, 00:08:03.153 "zone_append": false, 00:08:03.153 "compare": false, 00:08:03.153 "compare_and_write": false, 00:08:03.153 "abort": true, 00:08:03.153 "seek_hole": false, 00:08:03.153 "seek_data": false, 00:08:03.153 "copy": true, 00:08:03.153 "nvme_iov_md": false 00:08:03.153 }, 00:08:03.153 "memory_domains": [ 00:08:03.153 { 00:08:03.153 "dma_device_id": "system", 00:08:03.153 "dma_device_type": 1 00:08:03.153 }, 00:08:03.153 { 00:08:03.153 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:03.153 "dma_device_type": 2 00:08:03.153 } 00:08:03.153 ], 00:08:03.153 "driver_specific": {} 00:08:03.153 } 00:08:03.153 ] 00:08:03.153 23:41:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:03.153 23:41:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:08:03.153 23:41:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online concat 64 3 00:08:03.153 23:41:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:03.153 23:41:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:03.153 23:41:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:03.153 23:41:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:03.153 23:41:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:03.153 23:41:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:03.153 23:41:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:03.153 23:41:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:03.153 23:41:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:03.153 23:41:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:03.153 23:41:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:03.153 23:41:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:03.153 23:41:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:03.153 23:41:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:03.153 23:41:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:03.153 "name": "Existed_Raid", 00:08:03.153 "uuid": "900257f7-5c31-4f25-9ca0-9d817eea607f", 00:08:03.153 "strip_size_kb": 64, 00:08:03.153 "state": "online", 00:08:03.153 "raid_level": "concat", 00:08:03.153 "superblock": true, 00:08:03.153 "num_base_bdevs": 3, 00:08:03.153 "num_base_bdevs_discovered": 3, 00:08:03.153 "num_base_bdevs_operational": 3, 00:08:03.153 "base_bdevs_list": [ 00:08:03.153 { 00:08:03.153 "name": "NewBaseBdev", 00:08:03.153 "uuid": "ab84e5f1-7127-4ff1-b2cb-a5cf28bead2e", 00:08:03.153 "is_configured": true, 00:08:03.153 "data_offset": 2048, 00:08:03.153 "data_size": 63488 00:08:03.153 }, 00:08:03.153 { 00:08:03.153 "name": "BaseBdev2", 00:08:03.153 "uuid": "a77e9cec-136e-4839-b9c7-5bd5054f9619", 00:08:03.153 "is_configured": true, 00:08:03.153 "data_offset": 2048, 00:08:03.153 "data_size": 63488 00:08:03.153 }, 00:08:03.153 { 00:08:03.153 "name": "BaseBdev3", 00:08:03.153 "uuid": "1759d92b-10b4-4fd1-86c7-f8de8cc19125", 00:08:03.153 "is_configured": true, 00:08:03.153 "data_offset": 2048, 00:08:03.153 "data_size": 63488 00:08:03.153 } 00:08:03.153 ] 00:08:03.153 }' 00:08:03.153 23:41:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:03.153 23:41:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:03.414 23:41:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:08:03.414 23:41:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:08:03.414 23:41:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:03.414 23:41:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:03.414 23:41:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:08:03.414 23:41:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:03.414 23:41:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:08:03.414 23:41:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:03.414 23:41:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:03.414 23:41:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:03.414 [2024-11-26 23:41:51.466602] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:03.414 23:41:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:03.414 23:41:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:03.414 "name": "Existed_Raid", 00:08:03.414 "aliases": [ 00:08:03.414 "900257f7-5c31-4f25-9ca0-9d817eea607f" 00:08:03.414 ], 00:08:03.414 "product_name": "Raid Volume", 00:08:03.414 "block_size": 512, 00:08:03.414 "num_blocks": 190464, 00:08:03.414 "uuid": "900257f7-5c31-4f25-9ca0-9d817eea607f", 00:08:03.414 "assigned_rate_limits": { 00:08:03.414 "rw_ios_per_sec": 0, 00:08:03.414 "rw_mbytes_per_sec": 0, 00:08:03.414 "r_mbytes_per_sec": 0, 00:08:03.414 "w_mbytes_per_sec": 0 00:08:03.414 }, 00:08:03.414 "claimed": false, 00:08:03.414 "zoned": false, 00:08:03.414 "supported_io_types": { 00:08:03.414 "read": true, 00:08:03.414 "write": true, 00:08:03.414 "unmap": true, 00:08:03.414 "flush": true, 00:08:03.414 "reset": true, 00:08:03.414 "nvme_admin": false, 00:08:03.414 "nvme_io": false, 00:08:03.414 "nvme_io_md": false, 00:08:03.414 "write_zeroes": true, 00:08:03.414 "zcopy": false, 00:08:03.414 "get_zone_info": false, 00:08:03.414 "zone_management": false, 00:08:03.414 "zone_append": false, 00:08:03.414 "compare": false, 00:08:03.414 "compare_and_write": false, 00:08:03.414 "abort": false, 00:08:03.414 "seek_hole": false, 00:08:03.414 "seek_data": false, 00:08:03.414 "copy": false, 00:08:03.414 "nvme_iov_md": false 00:08:03.414 }, 00:08:03.414 "memory_domains": [ 00:08:03.414 { 00:08:03.414 "dma_device_id": "system", 00:08:03.414 "dma_device_type": 1 00:08:03.414 }, 00:08:03.414 { 00:08:03.414 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:03.414 "dma_device_type": 2 00:08:03.414 }, 00:08:03.414 { 00:08:03.414 "dma_device_id": "system", 00:08:03.414 "dma_device_type": 1 00:08:03.414 }, 00:08:03.414 { 00:08:03.414 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:03.414 "dma_device_type": 2 00:08:03.414 }, 00:08:03.414 { 00:08:03.414 "dma_device_id": "system", 00:08:03.414 "dma_device_type": 1 00:08:03.414 }, 00:08:03.414 { 00:08:03.414 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:03.414 "dma_device_type": 2 00:08:03.414 } 00:08:03.414 ], 00:08:03.414 "driver_specific": { 00:08:03.414 "raid": { 00:08:03.414 "uuid": "900257f7-5c31-4f25-9ca0-9d817eea607f", 00:08:03.414 "strip_size_kb": 64, 00:08:03.414 "state": "online", 00:08:03.414 "raid_level": "concat", 00:08:03.414 "superblock": true, 00:08:03.414 "num_base_bdevs": 3, 00:08:03.414 "num_base_bdevs_discovered": 3, 00:08:03.414 "num_base_bdevs_operational": 3, 00:08:03.414 "base_bdevs_list": [ 00:08:03.414 { 00:08:03.414 "name": "NewBaseBdev", 00:08:03.414 "uuid": "ab84e5f1-7127-4ff1-b2cb-a5cf28bead2e", 00:08:03.414 "is_configured": true, 00:08:03.414 "data_offset": 2048, 00:08:03.414 "data_size": 63488 00:08:03.414 }, 00:08:03.414 { 00:08:03.414 "name": "BaseBdev2", 00:08:03.414 "uuid": "a77e9cec-136e-4839-b9c7-5bd5054f9619", 00:08:03.414 "is_configured": true, 00:08:03.414 "data_offset": 2048, 00:08:03.414 "data_size": 63488 00:08:03.414 }, 00:08:03.414 { 00:08:03.414 "name": "BaseBdev3", 00:08:03.414 "uuid": "1759d92b-10b4-4fd1-86c7-f8de8cc19125", 00:08:03.414 "is_configured": true, 00:08:03.414 "data_offset": 2048, 00:08:03.414 "data_size": 63488 00:08:03.414 } 00:08:03.414 ] 00:08:03.414 } 00:08:03.414 } 00:08:03.414 }' 00:08:03.414 23:41:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:03.414 23:41:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:08:03.414 BaseBdev2 00:08:03.414 BaseBdev3' 00:08:03.414 23:41:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:03.674 23:41:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:03.674 23:41:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:03.674 23:41:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:08:03.674 23:41:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:03.674 23:41:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:03.674 23:41:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:03.674 23:41:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:03.674 23:41:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:03.674 23:41:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:03.674 23:41:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:03.674 23:41:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:08:03.674 23:41:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:03.674 23:41:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:03.674 23:41:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:03.674 23:41:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:03.674 23:41:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:03.674 23:41:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:03.674 23:41:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:03.675 23:41:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:08:03.675 23:41:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:03.675 23:41:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:03.675 23:41:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:03.675 23:41:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:03.675 23:41:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:03.675 23:41:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:03.675 23:41:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:03.675 23:41:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:03.675 23:41:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:03.675 [2024-11-26 23:41:51.721796] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:03.675 [2024-11-26 23:41:51.721821] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:03.675 [2024-11-26 23:41:51.721888] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:03.675 [2024-11-26 23:41:51.721941] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:03.675 [2024-11-26 23:41:51.721952] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name Existed_Raid, state offline 00:08:03.675 23:41:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:03.675 23:41:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 77060 00:08:03.675 23:41:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # '[' -z 77060 ']' 00:08:03.675 23:41:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@958 -- # kill -0 77060 00:08:03.675 23:41:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # uname 00:08:03.675 23:41:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:08:03.675 23:41:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 77060 00:08:03.675 23:41:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:08:03.675 23:41:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:08:03.675 23:41:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@972 -- # echo 'killing process with pid 77060' 00:08:03.675 killing process with pid 77060 00:08:03.675 23:41:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@973 -- # kill 77060 00:08:03.675 [2024-11-26 23:41:51.771628] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:03.675 23:41:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@978 -- # wait 77060 00:08:03.675 [2024-11-26 23:41:51.802378] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:03.935 23:41:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:08:03.935 00:08:03.935 real 0m8.761s 00:08:03.935 user 0m14.974s 00:08:03.935 sys 0m1.740s 00:08:03.935 23:41:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1130 -- # xtrace_disable 00:08:03.935 23:41:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:03.935 ************************************ 00:08:03.935 END TEST raid_state_function_test_sb 00:08:03.935 ************************************ 00:08:04.308 23:41:52 bdev_raid -- bdev/bdev_raid.sh@970 -- # run_test raid_superblock_test raid_superblock_test concat 3 00:08:04.308 23:41:52 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:08:04.308 23:41:52 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:08:04.308 23:41:52 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:04.308 ************************************ 00:08:04.308 START TEST raid_superblock_test 00:08:04.308 ************************************ 00:08:04.308 23:41:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1129 -- # raid_superblock_test concat 3 00:08:04.308 23:41:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=concat 00:08:04.308 23:41:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=3 00:08:04.308 23:41:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:08:04.308 23:41:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:08:04.308 23:41:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:08:04.308 23:41:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:08:04.308 23:41:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:08:04.308 23:41:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:08:04.308 23:41:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:08:04.308 23:41:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:08:04.308 23:41:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:08:04.308 23:41:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:08:04.308 23:41:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:08:04.308 23:41:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' concat '!=' raid1 ']' 00:08:04.308 23:41:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@405 -- # strip_size=64 00:08:04.308 23:41:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@406 -- # strip_size_create_arg='-z 64' 00:08:04.308 23:41:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=77658 00:08:04.308 23:41:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:08:04.308 23:41:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 77658 00:08:04.308 23:41:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@835 -- # '[' -z 77658 ']' 00:08:04.308 23:41:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:04.308 23:41:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:08:04.308 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:04.308 23:41:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:04.308 23:41:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:08:04.308 23:41:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:04.308 [2024-11-26 23:41:52.177247] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:08:04.308 [2024-11-26 23:41:52.177395] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid77658 ] 00:08:04.308 [2024-11-26 23:41:52.331586] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:04.308 [2024-11-26 23:41:52.356997] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:04.566 [2024-11-26 23:41:52.398878] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:04.566 [2024-11-26 23:41:52.398912] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:05.136 23:41:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:08:05.136 23:41:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@868 -- # return 0 00:08:05.136 23:41:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:08:05.136 23:41:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:08:05.136 23:41:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:08:05.136 23:41:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:08:05.136 23:41:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:08:05.136 23:41:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:08:05.136 23:41:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:08:05.136 23:41:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:08:05.136 23:41:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:08:05.136 23:41:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:05.136 23:41:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:05.136 malloc1 00:08:05.136 23:41:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:05.136 23:41:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:08:05.136 23:41:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:05.136 23:41:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:05.136 [2024-11-26 23:41:53.022047] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:08:05.136 [2024-11-26 23:41:53.022183] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:05.136 [2024-11-26 23:41:53.022227] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:08:05.136 [2024-11-26 23:41:53.022265] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:05.136 [2024-11-26 23:41:53.024425] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:05.136 [2024-11-26 23:41:53.024495] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:08:05.136 pt1 00:08:05.136 23:41:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:05.136 23:41:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:08:05.136 23:41:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:08:05.136 23:41:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:08:05.136 23:41:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:08:05.136 23:41:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:08:05.136 23:41:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:08:05.136 23:41:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:08:05.136 23:41:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:08:05.136 23:41:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:08:05.136 23:41:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:05.136 23:41:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:05.136 malloc2 00:08:05.137 23:41:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:05.137 23:41:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:08:05.137 23:41:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:05.137 23:41:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:05.137 [2024-11-26 23:41:53.054516] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:08:05.137 [2024-11-26 23:41:53.054632] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:05.137 [2024-11-26 23:41:53.054670] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:08:05.137 [2024-11-26 23:41:53.054705] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:05.137 [2024-11-26 23:41:53.056770] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:05.137 [2024-11-26 23:41:53.056839] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:08:05.137 pt2 00:08:05.137 23:41:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:05.137 23:41:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:08:05.137 23:41:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:08:05.137 23:41:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc3 00:08:05.137 23:41:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt3 00:08:05.137 23:41:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000003 00:08:05.137 23:41:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:08:05.137 23:41:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:08:05.137 23:41:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:08:05.137 23:41:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc3 00:08:05.137 23:41:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:05.137 23:41:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:05.137 malloc3 00:08:05.137 23:41:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:05.137 23:41:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:08:05.137 23:41:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:05.137 23:41:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:05.137 [2024-11-26 23:41:53.082979] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:08:05.137 [2024-11-26 23:41:53.083076] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:05.137 [2024-11-26 23:41:53.083113] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:08:05.137 [2024-11-26 23:41:53.083150] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:05.137 [2024-11-26 23:41:53.085193] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:05.137 [2024-11-26 23:41:53.085268] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:08:05.137 pt3 00:08:05.137 23:41:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:05.137 23:41:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:08:05.137 23:41:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:08:05.137 23:41:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''pt1 pt2 pt3'\''' -n raid_bdev1 -s 00:08:05.137 23:41:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:05.137 23:41:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:05.137 [2024-11-26 23:41:53.095024] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:08:05.137 [2024-11-26 23:41:53.096931] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:08:05.137 [2024-11-26 23:41:53.097045] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:08:05.137 [2024-11-26 23:41:53.097216] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:08:05.137 [2024-11-26 23:41:53.097264] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:08:05.137 [2024-11-26 23:41:53.097553] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:08:05.137 [2024-11-26 23:41:53.097731] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:08:05.137 [2024-11-26 23:41:53.097775] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:08:05.137 [2024-11-26 23:41:53.097919] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:05.137 23:41:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:05.137 23:41:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online concat 64 3 00:08:05.137 23:41:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:05.137 23:41:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:05.137 23:41:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:05.137 23:41:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:05.137 23:41:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:05.137 23:41:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:05.137 23:41:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:05.137 23:41:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:05.137 23:41:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:05.137 23:41:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:05.137 23:41:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:05.137 23:41:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:05.137 23:41:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:05.137 23:41:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:05.137 23:41:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:05.137 "name": "raid_bdev1", 00:08:05.137 "uuid": "7d64337f-30f2-4376-ad30-19e8c2ab2b94", 00:08:05.137 "strip_size_kb": 64, 00:08:05.137 "state": "online", 00:08:05.137 "raid_level": "concat", 00:08:05.137 "superblock": true, 00:08:05.137 "num_base_bdevs": 3, 00:08:05.137 "num_base_bdevs_discovered": 3, 00:08:05.137 "num_base_bdevs_operational": 3, 00:08:05.137 "base_bdevs_list": [ 00:08:05.137 { 00:08:05.137 "name": "pt1", 00:08:05.137 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:05.137 "is_configured": true, 00:08:05.137 "data_offset": 2048, 00:08:05.137 "data_size": 63488 00:08:05.137 }, 00:08:05.137 { 00:08:05.137 "name": "pt2", 00:08:05.137 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:05.137 "is_configured": true, 00:08:05.137 "data_offset": 2048, 00:08:05.137 "data_size": 63488 00:08:05.137 }, 00:08:05.137 { 00:08:05.137 "name": "pt3", 00:08:05.137 "uuid": "00000000-0000-0000-0000-000000000003", 00:08:05.137 "is_configured": true, 00:08:05.137 "data_offset": 2048, 00:08:05.137 "data_size": 63488 00:08:05.137 } 00:08:05.137 ] 00:08:05.137 }' 00:08:05.137 23:41:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:05.137 23:41:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:05.791 23:41:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:08:05.791 23:41:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:08:05.791 23:41:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:05.791 23:41:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:05.791 23:41:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:08:05.791 23:41:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:05.791 23:41:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:08:05.791 23:41:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:05.791 23:41:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:05.791 23:41:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:05.791 [2024-11-26 23:41:53.550568] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:05.791 23:41:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:05.791 23:41:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:05.791 "name": "raid_bdev1", 00:08:05.791 "aliases": [ 00:08:05.791 "7d64337f-30f2-4376-ad30-19e8c2ab2b94" 00:08:05.791 ], 00:08:05.791 "product_name": "Raid Volume", 00:08:05.791 "block_size": 512, 00:08:05.791 "num_blocks": 190464, 00:08:05.791 "uuid": "7d64337f-30f2-4376-ad30-19e8c2ab2b94", 00:08:05.791 "assigned_rate_limits": { 00:08:05.791 "rw_ios_per_sec": 0, 00:08:05.791 "rw_mbytes_per_sec": 0, 00:08:05.791 "r_mbytes_per_sec": 0, 00:08:05.791 "w_mbytes_per_sec": 0 00:08:05.791 }, 00:08:05.791 "claimed": false, 00:08:05.791 "zoned": false, 00:08:05.791 "supported_io_types": { 00:08:05.791 "read": true, 00:08:05.791 "write": true, 00:08:05.791 "unmap": true, 00:08:05.791 "flush": true, 00:08:05.791 "reset": true, 00:08:05.791 "nvme_admin": false, 00:08:05.791 "nvme_io": false, 00:08:05.791 "nvme_io_md": false, 00:08:05.791 "write_zeroes": true, 00:08:05.791 "zcopy": false, 00:08:05.791 "get_zone_info": false, 00:08:05.791 "zone_management": false, 00:08:05.791 "zone_append": false, 00:08:05.791 "compare": false, 00:08:05.791 "compare_and_write": false, 00:08:05.791 "abort": false, 00:08:05.791 "seek_hole": false, 00:08:05.791 "seek_data": false, 00:08:05.791 "copy": false, 00:08:05.791 "nvme_iov_md": false 00:08:05.791 }, 00:08:05.791 "memory_domains": [ 00:08:05.791 { 00:08:05.791 "dma_device_id": "system", 00:08:05.791 "dma_device_type": 1 00:08:05.791 }, 00:08:05.791 { 00:08:05.791 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:05.791 "dma_device_type": 2 00:08:05.791 }, 00:08:05.791 { 00:08:05.791 "dma_device_id": "system", 00:08:05.791 "dma_device_type": 1 00:08:05.791 }, 00:08:05.791 { 00:08:05.791 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:05.791 "dma_device_type": 2 00:08:05.791 }, 00:08:05.791 { 00:08:05.791 "dma_device_id": "system", 00:08:05.791 "dma_device_type": 1 00:08:05.791 }, 00:08:05.791 { 00:08:05.791 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:05.791 "dma_device_type": 2 00:08:05.791 } 00:08:05.791 ], 00:08:05.791 "driver_specific": { 00:08:05.791 "raid": { 00:08:05.791 "uuid": "7d64337f-30f2-4376-ad30-19e8c2ab2b94", 00:08:05.791 "strip_size_kb": 64, 00:08:05.791 "state": "online", 00:08:05.791 "raid_level": "concat", 00:08:05.791 "superblock": true, 00:08:05.791 "num_base_bdevs": 3, 00:08:05.791 "num_base_bdevs_discovered": 3, 00:08:05.791 "num_base_bdevs_operational": 3, 00:08:05.791 "base_bdevs_list": [ 00:08:05.791 { 00:08:05.791 "name": "pt1", 00:08:05.791 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:05.791 "is_configured": true, 00:08:05.791 "data_offset": 2048, 00:08:05.791 "data_size": 63488 00:08:05.791 }, 00:08:05.791 { 00:08:05.791 "name": "pt2", 00:08:05.791 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:05.791 "is_configured": true, 00:08:05.791 "data_offset": 2048, 00:08:05.791 "data_size": 63488 00:08:05.791 }, 00:08:05.791 { 00:08:05.791 "name": "pt3", 00:08:05.791 "uuid": "00000000-0000-0000-0000-000000000003", 00:08:05.791 "is_configured": true, 00:08:05.791 "data_offset": 2048, 00:08:05.791 "data_size": 63488 00:08:05.791 } 00:08:05.791 ] 00:08:05.791 } 00:08:05.791 } 00:08:05.791 }' 00:08:05.791 23:41:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:05.791 23:41:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:08:05.791 pt2 00:08:05.791 pt3' 00:08:05.791 23:41:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:05.791 23:41:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:05.791 23:41:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:05.791 23:41:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:05.791 23:41:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:08:05.791 23:41:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:05.791 23:41:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:05.791 23:41:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:05.791 23:41:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:05.791 23:41:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:05.791 23:41:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:05.791 23:41:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:08:05.791 23:41:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:05.791 23:41:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:05.791 23:41:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:05.791 23:41:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:05.792 23:41:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:05.792 23:41:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:05.792 23:41:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:05.792 23:41:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:05.792 23:41:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:08:05.792 23:41:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:05.792 23:41:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:05.792 23:41:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:05.792 23:41:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:05.792 23:41:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:05.792 23:41:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:08:05.792 23:41:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:08:05.792 23:41:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:05.792 23:41:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:05.792 [2024-11-26 23:41:53.850005] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:05.792 23:41:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:05.792 23:41:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=7d64337f-30f2-4376-ad30-19e8c2ab2b94 00:08:05.792 23:41:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z 7d64337f-30f2-4376-ad30-19e8c2ab2b94 ']' 00:08:05.792 23:41:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:08:05.792 23:41:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:05.792 23:41:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:05.792 [2024-11-26 23:41:53.881653] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:08:05.792 [2024-11-26 23:41:53.881725] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:05.792 [2024-11-26 23:41:53.881821] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:05.792 [2024-11-26 23:41:53.881889] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:05.792 [2024-11-26 23:41:53.881906] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:08:05.792 23:41:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:05.792 23:41:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:08:05.792 23:41:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:05.792 23:41:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:05.792 23:41:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:05.792 23:41:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:06.052 23:41:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:08:06.052 23:41:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:08:06.052 23:41:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:08:06.052 23:41:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:08:06.052 23:41:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:06.052 23:41:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:06.052 23:41:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:06.052 23:41:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:08:06.052 23:41:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:08:06.052 23:41:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:06.052 23:41:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:06.052 23:41:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:06.052 23:41:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:08:06.052 23:41:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt3 00:08:06.052 23:41:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:06.052 23:41:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:06.052 23:41:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:06.052 23:41:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:08:06.052 23:41:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:08:06.052 23:41:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:06.052 23:41:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:06.052 23:41:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:06.052 23:41:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:08:06.052 23:41:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:08:06.052 23:41:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@652 -- # local es=0 00:08:06.052 23:41:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:08:06.052 23:41:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:08:06.052 23:41:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:08:06.052 23:41:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:08:06.052 23:41:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:08:06.052 23:41:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:08:06.052 23:41:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:06.052 23:41:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:06.052 [2024-11-26 23:41:54.021491] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:08:06.052 [2024-11-26 23:41:54.023351] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:08:06.052 [2024-11-26 23:41:54.023392] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc3 is claimed 00:08:06.052 [2024-11-26 23:41:54.023438] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:08:06.052 [2024-11-26 23:41:54.023481] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:08:06.052 [2024-11-26 23:41:54.023512] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc3 00:08:06.052 [2024-11-26 23:41:54.023524] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:08:06.052 [2024-11-26 23:41:54.023534] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state configuring 00:08:06.052 request: 00:08:06.052 { 00:08:06.052 "name": "raid_bdev1", 00:08:06.052 "raid_level": "concat", 00:08:06.052 "base_bdevs": [ 00:08:06.052 "malloc1", 00:08:06.052 "malloc2", 00:08:06.052 "malloc3" 00:08:06.052 ], 00:08:06.052 "strip_size_kb": 64, 00:08:06.052 "superblock": false, 00:08:06.052 "method": "bdev_raid_create", 00:08:06.052 "req_id": 1 00:08:06.052 } 00:08:06.052 Got JSON-RPC error response 00:08:06.052 response: 00:08:06.052 { 00:08:06.052 "code": -17, 00:08:06.052 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:08:06.052 } 00:08:06.052 23:41:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:08:06.052 23:41:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # es=1 00:08:06.052 23:41:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:08:06.052 23:41:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:08:06.052 23:41:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:08:06.052 23:41:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:06.052 23:41:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:08:06.052 23:41:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:06.052 23:41:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:06.052 23:41:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:06.052 23:41:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:08:06.052 23:41:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:08:06.052 23:41:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:08:06.052 23:41:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:06.052 23:41:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:06.052 [2024-11-26 23:41:54.085336] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:08:06.052 [2024-11-26 23:41:54.085477] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:06.052 [2024-11-26 23:41:54.085530] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:08:06.052 [2024-11-26 23:41:54.085571] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:06.052 [2024-11-26 23:41:54.087988] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:06.052 [2024-11-26 23:41:54.088080] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:08:06.052 [2024-11-26 23:41:54.088189] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:08:06.052 [2024-11-26 23:41:54.088256] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:08:06.052 pt1 00:08:06.052 23:41:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:06.052 23:41:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring concat 64 3 00:08:06.052 23:41:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:06.052 23:41:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:06.052 23:41:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:06.052 23:41:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:06.052 23:41:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:06.052 23:41:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:06.052 23:41:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:06.052 23:41:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:06.052 23:41:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:06.052 23:41:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:06.052 23:41:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:06.052 23:41:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:06.052 23:41:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:06.052 23:41:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:06.052 23:41:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:06.052 "name": "raid_bdev1", 00:08:06.052 "uuid": "7d64337f-30f2-4376-ad30-19e8c2ab2b94", 00:08:06.052 "strip_size_kb": 64, 00:08:06.052 "state": "configuring", 00:08:06.052 "raid_level": "concat", 00:08:06.052 "superblock": true, 00:08:06.052 "num_base_bdevs": 3, 00:08:06.052 "num_base_bdevs_discovered": 1, 00:08:06.052 "num_base_bdevs_operational": 3, 00:08:06.052 "base_bdevs_list": [ 00:08:06.052 { 00:08:06.052 "name": "pt1", 00:08:06.052 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:06.052 "is_configured": true, 00:08:06.052 "data_offset": 2048, 00:08:06.053 "data_size": 63488 00:08:06.053 }, 00:08:06.053 { 00:08:06.053 "name": null, 00:08:06.053 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:06.053 "is_configured": false, 00:08:06.053 "data_offset": 2048, 00:08:06.053 "data_size": 63488 00:08:06.053 }, 00:08:06.053 { 00:08:06.053 "name": null, 00:08:06.053 "uuid": "00000000-0000-0000-0000-000000000003", 00:08:06.053 "is_configured": false, 00:08:06.053 "data_offset": 2048, 00:08:06.053 "data_size": 63488 00:08:06.053 } 00:08:06.053 ] 00:08:06.053 }' 00:08:06.053 23:41:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:06.053 23:41:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:06.633 23:41:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 3 -gt 2 ']' 00:08:06.633 23:41:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@472 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:08:06.633 23:41:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:06.633 23:41:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:06.633 [2024-11-26 23:41:54.532548] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:08:06.633 [2024-11-26 23:41:54.532623] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:06.633 [2024-11-26 23:41:54.532646] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:08:06.633 [2024-11-26 23:41:54.532658] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:06.633 [2024-11-26 23:41:54.533062] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:06.633 [2024-11-26 23:41:54.533088] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:08:06.633 [2024-11-26 23:41:54.533164] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:08:06.633 [2024-11-26 23:41:54.533188] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:08:06.633 pt2 00:08:06.633 23:41:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:06.633 23:41:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@473 -- # rpc_cmd bdev_passthru_delete pt2 00:08:06.633 23:41:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:06.633 23:41:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:06.633 [2024-11-26 23:41:54.544531] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt2 00:08:06.633 23:41:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:06.633 23:41:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@474 -- # verify_raid_bdev_state raid_bdev1 configuring concat 64 3 00:08:06.633 23:41:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:06.633 23:41:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:06.633 23:41:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:06.633 23:41:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:06.633 23:41:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:06.633 23:41:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:06.633 23:41:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:06.633 23:41:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:06.633 23:41:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:06.633 23:41:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:06.633 23:41:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:06.633 23:41:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:06.633 23:41:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:06.633 23:41:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:06.633 23:41:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:06.633 "name": "raid_bdev1", 00:08:06.634 "uuid": "7d64337f-30f2-4376-ad30-19e8c2ab2b94", 00:08:06.634 "strip_size_kb": 64, 00:08:06.634 "state": "configuring", 00:08:06.634 "raid_level": "concat", 00:08:06.634 "superblock": true, 00:08:06.634 "num_base_bdevs": 3, 00:08:06.634 "num_base_bdevs_discovered": 1, 00:08:06.634 "num_base_bdevs_operational": 3, 00:08:06.634 "base_bdevs_list": [ 00:08:06.634 { 00:08:06.634 "name": "pt1", 00:08:06.634 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:06.634 "is_configured": true, 00:08:06.634 "data_offset": 2048, 00:08:06.634 "data_size": 63488 00:08:06.634 }, 00:08:06.634 { 00:08:06.634 "name": null, 00:08:06.634 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:06.634 "is_configured": false, 00:08:06.634 "data_offset": 0, 00:08:06.634 "data_size": 63488 00:08:06.634 }, 00:08:06.634 { 00:08:06.634 "name": null, 00:08:06.634 "uuid": "00000000-0000-0000-0000-000000000003", 00:08:06.634 "is_configured": false, 00:08:06.634 "data_offset": 2048, 00:08:06.634 "data_size": 63488 00:08:06.634 } 00:08:06.634 ] 00:08:06.634 }' 00:08:06.634 23:41:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:06.634 23:41:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:06.896 23:41:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:08:06.896 23:41:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:08:06.896 23:41:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:08:06.896 23:41:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:06.896 23:41:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:06.896 [2024-11-26 23:41:55.019696] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:08:06.896 [2024-11-26 23:41:55.019817] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:06.896 [2024-11-26 23:41:55.019854] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:08:06.896 [2024-11-26 23:41:55.019882] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:06.896 [2024-11-26 23:41:55.020328] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:06.896 [2024-11-26 23:41:55.020394] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:08:06.896 [2024-11-26 23:41:55.020499] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:08:06.896 [2024-11-26 23:41:55.020560] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:08:06.896 pt2 00:08:06.896 23:41:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:06.896 23:41:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:08:06.896 23:41:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:08:06.896 23:41:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:08:07.156 23:41:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:07.156 23:41:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:07.156 [2024-11-26 23:41:55.031652] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:08:07.156 [2024-11-26 23:41:55.031751] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:07.156 [2024-11-26 23:41:55.031785] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:08:07.156 [2024-11-26 23:41:55.031810] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:07.156 [2024-11-26 23:41:55.032147] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:07.156 [2024-11-26 23:41:55.032199] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:08:07.156 [2024-11-26 23:41:55.032278] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:08:07.156 [2024-11-26 23:41:55.032322] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:08:07.156 [2024-11-26 23:41:55.032453] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:08:07.156 [2024-11-26 23:41:55.032494] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:08:07.156 [2024-11-26 23:41:55.032734] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:08:07.156 [2024-11-26 23:41:55.032877] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:08:07.156 [2024-11-26 23:41:55.032919] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:08:07.156 [2024-11-26 23:41:55.033050] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:07.156 pt3 00:08:07.156 23:41:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:07.156 23:41:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:08:07.156 23:41:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:08:07.156 23:41:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online concat 64 3 00:08:07.156 23:41:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:07.156 23:41:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:07.156 23:41:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:07.156 23:41:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:07.156 23:41:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:07.156 23:41:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:07.156 23:41:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:07.156 23:41:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:07.156 23:41:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:07.156 23:41:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:07.156 23:41:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:07.156 23:41:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:07.156 23:41:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:07.156 23:41:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:07.156 23:41:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:07.156 "name": "raid_bdev1", 00:08:07.156 "uuid": "7d64337f-30f2-4376-ad30-19e8c2ab2b94", 00:08:07.156 "strip_size_kb": 64, 00:08:07.156 "state": "online", 00:08:07.156 "raid_level": "concat", 00:08:07.156 "superblock": true, 00:08:07.156 "num_base_bdevs": 3, 00:08:07.156 "num_base_bdevs_discovered": 3, 00:08:07.156 "num_base_bdevs_operational": 3, 00:08:07.156 "base_bdevs_list": [ 00:08:07.156 { 00:08:07.156 "name": "pt1", 00:08:07.156 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:07.156 "is_configured": true, 00:08:07.156 "data_offset": 2048, 00:08:07.156 "data_size": 63488 00:08:07.156 }, 00:08:07.156 { 00:08:07.156 "name": "pt2", 00:08:07.156 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:07.156 "is_configured": true, 00:08:07.156 "data_offset": 2048, 00:08:07.156 "data_size": 63488 00:08:07.156 }, 00:08:07.156 { 00:08:07.156 "name": "pt3", 00:08:07.156 "uuid": "00000000-0000-0000-0000-000000000003", 00:08:07.156 "is_configured": true, 00:08:07.156 "data_offset": 2048, 00:08:07.156 "data_size": 63488 00:08:07.156 } 00:08:07.156 ] 00:08:07.156 }' 00:08:07.156 23:41:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:07.156 23:41:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:07.416 23:41:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:08:07.416 23:41:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:08:07.416 23:41:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:07.416 23:41:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:07.416 23:41:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:08:07.416 23:41:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:07.416 23:41:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:08:07.416 23:41:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:07.416 23:41:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:07.416 23:41:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:07.416 [2024-11-26 23:41:55.455236] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:07.416 23:41:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:07.416 23:41:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:07.416 "name": "raid_bdev1", 00:08:07.416 "aliases": [ 00:08:07.416 "7d64337f-30f2-4376-ad30-19e8c2ab2b94" 00:08:07.416 ], 00:08:07.416 "product_name": "Raid Volume", 00:08:07.416 "block_size": 512, 00:08:07.416 "num_blocks": 190464, 00:08:07.416 "uuid": "7d64337f-30f2-4376-ad30-19e8c2ab2b94", 00:08:07.416 "assigned_rate_limits": { 00:08:07.416 "rw_ios_per_sec": 0, 00:08:07.416 "rw_mbytes_per_sec": 0, 00:08:07.416 "r_mbytes_per_sec": 0, 00:08:07.416 "w_mbytes_per_sec": 0 00:08:07.416 }, 00:08:07.416 "claimed": false, 00:08:07.416 "zoned": false, 00:08:07.416 "supported_io_types": { 00:08:07.416 "read": true, 00:08:07.416 "write": true, 00:08:07.416 "unmap": true, 00:08:07.416 "flush": true, 00:08:07.416 "reset": true, 00:08:07.416 "nvme_admin": false, 00:08:07.416 "nvme_io": false, 00:08:07.416 "nvme_io_md": false, 00:08:07.416 "write_zeroes": true, 00:08:07.416 "zcopy": false, 00:08:07.416 "get_zone_info": false, 00:08:07.416 "zone_management": false, 00:08:07.416 "zone_append": false, 00:08:07.416 "compare": false, 00:08:07.416 "compare_and_write": false, 00:08:07.416 "abort": false, 00:08:07.416 "seek_hole": false, 00:08:07.416 "seek_data": false, 00:08:07.416 "copy": false, 00:08:07.416 "nvme_iov_md": false 00:08:07.416 }, 00:08:07.416 "memory_domains": [ 00:08:07.416 { 00:08:07.416 "dma_device_id": "system", 00:08:07.416 "dma_device_type": 1 00:08:07.416 }, 00:08:07.416 { 00:08:07.416 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:07.416 "dma_device_type": 2 00:08:07.416 }, 00:08:07.416 { 00:08:07.416 "dma_device_id": "system", 00:08:07.416 "dma_device_type": 1 00:08:07.416 }, 00:08:07.416 { 00:08:07.416 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:07.416 "dma_device_type": 2 00:08:07.416 }, 00:08:07.416 { 00:08:07.416 "dma_device_id": "system", 00:08:07.416 "dma_device_type": 1 00:08:07.416 }, 00:08:07.416 { 00:08:07.416 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:07.416 "dma_device_type": 2 00:08:07.416 } 00:08:07.416 ], 00:08:07.416 "driver_specific": { 00:08:07.416 "raid": { 00:08:07.416 "uuid": "7d64337f-30f2-4376-ad30-19e8c2ab2b94", 00:08:07.416 "strip_size_kb": 64, 00:08:07.416 "state": "online", 00:08:07.416 "raid_level": "concat", 00:08:07.416 "superblock": true, 00:08:07.416 "num_base_bdevs": 3, 00:08:07.416 "num_base_bdevs_discovered": 3, 00:08:07.416 "num_base_bdevs_operational": 3, 00:08:07.416 "base_bdevs_list": [ 00:08:07.416 { 00:08:07.416 "name": "pt1", 00:08:07.416 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:07.416 "is_configured": true, 00:08:07.416 "data_offset": 2048, 00:08:07.416 "data_size": 63488 00:08:07.416 }, 00:08:07.416 { 00:08:07.416 "name": "pt2", 00:08:07.416 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:07.416 "is_configured": true, 00:08:07.416 "data_offset": 2048, 00:08:07.416 "data_size": 63488 00:08:07.416 }, 00:08:07.416 { 00:08:07.416 "name": "pt3", 00:08:07.416 "uuid": "00000000-0000-0000-0000-000000000003", 00:08:07.416 "is_configured": true, 00:08:07.416 "data_offset": 2048, 00:08:07.416 "data_size": 63488 00:08:07.416 } 00:08:07.416 ] 00:08:07.416 } 00:08:07.416 } 00:08:07.416 }' 00:08:07.416 23:41:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:07.416 23:41:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:08:07.416 pt2 00:08:07.416 pt3' 00:08:07.677 23:41:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:07.677 23:41:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:07.677 23:41:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:07.677 23:41:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:08:07.677 23:41:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:07.677 23:41:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:07.677 23:41:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:07.677 23:41:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:07.677 23:41:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:07.677 23:41:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:07.677 23:41:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:07.677 23:41:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:08:07.677 23:41:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:07.677 23:41:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:07.677 23:41:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:07.677 23:41:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:07.677 23:41:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:07.677 23:41:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:07.677 23:41:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:07.677 23:41:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:07.677 23:41:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:08:07.677 23:41:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:07.677 23:41:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:07.677 23:41:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:07.677 23:41:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:07.677 23:41:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:07.677 23:41:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:08:07.677 23:41:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:08:07.677 23:41:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:07.677 23:41:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:07.677 [2024-11-26 23:41:55.730710] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:07.677 23:41:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:07.677 23:41:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' 7d64337f-30f2-4376-ad30-19e8c2ab2b94 '!=' 7d64337f-30f2-4376-ad30-19e8c2ab2b94 ']' 00:08:07.677 23:41:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy concat 00:08:07.677 23:41:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:08:07.677 23:41:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # return 1 00:08:07.677 23:41:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 77658 00:08:07.677 23:41:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # '[' -z 77658 ']' 00:08:07.677 23:41:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@958 -- # kill -0 77658 00:08:07.677 23:41:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # uname 00:08:07.677 23:41:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:08:07.677 23:41:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 77658 00:08:07.937 killing process with pid 77658 00:08:07.937 23:41:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:08:07.937 23:41:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:08:07.937 23:41:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 77658' 00:08:07.937 23:41:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@973 -- # kill 77658 00:08:07.937 [2024-11-26 23:41:55.813346] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:07.937 [2024-11-26 23:41:55.813461] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:07.937 [2024-11-26 23:41:55.813528] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:07.937 [2024-11-26 23:41:55.813538] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:08:07.937 23:41:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@978 -- # wait 77658 00:08:07.937 [2024-11-26 23:41:55.847387] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:07.937 23:41:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:08:07.937 00:08:07.937 real 0m3.980s 00:08:07.937 user 0m6.303s 00:08:07.937 sys 0m0.862s 00:08:07.937 23:41:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:08:07.937 ************************************ 00:08:07.937 END TEST raid_superblock_test 00:08:07.937 ************************************ 00:08:07.937 23:41:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:08.197 23:41:56 bdev_raid -- bdev/bdev_raid.sh@971 -- # run_test raid_read_error_test raid_io_error_test concat 3 read 00:08:08.197 23:41:56 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:08:08.197 23:41:56 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:08:08.197 23:41:56 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:08.197 ************************************ 00:08:08.197 START TEST raid_read_error_test 00:08:08.197 ************************************ 00:08:08.197 23:41:56 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test concat 3 read 00:08:08.197 23:41:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=concat 00:08:08.197 23:41:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=3 00:08:08.197 23:41:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=read 00:08:08.197 23:41:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:08:08.197 23:41:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:08.197 23:41:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:08:08.197 23:41:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:08:08.197 23:41:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:08.197 23:41:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:08:08.197 23:41:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:08:08.197 23:41:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:08.197 23:41:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:08:08.197 23:41:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:08:08.197 23:41:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:08.197 23:41:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:08:08.197 23:41:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:08:08.197 23:41:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:08:08.197 23:41:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:08:08.197 23:41:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:08:08.197 23:41:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:08:08.197 23:41:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:08:08.197 23:41:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # '[' concat '!=' raid1 ']' 00:08:08.197 23:41:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:08:08.197 23:41:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:08:08.197 23:41:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:08:08.197 23:41:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.qSWovo8x0k 00:08:08.197 23:41:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=77900 00:08:08.197 23:41:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:08:08.197 23:41:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 77900 00:08:08.197 23:41:56 bdev_raid.raid_read_error_test -- common/autotest_common.sh@835 -- # '[' -z 77900 ']' 00:08:08.197 23:41:56 bdev_raid.raid_read_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:08.197 23:41:56 bdev_raid.raid_read_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:08:08.197 23:41:56 bdev_raid.raid_read_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:08.197 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:08.197 23:41:56 bdev_raid.raid_read_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:08:08.197 23:41:56 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:08.197 [2024-11-26 23:41:56.239422] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:08:08.197 [2024-11-26 23:41:56.239658] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid77900 ] 00:08:08.455 [2024-11-26 23:41:56.394554] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:08.455 [2024-11-26 23:41:56.420078] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:08.455 [2024-11-26 23:41:56.462202] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:08.455 [2024-11-26 23:41:56.462234] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:09.025 23:41:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:08:09.025 23:41:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@868 -- # return 0 00:08:09.025 23:41:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:08:09.025 23:41:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:08:09.025 23:41:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:09.025 23:41:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:09.025 BaseBdev1_malloc 00:08:09.025 23:41:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:09.025 23:41:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:08:09.025 23:41:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:09.025 23:41:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:09.025 true 00:08:09.025 23:41:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:09.025 23:41:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:08:09.025 23:41:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:09.025 23:41:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:09.025 [2024-11-26 23:41:57.097096] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:08:09.025 [2024-11-26 23:41:57.097142] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:09.025 [2024-11-26 23:41:57.097176] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006980 00:08:09.025 [2024-11-26 23:41:57.097184] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:09.025 [2024-11-26 23:41:57.099257] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:09.025 [2024-11-26 23:41:57.099294] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:08:09.025 BaseBdev1 00:08:09.025 23:41:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:09.025 23:41:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:08:09.025 23:41:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:08:09.025 23:41:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:09.025 23:41:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:09.025 BaseBdev2_malloc 00:08:09.025 23:41:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:09.025 23:41:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:08:09.025 23:41:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:09.025 23:41:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:09.025 true 00:08:09.025 23:41:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:09.025 23:41:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:08:09.025 23:41:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:09.025 23:41:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:09.025 [2024-11-26 23:41:57.137431] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:08:09.025 [2024-11-26 23:41:57.137476] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:09.025 [2024-11-26 23:41:57.137509] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:08:09.025 [2024-11-26 23:41:57.137525] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:09.025 [2024-11-26 23:41:57.139542] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:09.025 [2024-11-26 23:41:57.139577] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:08:09.025 BaseBdev2 00:08:09.025 23:41:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:09.025 23:41:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:08:09.025 23:41:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:08:09.025 23:41:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:09.025 23:41:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:09.284 BaseBdev3_malloc 00:08:09.284 23:41:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:09.284 23:41:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:08:09.284 23:41:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:09.284 23:41:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:09.284 true 00:08:09.284 23:41:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:09.284 23:41:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:08:09.284 23:41:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:09.284 23:41:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:09.284 [2024-11-26 23:41:57.177745] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:08:09.284 [2024-11-26 23:41:57.177790] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:09.285 [2024-11-26 23:41:57.177808] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008780 00:08:09.285 [2024-11-26 23:41:57.177816] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:09.285 [2024-11-26 23:41:57.179855] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:09.285 [2024-11-26 23:41:57.179889] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:08:09.285 BaseBdev3 00:08:09.285 23:41:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:09.285 23:41:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n raid_bdev1 -s 00:08:09.285 23:41:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:09.285 23:41:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:09.285 [2024-11-26 23:41:57.189766] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:09.285 [2024-11-26 23:41:57.191679] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:09.285 [2024-11-26 23:41:57.191755] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:09.285 [2024-11-26 23:41:57.191929] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:08:09.285 [2024-11-26 23:41:57.191943] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:08:09.285 [2024-11-26 23:41:57.192215] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002bb0 00:08:09.285 [2024-11-26 23:41:57.192357] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:08:09.285 [2024-11-26 23:41:57.192369] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001c80 00:08:09.285 [2024-11-26 23:41:57.192502] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:09.285 23:41:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:09.285 23:41:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online concat 64 3 00:08:09.285 23:41:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:09.285 23:41:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:09.285 23:41:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:09.285 23:41:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:09.285 23:41:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:09.285 23:41:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:09.285 23:41:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:09.285 23:41:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:09.285 23:41:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:09.285 23:41:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:09.285 23:41:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:09.285 23:41:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:09.285 23:41:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:09.285 23:41:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:09.285 23:41:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:09.285 "name": "raid_bdev1", 00:08:09.285 "uuid": "78678a48-0cb1-4190-a1cb-95f36fe453f7", 00:08:09.285 "strip_size_kb": 64, 00:08:09.285 "state": "online", 00:08:09.285 "raid_level": "concat", 00:08:09.285 "superblock": true, 00:08:09.285 "num_base_bdevs": 3, 00:08:09.285 "num_base_bdevs_discovered": 3, 00:08:09.285 "num_base_bdevs_operational": 3, 00:08:09.285 "base_bdevs_list": [ 00:08:09.285 { 00:08:09.285 "name": "BaseBdev1", 00:08:09.285 "uuid": "4498f45b-ab6f-556b-93a1-6ac4bec49bea", 00:08:09.285 "is_configured": true, 00:08:09.285 "data_offset": 2048, 00:08:09.285 "data_size": 63488 00:08:09.285 }, 00:08:09.285 { 00:08:09.285 "name": "BaseBdev2", 00:08:09.285 "uuid": "006d66a3-7897-5041-a172-14a4c948bd93", 00:08:09.285 "is_configured": true, 00:08:09.285 "data_offset": 2048, 00:08:09.285 "data_size": 63488 00:08:09.285 }, 00:08:09.285 { 00:08:09.285 "name": "BaseBdev3", 00:08:09.285 "uuid": "b6be85b9-0ce6-522c-88d5-fba46da917a8", 00:08:09.285 "is_configured": true, 00:08:09.285 "data_offset": 2048, 00:08:09.285 "data_size": 63488 00:08:09.285 } 00:08:09.285 ] 00:08:09.285 }' 00:08:09.285 23:41:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:09.285 23:41:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:09.545 23:41:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:08:09.545 23:41:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:08:09.803 [2024-11-26 23:41:57.709375] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002d50 00:08:10.738 23:41:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:08:10.738 23:41:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:10.738 23:41:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:10.738 23:41:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:10.738 23:41:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:08:10.738 23:41:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ concat = \r\a\i\d\1 ]] 00:08:10.738 23:41:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=3 00:08:10.738 23:41:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online concat 64 3 00:08:10.738 23:41:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:10.738 23:41:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:10.738 23:41:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:10.738 23:41:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:10.738 23:41:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:10.738 23:41:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:10.738 23:41:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:10.738 23:41:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:10.738 23:41:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:10.738 23:41:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:10.738 23:41:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:10.738 23:41:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:10.739 23:41:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:10.739 23:41:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:10.739 23:41:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:10.739 "name": "raid_bdev1", 00:08:10.739 "uuid": "78678a48-0cb1-4190-a1cb-95f36fe453f7", 00:08:10.739 "strip_size_kb": 64, 00:08:10.739 "state": "online", 00:08:10.739 "raid_level": "concat", 00:08:10.739 "superblock": true, 00:08:10.739 "num_base_bdevs": 3, 00:08:10.739 "num_base_bdevs_discovered": 3, 00:08:10.739 "num_base_bdevs_operational": 3, 00:08:10.739 "base_bdevs_list": [ 00:08:10.739 { 00:08:10.739 "name": "BaseBdev1", 00:08:10.739 "uuid": "4498f45b-ab6f-556b-93a1-6ac4bec49bea", 00:08:10.739 "is_configured": true, 00:08:10.739 "data_offset": 2048, 00:08:10.739 "data_size": 63488 00:08:10.739 }, 00:08:10.739 { 00:08:10.739 "name": "BaseBdev2", 00:08:10.739 "uuid": "006d66a3-7897-5041-a172-14a4c948bd93", 00:08:10.739 "is_configured": true, 00:08:10.739 "data_offset": 2048, 00:08:10.739 "data_size": 63488 00:08:10.739 }, 00:08:10.739 { 00:08:10.739 "name": "BaseBdev3", 00:08:10.739 "uuid": "b6be85b9-0ce6-522c-88d5-fba46da917a8", 00:08:10.739 "is_configured": true, 00:08:10.739 "data_offset": 2048, 00:08:10.739 "data_size": 63488 00:08:10.739 } 00:08:10.739 ] 00:08:10.739 }' 00:08:10.739 23:41:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:10.739 23:41:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:10.998 23:41:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:08:10.998 23:41:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:10.998 23:41:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:10.998 [2024-11-26 23:41:59.077431] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:08:10.998 [2024-11-26 23:41:59.077533] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:10.998 [2024-11-26 23:41:59.080177] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:10.998 [2024-11-26 23:41:59.080293] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:10.998 [2024-11-26 23:41:59.080350] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:10.998 [2024-11-26 23:41:59.080406] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name raid_bdev1, state offline 00:08:10.998 { 00:08:10.998 "results": [ 00:08:10.998 { 00:08:10.998 "job": "raid_bdev1", 00:08:10.998 "core_mask": "0x1", 00:08:10.998 "workload": "randrw", 00:08:10.998 "percentage": 50, 00:08:10.998 "status": "finished", 00:08:10.998 "queue_depth": 1, 00:08:10.998 "io_size": 131072, 00:08:10.998 "runtime": 1.369039, 00:08:10.998 "iops": 16391.05971414985, 00:08:10.998 "mibps": 2048.8824642687314, 00:08:10.998 "io_failed": 1, 00:08:10.998 "io_timeout": 0, 00:08:10.998 "avg_latency_us": 83.99605385417247, 00:08:10.998 "min_latency_us": 25.041048034934498, 00:08:10.998 "max_latency_us": 1566.8541484716156 00:08:10.998 } 00:08:10.998 ], 00:08:10.998 "core_count": 1 00:08:10.998 } 00:08:10.998 23:41:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:10.998 23:41:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 77900 00:08:10.998 23:41:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # '[' -z 77900 ']' 00:08:10.998 23:41:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@958 -- # kill -0 77900 00:08:10.998 23:41:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # uname 00:08:10.998 23:41:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:08:10.998 23:41:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 77900 00:08:10.998 23:41:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:08:10.998 23:41:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:08:10.998 killing process with pid 77900 00:08:10.998 23:41:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 77900' 00:08:10.998 23:41:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@973 -- # kill 77900 00:08:10.998 [2024-11-26 23:41:59.119506] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:10.998 23:41:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@978 -- # wait 77900 00:08:11.257 [2024-11-26 23:41:59.145292] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:11.257 23:41:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.qSWovo8x0k 00:08:11.257 23:41:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:08:11.257 23:41:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:08:11.257 23:41:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.73 00:08:11.257 23:41:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy concat 00:08:11.257 23:41:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:08:11.257 23:41:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:08:11.257 23:41:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.73 != \0\.\0\0 ]] 00:08:11.257 00:08:11.257 real 0m3.225s 00:08:11.257 user 0m4.075s 00:08:11.257 sys 0m0.542s 00:08:11.257 23:41:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:08:11.257 23:41:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:11.257 ************************************ 00:08:11.257 END TEST raid_read_error_test 00:08:11.257 ************************************ 00:08:11.515 23:41:59 bdev_raid -- bdev/bdev_raid.sh@972 -- # run_test raid_write_error_test raid_io_error_test concat 3 write 00:08:11.515 23:41:59 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:08:11.515 23:41:59 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:08:11.515 23:41:59 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:11.515 ************************************ 00:08:11.515 START TEST raid_write_error_test 00:08:11.515 ************************************ 00:08:11.516 23:41:59 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test concat 3 write 00:08:11.516 23:41:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=concat 00:08:11.516 23:41:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=3 00:08:11.516 23:41:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=write 00:08:11.516 23:41:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:08:11.516 23:41:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:11.516 23:41:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:08:11.516 23:41:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:08:11.516 23:41:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:11.516 23:41:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:08:11.516 23:41:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:08:11.516 23:41:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:11.516 23:41:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:08:11.516 23:41:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:08:11.516 23:41:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:11.516 23:41:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:08:11.516 23:41:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:08:11.516 23:41:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:08:11.516 23:41:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:08:11.516 23:41:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:08:11.516 23:41:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:08:11.516 23:41:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:08:11.516 23:41:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # '[' concat '!=' raid1 ']' 00:08:11.516 23:41:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:08:11.516 23:41:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:08:11.516 23:41:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:08:11.516 23:41:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.RK0RE8qILT 00:08:11.516 23:41:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=78029 00:08:11.516 23:41:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 78029 00:08:11.516 23:41:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:08:11.516 23:41:59 bdev_raid.raid_write_error_test -- common/autotest_common.sh@835 -- # '[' -z 78029 ']' 00:08:11.516 23:41:59 bdev_raid.raid_write_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:11.516 23:41:59 bdev_raid.raid_write_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:08:11.516 23:41:59 bdev_raid.raid_write_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:11.516 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:11.516 23:41:59 bdev_raid.raid_write_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:08:11.516 23:41:59 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:11.516 [2024-11-26 23:41:59.530222] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:08:11.516 [2024-11-26 23:41:59.530889] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid78029 ] 00:08:11.775 [2024-11-26 23:41:59.683710] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:11.775 [2024-11-26 23:41:59.709980] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:11.775 [2024-11-26 23:41:59.752680] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:11.775 [2024-11-26 23:41:59.752792] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:12.343 23:42:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:08:12.343 23:42:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@868 -- # return 0 00:08:12.343 23:42:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:08:12.343 23:42:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:08:12.343 23:42:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:12.343 23:42:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:12.343 BaseBdev1_malloc 00:08:12.343 23:42:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:12.343 23:42:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:08:12.343 23:42:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:12.343 23:42:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:12.343 true 00:08:12.343 23:42:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:12.343 23:42:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:08:12.343 23:42:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:12.343 23:42:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:12.343 [2024-11-26 23:42:00.392182] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:08:12.343 [2024-11-26 23:42:00.392290] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:12.343 [2024-11-26 23:42:00.392322] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006980 00:08:12.343 [2024-11-26 23:42:00.392332] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:12.343 [2024-11-26 23:42:00.394522] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:12.343 [2024-11-26 23:42:00.394560] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:08:12.343 BaseBdev1 00:08:12.343 23:42:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:12.343 23:42:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:08:12.343 23:42:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:08:12.343 23:42:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:12.343 23:42:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:12.343 BaseBdev2_malloc 00:08:12.343 23:42:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:12.343 23:42:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:08:12.343 23:42:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:12.343 23:42:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:12.343 true 00:08:12.343 23:42:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:12.343 23:42:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:08:12.343 23:42:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:12.343 23:42:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:12.343 [2024-11-26 23:42:00.432650] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:08:12.343 [2024-11-26 23:42:00.432694] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:12.343 [2024-11-26 23:42:00.432710] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:08:12.343 [2024-11-26 23:42:00.432726] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:12.343 [2024-11-26 23:42:00.434732] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:12.343 [2024-11-26 23:42:00.434827] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:08:12.343 BaseBdev2 00:08:12.343 23:42:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:12.343 23:42:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:08:12.343 23:42:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:08:12.343 23:42:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:12.343 23:42:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:12.343 BaseBdev3_malloc 00:08:12.343 23:42:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:12.343 23:42:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:08:12.343 23:42:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:12.343 23:42:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:12.343 true 00:08:12.343 23:42:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:12.343 23:42:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:08:12.343 23:42:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:12.343 23:42:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:12.603 [2024-11-26 23:42:00.473112] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:08:12.603 [2024-11-26 23:42:00.473156] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:12.603 [2024-11-26 23:42:00.473174] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008780 00:08:12.603 [2024-11-26 23:42:00.473183] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:12.603 [2024-11-26 23:42:00.475255] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:12.603 [2024-11-26 23:42:00.475335] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:08:12.603 BaseBdev3 00:08:12.603 23:42:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:12.603 23:42:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n raid_bdev1 -s 00:08:12.603 23:42:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:12.603 23:42:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:12.603 [2024-11-26 23:42:00.485154] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:12.603 [2024-11-26 23:42:00.486988] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:12.603 [2024-11-26 23:42:00.487063] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:12.603 [2024-11-26 23:42:00.487232] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:08:12.603 [2024-11-26 23:42:00.487246] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:08:12.603 [2024-11-26 23:42:00.487526] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002bb0 00:08:12.603 [2024-11-26 23:42:00.487655] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:08:12.603 [2024-11-26 23:42:00.487687] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001c80 00:08:12.603 [2024-11-26 23:42:00.487796] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:12.603 23:42:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:12.603 23:42:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online concat 64 3 00:08:12.603 23:42:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:12.603 23:42:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:12.603 23:42:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:12.603 23:42:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:12.603 23:42:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:12.603 23:42:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:12.603 23:42:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:12.603 23:42:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:12.603 23:42:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:12.603 23:42:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:12.603 23:42:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:12.603 23:42:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:12.603 23:42:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:12.603 23:42:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:12.603 23:42:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:12.603 "name": "raid_bdev1", 00:08:12.603 "uuid": "87145bb7-850c-47b5-8555-00ca2e946dad", 00:08:12.603 "strip_size_kb": 64, 00:08:12.603 "state": "online", 00:08:12.604 "raid_level": "concat", 00:08:12.604 "superblock": true, 00:08:12.604 "num_base_bdevs": 3, 00:08:12.604 "num_base_bdevs_discovered": 3, 00:08:12.604 "num_base_bdevs_operational": 3, 00:08:12.604 "base_bdevs_list": [ 00:08:12.604 { 00:08:12.604 "name": "BaseBdev1", 00:08:12.604 "uuid": "96584eb9-29e8-5ada-ad84-e669e8a6f2df", 00:08:12.604 "is_configured": true, 00:08:12.604 "data_offset": 2048, 00:08:12.604 "data_size": 63488 00:08:12.604 }, 00:08:12.604 { 00:08:12.604 "name": "BaseBdev2", 00:08:12.604 "uuid": "47b5e90a-e780-5a89-be64-cacbb396530a", 00:08:12.604 "is_configured": true, 00:08:12.604 "data_offset": 2048, 00:08:12.604 "data_size": 63488 00:08:12.604 }, 00:08:12.604 { 00:08:12.604 "name": "BaseBdev3", 00:08:12.604 "uuid": "2ce999b0-daa2-5c0c-a9d5-d346314499cf", 00:08:12.604 "is_configured": true, 00:08:12.604 "data_offset": 2048, 00:08:12.604 "data_size": 63488 00:08:12.604 } 00:08:12.604 ] 00:08:12.604 }' 00:08:12.604 23:42:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:12.604 23:42:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:12.862 23:42:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:08:12.862 23:42:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:08:13.120 [2024-11-26 23:42:01.016675] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002d50 00:08:14.060 23:42:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:08:14.060 23:42:01 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:14.060 23:42:01 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:14.060 23:42:01 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:14.060 23:42:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:08:14.060 23:42:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ concat = \r\a\i\d\1 ]] 00:08:14.060 23:42:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=3 00:08:14.060 23:42:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online concat 64 3 00:08:14.060 23:42:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:14.060 23:42:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:14.060 23:42:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:14.060 23:42:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:14.060 23:42:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:14.060 23:42:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:14.060 23:42:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:14.060 23:42:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:14.060 23:42:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:14.060 23:42:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:14.060 23:42:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:14.060 23:42:01 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:14.060 23:42:01 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:14.060 23:42:01 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:14.060 23:42:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:14.060 "name": "raid_bdev1", 00:08:14.060 "uuid": "87145bb7-850c-47b5-8555-00ca2e946dad", 00:08:14.060 "strip_size_kb": 64, 00:08:14.060 "state": "online", 00:08:14.060 "raid_level": "concat", 00:08:14.060 "superblock": true, 00:08:14.060 "num_base_bdevs": 3, 00:08:14.060 "num_base_bdevs_discovered": 3, 00:08:14.060 "num_base_bdevs_operational": 3, 00:08:14.060 "base_bdevs_list": [ 00:08:14.060 { 00:08:14.060 "name": "BaseBdev1", 00:08:14.060 "uuid": "96584eb9-29e8-5ada-ad84-e669e8a6f2df", 00:08:14.060 "is_configured": true, 00:08:14.060 "data_offset": 2048, 00:08:14.060 "data_size": 63488 00:08:14.060 }, 00:08:14.060 { 00:08:14.060 "name": "BaseBdev2", 00:08:14.060 "uuid": "47b5e90a-e780-5a89-be64-cacbb396530a", 00:08:14.060 "is_configured": true, 00:08:14.060 "data_offset": 2048, 00:08:14.060 "data_size": 63488 00:08:14.060 }, 00:08:14.060 { 00:08:14.060 "name": "BaseBdev3", 00:08:14.060 "uuid": "2ce999b0-daa2-5c0c-a9d5-d346314499cf", 00:08:14.060 "is_configured": true, 00:08:14.060 "data_offset": 2048, 00:08:14.060 "data_size": 63488 00:08:14.060 } 00:08:14.060 ] 00:08:14.060 }' 00:08:14.060 23:42:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:14.060 23:42:01 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:14.320 23:42:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:08:14.320 23:42:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:14.320 23:42:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:14.320 [2024-11-26 23:42:02.404658] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:08:14.320 [2024-11-26 23:42:02.404755] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:14.320 [2024-11-26 23:42:02.407250] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:14.320 [2024-11-26 23:42:02.407358] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:14.320 [2024-11-26 23:42:02.407413] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:14.320 [2024-11-26 23:42:02.407458] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name raid_bdev1, state offline 00:08:14.320 { 00:08:14.320 "results": [ 00:08:14.320 { 00:08:14.320 "job": "raid_bdev1", 00:08:14.320 "core_mask": "0x1", 00:08:14.320 "workload": "randrw", 00:08:14.320 "percentage": 50, 00:08:14.320 "status": "finished", 00:08:14.320 "queue_depth": 1, 00:08:14.320 "io_size": 131072, 00:08:14.320 "runtime": 1.388834, 00:08:14.320 "iops": 16462.730607113594, 00:08:14.320 "mibps": 2057.841325889199, 00:08:14.320 "io_failed": 1, 00:08:14.320 "io_timeout": 0, 00:08:14.320 "avg_latency_us": 83.75615414952202, 00:08:14.320 "min_latency_us": 24.705676855895195, 00:08:14.320 "max_latency_us": 1366.5257641921398 00:08:14.320 } 00:08:14.320 ], 00:08:14.320 "core_count": 1 00:08:14.320 } 00:08:14.320 23:42:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:14.320 23:42:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 78029 00:08:14.320 23:42:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # '[' -z 78029 ']' 00:08:14.320 23:42:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@958 -- # kill -0 78029 00:08:14.320 23:42:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # uname 00:08:14.320 23:42:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:08:14.320 23:42:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 78029 00:08:14.580 killing process with pid 78029 00:08:14.580 23:42:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:08:14.580 23:42:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:08:14.580 23:42:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 78029' 00:08:14.580 23:42:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@973 -- # kill 78029 00:08:14.580 [2024-11-26 23:42:02.451714] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:14.580 23:42:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@978 -- # wait 78029 00:08:14.580 [2024-11-26 23:42:02.477125] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:14.580 23:42:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.RK0RE8qILT 00:08:14.580 23:42:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:08:14.580 23:42:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:08:14.580 23:42:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.72 00:08:14.580 23:42:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy concat 00:08:14.580 ************************************ 00:08:14.580 END TEST raid_write_error_test 00:08:14.580 ************************************ 00:08:14.580 23:42:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:08:14.580 23:42:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:08:14.580 23:42:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.72 != \0\.\0\0 ]] 00:08:14.580 00:08:14.580 real 0m3.261s 00:08:14.580 user 0m4.160s 00:08:14.580 sys 0m0.532s 00:08:14.580 23:42:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:08:14.580 23:42:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:14.848 23:42:02 bdev_raid -- bdev/bdev_raid.sh@967 -- # for level in raid0 concat raid1 00:08:14.848 23:42:02 bdev_raid -- bdev/bdev_raid.sh@968 -- # run_test raid_state_function_test raid_state_function_test raid1 3 false 00:08:14.848 23:42:02 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:08:14.848 23:42:02 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:08:14.848 23:42:02 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:14.848 ************************************ 00:08:14.848 START TEST raid_state_function_test 00:08:14.848 ************************************ 00:08:14.848 23:42:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1129 -- # raid_state_function_test raid1 3 false 00:08:14.848 23:42:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=raid1 00:08:14.848 23:42:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=3 00:08:14.848 23:42:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:08:14.848 23:42:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:08:14.848 23:42:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:08:14.849 23:42:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:14.849 23:42:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:08:14.849 23:42:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:14.849 23:42:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:14.849 23:42:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:08:14.849 23:42:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:14.849 23:42:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:14.849 23:42:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:08:14.849 23:42:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:14.849 23:42:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:14.849 23:42:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:08:14.849 23:42:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:08:14.849 23:42:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:08:14.849 23:42:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:08:14.849 23:42:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:08:14.849 23:42:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:08:14.849 23:42:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' raid1 '!=' raid1 ']' 00:08:14.849 23:42:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@219 -- # strip_size=0 00:08:14.849 23:42:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:08:14.849 23:42:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:08:14.849 23:42:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=78162 00:08:14.849 23:42:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:08:14.849 23:42:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 78162' 00:08:14.850 Process raid pid: 78162 00:08:14.850 23:42:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 78162 00:08:14.850 23:42:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@835 -- # '[' -z 78162 ']' 00:08:14.850 23:42:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:14.850 23:42:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:08:14.850 23:42:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:14.850 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:14.850 23:42:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:08:14.850 23:42:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:14.850 [2024-11-26 23:42:02.854504] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:08:14.850 [2024-11-26 23:42:02.854700] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:08:15.115 [2024-11-26 23:42:02.989719] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:15.115 [2024-11-26 23:42:03.018157] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:15.115 [2024-11-26 23:42:03.060438] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:15.115 [2024-11-26 23:42:03.060549] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:15.684 23:42:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:08:15.684 23:42:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@868 -- # return 0 00:08:15.684 23:42:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:15.684 23:42:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:15.684 23:42:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:15.684 [2024-11-26 23:42:03.722825] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:15.684 [2024-11-26 23:42:03.722881] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:15.684 [2024-11-26 23:42:03.722892] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:15.685 [2024-11-26 23:42:03.722901] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:15.685 [2024-11-26 23:42:03.722907] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:08:15.685 [2024-11-26 23:42:03.722919] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:08:15.685 23:42:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:15.685 23:42:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:08:15.685 23:42:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:15.685 23:42:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:15.685 23:42:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:15.685 23:42:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:15.685 23:42:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:15.685 23:42:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:15.685 23:42:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:15.685 23:42:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:15.685 23:42:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:15.685 23:42:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:15.685 23:42:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:15.685 23:42:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:15.685 23:42:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:15.685 23:42:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:15.685 23:42:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:15.685 "name": "Existed_Raid", 00:08:15.685 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:15.685 "strip_size_kb": 0, 00:08:15.685 "state": "configuring", 00:08:15.685 "raid_level": "raid1", 00:08:15.685 "superblock": false, 00:08:15.685 "num_base_bdevs": 3, 00:08:15.685 "num_base_bdevs_discovered": 0, 00:08:15.685 "num_base_bdevs_operational": 3, 00:08:15.685 "base_bdevs_list": [ 00:08:15.685 { 00:08:15.685 "name": "BaseBdev1", 00:08:15.685 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:15.685 "is_configured": false, 00:08:15.685 "data_offset": 0, 00:08:15.685 "data_size": 0 00:08:15.685 }, 00:08:15.685 { 00:08:15.685 "name": "BaseBdev2", 00:08:15.685 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:15.685 "is_configured": false, 00:08:15.685 "data_offset": 0, 00:08:15.685 "data_size": 0 00:08:15.685 }, 00:08:15.685 { 00:08:15.685 "name": "BaseBdev3", 00:08:15.685 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:15.685 "is_configured": false, 00:08:15.685 "data_offset": 0, 00:08:15.685 "data_size": 0 00:08:15.685 } 00:08:15.685 ] 00:08:15.685 }' 00:08:15.685 23:42:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:15.685 23:42:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:16.256 23:42:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:16.256 23:42:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:16.256 23:42:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:16.256 [2024-11-26 23:42:04.193978] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:16.256 [2024-11-26 23:42:04.194067] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:08:16.256 23:42:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:16.256 23:42:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:16.256 23:42:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:16.256 23:42:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:16.256 [2024-11-26 23:42:04.205977] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:16.256 [2024-11-26 23:42:04.206059] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:16.256 [2024-11-26 23:42:04.206112] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:16.256 [2024-11-26 23:42:04.206137] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:16.256 [2024-11-26 23:42:04.206165] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:08:16.256 [2024-11-26 23:42:04.206209] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:08:16.256 23:42:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:16.256 23:42:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:08:16.256 23:42:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:16.256 23:42:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:16.256 [2024-11-26 23:42:04.226690] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:16.256 BaseBdev1 00:08:16.256 23:42:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:16.256 23:42:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:08:16.256 23:42:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:08:16.256 23:42:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:16.256 23:42:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:08:16.256 23:42:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:16.256 23:42:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:16.256 23:42:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:16.256 23:42:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:16.256 23:42:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:16.256 23:42:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:16.256 23:42:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:08:16.256 23:42:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:16.256 23:42:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:16.256 [ 00:08:16.256 { 00:08:16.256 "name": "BaseBdev1", 00:08:16.256 "aliases": [ 00:08:16.256 "32def0d6-7f2d-4f5b-b3b0-089380ac7c3f" 00:08:16.256 ], 00:08:16.256 "product_name": "Malloc disk", 00:08:16.256 "block_size": 512, 00:08:16.256 "num_blocks": 65536, 00:08:16.256 "uuid": "32def0d6-7f2d-4f5b-b3b0-089380ac7c3f", 00:08:16.256 "assigned_rate_limits": { 00:08:16.256 "rw_ios_per_sec": 0, 00:08:16.256 "rw_mbytes_per_sec": 0, 00:08:16.256 "r_mbytes_per_sec": 0, 00:08:16.256 "w_mbytes_per_sec": 0 00:08:16.256 }, 00:08:16.256 "claimed": true, 00:08:16.256 "claim_type": "exclusive_write", 00:08:16.256 "zoned": false, 00:08:16.256 "supported_io_types": { 00:08:16.256 "read": true, 00:08:16.256 "write": true, 00:08:16.256 "unmap": true, 00:08:16.256 "flush": true, 00:08:16.256 "reset": true, 00:08:16.256 "nvme_admin": false, 00:08:16.256 "nvme_io": false, 00:08:16.256 "nvme_io_md": false, 00:08:16.256 "write_zeroes": true, 00:08:16.256 "zcopy": true, 00:08:16.256 "get_zone_info": false, 00:08:16.256 "zone_management": false, 00:08:16.256 "zone_append": false, 00:08:16.256 "compare": false, 00:08:16.256 "compare_and_write": false, 00:08:16.256 "abort": true, 00:08:16.256 "seek_hole": false, 00:08:16.256 "seek_data": false, 00:08:16.256 "copy": true, 00:08:16.256 "nvme_iov_md": false 00:08:16.256 }, 00:08:16.256 "memory_domains": [ 00:08:16.256 { 00:08:16.256 "dma_device_id": "system", 00:08:16.256 "dma_device_type": 1 00:08:16.256 }, 00:08:16.256 { 00:08:16.256 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:16.256 "dma_device_type": 2 00:08:16.256 } 00:08:16.256 ], 00:08:16.256 "driver_specific": {} 00:08:16.256 } 00:08:16.256 ] 00:08:16.256 23:42:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:16.256 23:42:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:08:16.256 23:42:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:08:16.256 23:42:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:16.256 23:42:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:16.256 23:42:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:16.256 23:42:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:16.256 23:42:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:16.256 23:42:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:16.256 23:42:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:16.256 23:42:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:16.256 23:42:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:16.256 23:42:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:16.256 23:42:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:16.256 23:42:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:16.256 23:42:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:16.256 23:42:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:16.256 23:42:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:16.256 "name": "Existed_Raid", 00:08:16.256 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:16.256 "strip_size_kb": 0, 00:08:16.256 "state": "configuring", 00:08:16.256 "raid_level": "raid1", 00:08:16.256 "superblock": false, 00:08:16.256 "num_base_bdevs": 3, 00:08:16.256 "num_base_bdevs_discovered": 1, 00:08:16.256 "num_base_bdevs_operational": 3, 00:08:16.256 "base_bdevs_list": [ 00:08:16.256 { 00:08:16.256 "name": "BaseBdev1", 00:08:16.256 "uuid": "32def0d6-7f2d-4f5b-b3b0-089380ac7c3f", 00:08:16.256 "is_configured": true, 00:08:16.256 "data_offset": 0, 00:08:16.256 "data_size": 65536 00:08:16.256 }, 00:08:16.256 { 00:08:16.256 "name": "BaseBdev2", 00:08:16.256 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:16.256 "is_configured": false, 00:08:16.256 "data_offset": 0, 00:08:16.256 "data_size": 0 00:08:16.256 }, 00:08:16.256 { 00:08:16.256 "name": "BaseBdev3", 00:08:16.256 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:16.256 "is_configured": false, 00:08:16.256 "data_offset": 0, 00:08:16.256 "data_size": 0 00:08:16.256 } 00:08:16.256 ] 00:08:16.256 }' 00:08:16.257 23:42:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:16.257 23:42:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:16.827 23:42:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:16.827 23:42:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:16.827 23:42:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:16.827 [2024-11-26 23:42:04.701999] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:16.827 [2024-11-26 23:42:04.702056] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:08:16.827 23:42:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:16.827 23:42:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:16.827 23:42:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:16.827 23:42:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:16.827 [2024-11-26 23:42:04.714005] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:16.827 [2024-11-26 23:42:04.715936] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:16.827 [2024-11-26 23:42:04.716014] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:16.827 [2024-11-26 23:42:04.716027] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:08:16.827 [2024-11-26 23:42:04.716038] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:08:16.827 23:42:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:16.827 23:42:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:08:16.827 23:42:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:16.827 23:42:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:08:16.827 23:42:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:16.827 23:42:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:16.827 23:42:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:16.827 23:42:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:16.827 23:42:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:16.827 23:42:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:16.827 23:42:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:16.827 23:42:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:16.827 23:42:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:16.827 23:42:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:16.827 23:42:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:16.827 23:42:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:16.827 23:42:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:16.827 23:42:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:16.827 23:42:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:16.827 "name": "Existed_Raid", 00:08:16.827 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:16.827 "strip_size_kb": 0, 00:08:16.827 "state": "configuring", 00:08:16.827 "raid_level": "raid1", 00:08:16.827 "superblock": false, 00:08:16.827 "num_base_bdevs": 3, 00:08:16.828 "num_base_bdevs_discovered": 1, 00:08:16.828 "num_base_bdevs_operational": 3, 00:08:16.828 "base_bdevs_list": [ 00:08:16.828 { 00:08:16.828 "name": "BaseBdev1", 00:08:16.828 "uuid": "32def0d6-7f2d-4f5b-b3b0-089380ac7c3f", 00:08:16.828 "is_configured": true, 00:08:16.828 "data_offset": 0, 00:08:16.828 "data_size": 65536 00:08:16.828 }, 00:08:16.828 { 00:08:16.828 "name": "BaseBdev2", 00:08:16.828 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:16.828 "is_configured": false, 00:08:16.828 "data_offset": 0, 00:08:16.828 "data_size": 0 00:08:16.828 }, 00:08:16.828 { 00:08:16.828 "name": "BaseBdev3", 00:08:16.828 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:16.828 "is_configured": false, 00:08:16.828 "data_offset": 0, 00:08:16.828 "data_size": 0 00:08:16.828 } 00:08:16.828 ] 00:08:16.828 }' 00:08:16.828 23:42:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:16.828 23:42:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:17.087 23:42:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:08:17.088 23:42:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:17.088 23:42:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:17.088 [2024-11-26 23:42:05.172388] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:17.088 BaseBdev2 00:08:17.088 23:42:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:17.088 23:42:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:08:17.088 23:42:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:08:17.088 23:42:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:17.088 23:42:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:08:17.088 23:42:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:17.088 23:42:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:17.088 23:42:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:17.088 23:42:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:17.088 23:42:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:17.088 23:42:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:17.088 23:42:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:08:17.088 23:42:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:17.088 23:42:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:17.088 [ 00:08:17.088 { 00:08:17.088 "name": "BaseBdev2", 00:08:17.088 "aliases": [ 00:08:17.088 "ed589290-e998-4956-8cab-f10b0163b609" 00:08:17.088 ], 00:08:17.088 "product_name": "Malloc disk", 00:08:17.088 "block_size": 512, 00:08:17.088 "num_blocks": 65536, 00:08:17.088 "uuid": "ed589290-e998-4956-8cab-f10b0163b609", 00:08:17.088 "assigned_rate_limits": { 00:08:17.088 "rw_ios_per_sec": 0, 00:08:17.088 "rw_mbytes_per_sec": 0, 00:08:17.088 "r_mbytes_per_sec": 0, 00:08:17.088 "w_mbytes_per_sec": 0 00:08:17.088 }, 00:08:17.088 "claimed": true, 00:08:17.088 "claim_type": "exclusive_write", 00:08:17.088 "zoned": false, 00:08:17.088 "supported_io_types": { 00:08:17.088 "read": true, 00:08:17.088 "write": true, 00:08:17.088 "unmap": true, 00:08:17.088 "flush": true, 00:08:17.088 "reset": true, 00:08:17.088 "nvme_admin": false, 00:08:17.088 "nvme_io": false, 00:08:17.088 "nvme_io_md": false, 00:08:17.088 "write_zeroes": true, 00:08:17.088 "zcopy": true, 00:08:17.088 "get_zone_info": false, 00:08:17.088 "zone_management": false, 00:08:17.088 "zone_append": false, 00:08:17.088 "compare": false, 00:08:17.088 "compare_and_write": false, 00:08:17.088 "abort": true, 00:08:17.088 "seek_hole": false, 00:08:17.088 "seek_data": false, 00:08:17.088 "copy": true, 00:08:17.088 "nvme_iov_md": false 00:08:17.088 }, 00:08:17.088 "memory_domains": [ 00:08:17.088 { 00:08:17.088 "dma_device_id": "system", 00:08:17.088 "dma_device_type": 1 00:08:17.088 }, 00:08:17.088 { 00:08:17.088 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:17.088 "dma_device_type": 2 00:08:17.088 } 00:08:17.088 ], 00:08:17.088 "driver_specific": {} 00:08:17.088 } 00:08:17.088 ] 00:08:17.088 23:42:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:17.088 23:42:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:08:17.088 23:42:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:08:17.088 23:42:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:17.088 23:42:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:08:17.088 23:42:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:17.088 23:42:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:17.088 23:42:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:17.088 23:42:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:17.088 23:42:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:17.088 23:42:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:17.088 23:42:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:17.088 23:42:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:17.088 23:42:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:17.347 23:42:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:17.347 23:42:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:17.347 23:42:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:17.347 23:42:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:17.347 23:42:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:17.347 23:42:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:17.347 "name": "Existed_Raid", 00:08:17.347 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:17.347 "strip_size_kb": 0, 00:08:17.347 "state": "configuring", 00:08:17.347 "raid_level": "raid1", 00:08:17.347 "superblock": false, 00:08:17.347 "num_base_bdevs": 3, 00:08:17.347 "num_base_bdevs_discovered": 2, 00:08:17.347 "num_base_bdevs_operational": 3, 00:08:17.347 "base_bdevs_list": [ 00:08:17.347 { 00:08:17.347 "name": "BaseBdev1", 00:08:17.347 "uuid": "32def0d6-7f2d-4f5b-b3b0-089380ac7c3f", 00:08:17.347 "is_configured": true, 00:08:17.347 "data_offset": 0, 00:08:17.347 "data_size": 65536 00:08:17.347 }, 00:08:17.347 { 00:08:17.347 "name": "BaseBdev2", 00:08:17.347 "uuid": "ed589290-e998-4956-8cab-f10b0163b609", 00:08:17.347 "is_configured": true, 00:08:17.347 "data_offset": 0, 00:08:17.347 "data_size": 65536 00:08:17.347 }, 00:08:17.347 { 00:08:17.347 "name": "BaseBdev3", 00:08:17.347 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:17.347 "is_configured": false, 00:08:17.347 "data_offset": 0, 00:08:17.347 "data_size": 0 00:08:17.347 } 00:08:17.347 ] 00:08:17.347 }' 00:08:17.347 23:42:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:17.347 23:42:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:17.607 23:42:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:08:17.607 23:42:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:17.607 23:42:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:17.607 [2024-11-26 23:42:05.672495] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:17.607 [2024-11-26 23:42:05.672610] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:08:17.607 [2024-11-26 23:42:05.672640] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:08:17.607 [2024-11-26 23:42:05.672982] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:08:17.607 [2024-11-26 23:42:05.673140] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:08:17.607 [2024-11-26 23:42:05.673152] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:08:17.607 [2024-11-26 23:42:05.673384] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:17.607 BaseBdev3 00:08:17.607 23:42:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:17.607 23:42:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:08:17.607 23:42:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:08:17.607 23:42:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:17.607 23:42:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:08:17.607 23:42:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:17.607 23:42:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:17.607 23:42:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:17.607 23:42:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:17.607 23:42:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:17.607 23:42:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:17.607 23:42:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:08:17.607 23:42:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:17.607 23:42:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:17.607 [ 00:08:17.607 { 00:08:17.607 "name": "BaseBdev3", 00:08:17.607 "aliases": [ 00:08:17.607 "74381ae0-e7ea-4688-82dc-aa1f3191b179" 00:08:17.607 ], 00:08:17.607 "product_name": "Malloc disk", 00:08:17.607 "block_size": 512, 00:08:17.607 "num_blocks": 65536, 00:08:17.607 "uuid": "74381ae0-e7ea-4688-82dc-aa1f3191b179", 00:08:17.607 "assigned_rate_limits": { 00:08:17.607 "rw_ios_per_sec": 0, 00:08:17.607 "rw_mbytes_per_sec": 0, 00:08:17.607 "r_mbytes_per_sec": 0, 00:08:17.607 "w_mbytes_per_sec": 0 00:08:17.607 }, 00:08:17.607 "claimed": true, 00:08:17.607 "claim_type": "exclusive_write", 00:08:17.608 "zoned": false, 00:08:17.608 "supported_io_types": { 00:08:17.608 "read": true, 00:08:17.608 "write": true, 00:08:17.608 "unmap": true, 00:08:17.608 "flush": true, 00:08:17.608 "reset": true, 00:08:17.608 "nvme_admin": false, 00:08:17.608 "nvme_io": false, 00:08:17.608 "nvme_io_md": false, 00:08:17.608 "write_zeroes": true, 00:08:17.608 "zcopy": true, 00:08:17.608 "get_zone_info": false, 00:08:17.608 "zone_management": false, 00:08:17.608 "zone_append": false, 00:08:17.608 "compare": false, 00:08:17.608 "compare_and_write": false, 00:08:17.608 "abort": true, 00:08:17.608 "seek_hole": false, 00:08:17.608 "seek_data": false, 00:08:17.608 "copy": true, 00:08:17.608 "nvme_iov_md": false 00:08:17.608 }, 00:08:17.608 "memory_domains": [ 00:08:17.608 { 00:08:17.608 "dma_device_id": "system", 00:08:17.608 "dma_device_type": 1 00:08:17.608 }, 00:08:17.608 { 00:08:17.608 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:17.608 "dma_device_type": 2 00:08:17.608 } 00:08:17.608 ], 00:08:17.608 "driver_specific": {} 00:08:17.608 } 00:08:17.608 ] 00:08:17.608 23:42:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:17.608 23:42:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:08:17.608 23:42:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:08:17.608 23:42:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:17.608 23:42:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid1 0 3 00:08:17.608 23:42:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:17.608 23:42:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:17.608 23:42:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:17.608 23:42:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:17.608 23:42:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:17.608 23:42:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:17.608 23:42:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:17.608 23:42:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:17.608 23:42:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:17.608 23:42:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:17.608 23:42:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:17.608 23:42:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:17.608 23:42:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:17.868 23:42:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:17.868 23:42:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:17.868 "name": "Existed_Raid", 00:08:17.868 "uuid": "b03582e4-a081-4e8e-90e4-be690087a2af", 00:08:17.868 "strip_size_kb": 0, 00:08:17.868 "state": "online", 00:08:17.868 "raid_level": "raid1", 00:08:17.868 "superblock": false, 00:08:17.868 "num_base_bdevs": 3, 00:08:17.868 "num_base_bdevs_discovered": 3, 00:08:17.868 "num_base_bdevs_operational": 3, 00:08:17.868 "base_bdevs_list": [ 00:08:17.868 { 00:08:17.868 "name": "BaseBdev1", 00:08:17.868 "uuid": "32def0d6-7f2d-4f5b-b3b0-089380ac7c3f", 00:08:17.868 "is_configured": true, 00:08:17.868 "data_offset": 0, 00:08:17.868 "data_size": 65536 00:08:17.868 }, 00:08:17.868 { 00:08:17.868 "name": "BaseBdev2", 00:08:17.868 "uuid": "ed589290-e998-4956-8cab-f10b0163b609", 00:08:17.868 "is_configured": true, 00:08:17.868 "data_offset": 0, 00:08:17.868 "data_size": 65536 00:08:17.868 }, 00:08:17.868 { 00:08:17.868 "name": "BaseBdev3", 00:08:17.868 "uuid": "74381ae0-e7ea-4688-82dc-aa1f3191b179", 00:08:17.868 "is_configured": true, 00:08:17.868 "data_offset": 0, 00:08:17.868 "data_size": 65536 00:08:17.868 } 00:08:17.868 ] 00:08:17.868 }' 00:08:17.868 23:42:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:17.868 23:42:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:18.128 23:42:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:08:18.128 23:42:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:08:18.128 23:42:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:18.128 23:42:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:18.128 23:42:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:08:18.128 23:42:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:18.128 23:42:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:08:18.128 23:42:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:18.128 23:42:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:18.128 23:42:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:18.128 [2024-11-26 23:42:06.188050] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:18.128 23:42:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:18.128 23:42:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:18.128 "name": "Existed_Raid", 00:08:18.128 "aliases": [ 00:08:18.128 "b03582e4-a081-4e8e-90e4-be690087a2af" 00:08:18.128 ], 00:08:18.128 "product_name": "Raid Volume", 00:08:18.128 "block_size": 512, 00:08:18.128 "num_blocks": 65536, 00:08:18.128 "uuid": "b03582e4-a081-4e8e-90e4-be690087a2af", 00:08:18.128 "assigned_rate_limits": { 00:08:18.128 "rw_ios_per_sec": 0, 00:08:18.128 "rw_mbytes_per_sec": 0, 00:08:18.128 "r_mbytes_per_sec": 0, 00:08:18.128 "w_mbytes_per_sec": 0 00:08:18.128 }, 00:08:18.128 "claimed": false, 00:08:18.128 "zoned": false, 00:08:18.128 "supported_io_types": { 00:08:18.128 "read": true, 00:08:18.128 "write": true, 00:08:18.128 "unmap": false, 00:08:18.128 "flush": false, 00:08:18.128 "reset": true, 00:08:18.128 "nvme_admin": false, 00:08:18.128 "nvme_io": false, 00:08:18.128 "nvme_io_md": false, 00:08:18.128 "write_zeroes": true, 00:08:18.128 "zcopy": false, 00:08:18.128 "get_zone_info": false, 00:08:18.128 "zone_management": false, 00:08:18.128 "zone_append": false, 00:08:18.128 "compare": false, 00:08:18.128 "compare_and_write": false, 00:08:18.128 "abort": false, 00:08:18.128 "seek_hole": false, 00:08:18.128 "seek_data": false, 00:08:18.128 "copy": false, 00:08:18.128 "nvme_iov_md": false 00:08:18.128 }, 00:08:18.128 "memory_domains": [ 00:08:18.128 { 00:08:18.128 "dma_device_id": "system", 00:08:18.128 "dma_device_type": 1 00:08:18.128 }, 00:08:18.128 { 00:08:18.128 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:18.128 "dma_device_type": 2 00:08:18.128 }, 00:08:18.128 { 00:08:18.128 "dma_device_id": "system", 00:08:18.128 "dma_device_type": 1 00:08:18.128 }, 00:08:18.128 { 00:08:18.128 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:18.128 "dma_device_type": 2 00:08:18.128 }, 00:08:18.128 { 00:08:18.128 "dma_device_id": "system", 00:08:18.128 "dma_device_type": 1 00:08:18.128 }, 00:08:18.128 { 00:08:18.128 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:18.128 "dma_device_type": 2 00:08:18.128 } 00:08:18.128 ], 00:08:18.128 "driver_specific": { 00:08:18.128 "raid": { 00:08:18.128 "uuid": "b03582e4-a081-4e8e-90e4-be690087a2af", 00:08:18.128 "strip_size_kb": 0, 00:08:18.128 "state": "online", 00:08:18.128 "raid_level": "raid1", 00:08:18.128 "superblock": false, 00:08:18.128 "num_base_bdevs": 3, 00:08:18.128 "num_base_bdevs_discovered": 3, 00:08:18.128 "num_base_bdevs_operational": 3, 00:08:18.128 "base_bdevs_list": [ 00:08:18.128 { 00:08:18.128 "name": "BaseBdev1", 00:08:18.128 "uuid": "32def0d6-7f2d-4f5b-b3b0-089380ac7c3f", 00:08:18.128 "is_configured": true, 00:08:18.128 "data_offset": 0, 00:08:18.128 "data_size": 65536 00:08:18.128 }, 00:08:18.128 { 00:08:18.128 "name": "BaseBdev2", 00:08:18.128 "uuid": "ed589290-e998-4956-8cab-f10b0163b609", 00:08:18.128 "is_configured": true, 00:08:18.128 "data_offset": 0, 00:08:18.128 "data_size": 65536 00:08:18.128 }, 00:08:18.128 { 00:08:18.128 "name": "BaseBdev3", 00:08:18.128 "uuid": "74381ae0-e7ea-4688-82dc-aa1f3191b179", 00:08:18.128 "is_configured": true, 00:08:18.128 "data_offset": 0, 00:08:18.128 "data_size": 65536 00:08:18.128 } 00:08:18.128 ] 00:08:18.128 } 00:08:18.128 } 00:08:18.128 }' 00:08:18.128 23:42:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:18.388 23:42:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:08:18.388 BaseBdev2 00:08:18.388 BaseBdev3' 00:08:18.388 23:42:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:18.388 23:42:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:18.388 23:42:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:18.388 23:42:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:08:18.388 23:42:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:18.388 23:42:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:18.388 23:42:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:18.388 23:42:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:18.388 23:42:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:18.388 23:42:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:18.388 23:42:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:18.388 23:42:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:08:18.388 23:42:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:18.389 23:42:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:18.389 23:42:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:18.389 23:42:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:18.389 23:42:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:18.389 23:42:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:18.389 23:42:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:18.389 23:42:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:08:18.389 23:42:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:18.389 23:42:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:18.389 23:42:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:18.389 23:42:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:18.389 23:42:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:18.389 23:42:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:18.389 23:42:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:08:18.389 23:42:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:18.389 23:42:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:18.389 [2024-11-26 23:42:06.443289] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:08:18.389 23:42:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:18.389 23:42:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:08:18.389 23:42:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy raid1 00:08:18.389 23:42:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:08:18.389 23:42:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@199 -- # return 0 00:08:18.389 23:42:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:08:18.389 23:42:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid1 0 2 00:08:18.389 23:42:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:18.389 23:42:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:18.389 23:42:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:18.389 23:42:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:18.389 23:42:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:18.389 23:42:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:18.389 23:42:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:18.389 23:42:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:18.389 23:42:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:18.389 23:42:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:18.389 23:42:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:18.389 23:42:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:18.389 23:42:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:18.389 23:42:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:18.389 23:42:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:18.389 "name": "Existed_Raid", 00:08:18.389 "uuid": "b03582e4-a081-4e8e-90e4-be690087a2af", 00:08:18.389 "strip_size_kb": 0, 00:08:18.389 "state": "online", 00:08:18.389 "raid_level": "raid1", 00:08:18.389 "superblock": false, 00:08:18.389 "num_base_bdevs": 3, 00:08:18.389 "num_base_bdevs_discovered": 2, 00:08:18.389 "num_base_bdevs_operational": 2, 00:08:18.389 "base_bdevs_list": [ 00:08:18.389 { 00:08:18.389 "name": null, 00:08:18.389 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:18.389 "is_configured": false, 00:08:18.389 "data_offset": 0, 00:08:18.389 "data_size": 65536 00:08:18.389 }, 00:08:18.389 { 00:08:18.389 "name": "BaseBdev2", 00:08:18.389 "uuid": "ed589290-e998-4956-8cab-f10b0163b609", 00:08:18.389 "is_configured": true, 00:08:18.389 "data_offset": 0, 00:08:18.389 "data_size": 65536 00:08:18.389 }, 00:08:18.389 { 00:08:18.389 "name": "BaseBdev3", 00:08:18.389 "uuid": "74381ae0-e7ea-4688-82dc-aa1f3191b179", 00:08:18.389 "is_configured": true, 00:08:18.389 "data_offset": 0, 00:08:18.389 "data_size": 65536 00:08:18.389 } 00:08:18.389 ] 00:08:18.389 }' 00:08:18.389 23:42:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:18.389 23:42:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:18.957 23:42:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:08:18.957 23:42:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:18.957 23:42:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:18.957 23:42:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:08:18.957 23:42:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:18.957 23:42:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:18.957 23:42:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:18.957 23:42:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:08:18.957 23:42:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:08:18.957 23:42:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:08:18.957 23:42:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:18.957 23:42:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:18.957 [2024-11-26 23:42:06.958211] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:08:18.957 23:42:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:18.957 23:42:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:08:18.957 23:42:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:18.957 23:42:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:18.957 23:42:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:18.957 23:42:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:18.957 23:42:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:08:18.957 23:42:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:18.957 23:42:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:08:18.957 23:42:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:08:18.957 23:42:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:08:18.957 23:42:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:18.957 23:42:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:18.957 [2024-11-26 23:42:07.021475] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:08:18.957 [2024-11-26 23:42:07.021564] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:18.957 [2024-11-26 23:42:07.033267] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:18.957 [2024-11-26 23:42:07.033417] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:18.957 [2024-11-26 23:42:07.033447] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:08:18.957 23:42:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:18.957 23:42:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:08:18.957 23:42:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:18.957 23:42:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:18.957 23:42:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:08:18.957 23:42:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:18.958 23:42:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:18.958 23:42:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:18.958 23:42:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:08:18.958 23:42:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:08:18.958 23:42:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 3 -gt 2 ']' 00:08:18.958 23:42:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:08:18.958 23:42:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:08:18.958 23:42:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:08:18.958 23:42:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:18.958 23:42:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:19.218 BaseBdev2 00:08:19.218 23:42:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:19.218 23:42:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:08:19.218 23:42:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:08:19.219 23:42:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:19.219 23:42:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:08:19.219 23:42:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:19.219 23:42:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:19.219 23:42:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:19.219 23:42:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:19.219 23:42:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:19.219 23:42:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:19.219 23:42:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:08:19.219 23:42:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:19.219 23:42:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:19.219 [ 00:08:19.219 { 00:08:19.219 "name": "BaseBdev2", 00:08:19.219 "aliases": [ 00:08:19.219 "20cfbe28-a743-4f5c-b433-e5ee184a7ed7" 00:08:19.219 ], 00:08:19.219 "product_name": "Malloc disk", 00:08:19.219 "block_size": 512, 00:08:19.219 "num_blocks": 65536, 00:08:19.219 "uuid": "20cfbe28-a743-4f5c-b433-e5ee184a7ed7", 00:08:19.219 "assigned_rate_limits": { 00:08:19.219 "rw_ios_per_sec": 0, 00:08:19.219 "rw_mbytes_per_sec": 0, 00:08:19.219 "r_mbytes_per_sec": 0, 00:08:19.219 "w_mbytes_per_sec": 0 00:08:19.219 }, 00:08:19.219 "claimed": false, 00:08:19.219 "zoned": false, 00:08:19.219 "supported_io_types": { 00:08:19.219 "read": true, 00:08:19.219 "write": true, 00:08:19.219 "unmap": true, 00:08:19.219 "flush": true, 00:08:19.219 "reset": true, 00:08:19.219 "nvme_admin": false, 00:08:19.219 "nvme_io": false, 00:08:19.219 "nvme_io_md": false, 00:08:19.219 "write_zeroes": true, 00:08:19.219 "zcopy": true, 00:08:19.219 "get_zone_info": false, 00:08:19.219 "zone_management": false, 00:08:19.219 "zone_append": false, 00:08:19.219 "compare": false, 00:08:19.219 "compare_and_write": false, 00:08:19.219 "abort": true, 00:08:19.219 "seek_hole": false, 00:08:19.219 "seek_data": false, 00:08:19.219 "copy": true, 00:08:19.219 "nvme_iov_md": false 00:08:19.219 }, 00:08:19.219 "memory_domains": [ 00:08:19.219 { 00:08:19.219 "dma_device_id": "system", 00:08:19.219 "dma_device_type": 1 00:08:19.219 }, 00:08:19.219 { 00:08:19.219 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:19.219 "dma_device_type": 2 00:08:19.219 } 00:08:19.219 ], 00:08:19.219 "driver_specific": {} 00:08:19.219 } 00:08:19.219 ] 00:08:19.219 23:42:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:19.219 23:42:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:08:19.219 23:42:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:08:19.219 23:42:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:08:19.219 23:42:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:08:19.219 23:42:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:19.219 23:42:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:19.219 BaseBdev3 00:08:19.219 23:42:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:19.219 23:42:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:08:19.219 23:42:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:08:19.219 23:42:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:19.219 23:42:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:08:19.219 23:42:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:19.219 23:42:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:19.219 23:42:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:19.219 23:42:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:19.219 23:42:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:19.219 23:42:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:19.219 23:42:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:08:19.219 23:42:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:19.219 23:42:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:19.219 [ 00:08:19.219 { 00:08:19.219 "name": "BaseBdev3", 00:08:19.219 "aliases": [ 00:08:19.219 "807922d6-be55-45da-88f6-d7afc6ddcfe3" 00:08:19.219 ], 00:08:19.219 "product_name": "Malloc disk", 00:08:19.219 "block_size": 512, 00:08:19.219 "num_blocks": 65536, 00:08:19.219 "uuid": "807922d6-be55-45da-88f6-d7afc6ddcfe3", 00:08:19.219 "assigned_rate_limits": { 00:08:19.219 "rw_ios_per_sec": 0, 00:08:19.219 "rw_mbytes_per_sec": 0, 00:08:19.219 "r_mbytes_per_sec": 0, 00:08:19.219 "w_mbytes_per_sec": 0 00:08:19.219 }, 00:08:19.219 "claimed": false, 00:08:19.219 "zoned": false, 00:08:19.219 "supported_io_types": { 00:08:19.219 "read": true, 00:08:19.219 "write": true, 00:08:19.219 "unmap": true, 00:08:19.219 "flush": true, 00:08:19.219 "reset": true, 00:08:19.219 "nvme_admin": false, 00:08:19.219 "nvme_io": false, 00:08:19.219 "nvme_io_md": false, 00:08:19.219 "write_zeroes": true, 00:08:19.219 "zcopy": true, 00:08:19.219 "get_zone_info": false, 00:08:19.219 "zone_management": false, 00:08:19.219 "zone_append": false, 00:08:19.219 "compare": false, 00:08:19.219 "compare_and_write": false, 00:08:19.219 "abort": true, 00:08:19.219 "seek_hole": false, 00:08:19.219 "seek_data": false, 00:08:19.219 "copy": true, 00:08:19.219 "nvme_iov_md": false 00:08:19.219 }, 00:08:19.219 "memory_domains": [ 00:08:19.219 { 00:08:19.219 "dma_device_id": "system", 00:08:19.219 "dma_device_type": 1 00:08:19.219 }, 00:08:19.219 { 00:08:19.219 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:19.219 "dma_device_type": 2 00:08:19.219 } 00:08:19.219 ], 00:08:19.219 "driver_specific": {} 00:08:19.219 } 00:08:19.219 ] 00:08:19.219 23:42:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:19.219 23:42:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:08:19.219 23:42:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:08:19.219 23:42:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:08:19.219 23:42:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:19.219 23:42:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:19.219 23:42:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:19.219 [2024-11-26 23:42:07.196888] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:19.219 [2024-11-26 23:42:07.196982] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:19.219 [2024-11-26 23:42:07.197053] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:19.219 [2024-11-26 23:42:07.199104] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:19.219 23:42:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:19.219 23:42:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:08:19.219 23:42:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:19.219 23:42:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:19.219 23:42:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:19.219 23:42:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:19.219 23:42:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:19.219 23:42:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:19.219 23:42:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:19.219 23:42:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:19.219 23:42:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:19.219 23:42:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:19.219 23:42:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:19.219 23:42:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:19.219 23:42:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:19.219 23:42:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:19.219 23:42:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:19.219 "name": "Existed_Raid", 00:08:19.219 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:19.219 "strip_size_kb": 0, 00:08:19.219 "state": "configuring", 00:08:19.219 "raid_level": "raid1", 00:08:19.219 "superblock": false, 00:08:19.219 "num_base_bdevs": 3, 00:08:19.219 "num_base_bdevs_discovered": 2, 00:08:19.219 "num_base_bdevs_operational": 3, 00:08:19.219 "base_bdevs_list": [ 00:08:19.219 { 00:08:19.219 "name": "BaseBdev1", 00:08:19.219 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:19.219 "is_configured": false, 00:08:19.219 "data_offset": 0, 00:08:19.219 "data_size": 0 00:08:19.219 }, 00:08:19.219 { 00:08:19.219 "name": "BaseBdev2", 00:08:19.219 "uuid": "20cfbe28-a743-4f5c-b433-e5ee184a7ed7", 00:08:19.219 "is_configured": true, 00:08:19.219 "data_offset": 0, 00:08:19.219 "data_size": 65536 00:08:19.219 }, 00:08:19.220 { 00:08:19.220 "name": "BaseBdev3", 00:08:19.220 "uuid": "807922d6-be55-45da-88f6-d7afc6ddcfe3", 00:08:19.220 "is_configured": true, 00:08:19.220 "data_offset": 0, 00:08:19.220 "data_size": 65536 00:08:19.220 } 00:08:19.220 ] 00:08:19.220 }' 00:08:19.220 23:42:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:19.220 23:42:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:19.790 23:42:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:08:19.790 23:42:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:19.790 23:42:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:19.790 [2024-11-26 23:42:07.664118] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:08:19.790 23:42:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:19.790 23:42:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:08:19.790 23:42:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:19.790 23:42:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:19.790 23:42:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:19.790 23:42:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:19.790 23:42:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:19.790 23:42:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:19.790 23:42:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:19.790 23:42:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:19.790 23:42:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:19.790 23:42:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:19.790 23:42:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:19.790 23:42:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:19.790 23:42:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:19.790 23:42:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:19.790 23:42:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:19.790 "name": "Existed_Raid", 00:08:19.790 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:19.790 "strip_size_kb": 0, 00:08:19.790 "state": "configuring", 00:08:19.790 "raid_level": "raid1", 00:08:19.790 "superblock": false, 00:08:19.790 "num_base_bdevs": 3, 00:08:19.790 "num_base_bdevs_discovered": 1, 00:08:19.790 "num_base_bdevs_operational": 3, 00:08:19.790 "base_bdevs_list": [ 00:08:19.790 { 00:08:19.790 "name": "BaseBdev1", 00:08:19.790 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:19.790 "is_configured": false, 00:08:19.790 "data_offset": 0, 00:08:19.790 "data_size": 0 00:08:19.790 }, 00:08:19.790 { 00:08:19.790 "name": null, 00:08:19.790 "uuid": "20cfbe28-a743-4f5c-b433-e5ee184a7ed7", 00:08:19.790 "is_configured": false, 00:08:19.790 "data_offset": 0, 00:08:19.790 "data_size": 65536 00:08:19.790 }, 00:08:19.790 { 00:08:19.790 "name": "BaseBdev3", 00:08:19.790 "uuid": "807922d6-be55-45da-88f6-d7afc6ddcfe3", 00:08:19.790 "is_configured": true, 00:08:19.790 "data_offset": 0, 00:08:19.790 "data_size": 65536 00:08:19.790 } 00:08:19.790 ] 00:08:19.790 }' 00:08:19.790 23:42:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:19.790 23:42:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:20.049 23:42:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:20.049 23:42:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:20.049 23:42:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:20.049 23:42:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:08:20.049 23:42:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:20.049 23:42:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:08:20.049 23:42:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:08:20.049 23:42:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:20.049 23:42:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:20.049 [2024-11-26 23:42:08.162245] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:20.049 BaseBdev1 00:08:20.049 23:42:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:20.049 23:42:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:08:20.049 23:42:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:08:20.049 23:42:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:20.049 23:42:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:08:20.049 23:42:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:20.049 23:42:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:20.049 23:42:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:20.049 23:42:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:20.049 23:42:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:20.049 23:42:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:20.049 23:42:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:08:20.049 23:42:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:20.049 23:42:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:20.308 [ 00:08:20.308 { 00:08:20.308 "name": "BaseBdev1", 00:08:20.308 "aliases": [ 00:08:20.308 "a5052e26-9954-47a5-8887-305a2ee7689f" 00:08:20.308 ], 00:08:20.308 "product_name": "Malloc disk", 00:08:20.308 "block_size": 512, 00:08:20.308 "num_blocks": 65536, 00:08:20.308 "uuid": "a5052e26-9954-47a5-8887-305a2ee7689f", 00:08:20.308 "assigned_rate_limits": { 00:08:20.308 "rw_ios_per_sec": 0, 00:08:20.308 "rw_mbytes_per_sec": 0, 00:08:20.308 "r_mbytes_per_sec": 0, 00:08:20.308 "w_mbytes_per_sec": 0 00:08:20.308 }, 00:08:20.308 "claimed": true, 00:08:20.308 "claim_type": "exclusive_write", 00:08:20.308 "zoned": false, 00:08:20.308 "supported_io_types": { 00:08:20.308 "read": true, 00:08:20.308 "write": true, 00:08:20.308 "unmap": true, 00:08:20.309 "flush": true, 00:08:20.309 "reset": true, 00:08:20.309 "nvme_admin": false, 00:08:20.309 "nvme_io": false, 00:08:20.309 "nvme_io_md": false, 00:08:20.309 "write_zeroes": true, 00:08:20.309 "zcopy": true, 00:08:20.309 "get_zone_info": false, 00:08:20.309 "zone_management": false, 00:08:20.309 "zone_append": false, 00:08:20.309 "compare": false, 00:08:20.309 "compare_and_write": false, 00:08:20.309 "abort": true, 00:08:20.309 "seek_hole": false, 00:08:20.309 "seek_data": false, 00:08:20.309 "copy": true, 00:08:20.309 "nvme_iov_md": false 00:08:20.309 }, 00:08:20.309 "memory_domains": [ 00:08:20.309 { 00:08:20.309 "dma_device_id": "system", 00:08:20.309 "dma_device_type": 1 00:08:20.309 }, 00:08:20.309 { 00:08:20.309 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:20.309 "dma_device_type": 2 00:08:20.309 } 00:08:20.309 ], 00:08:20.309 "driver_specific": {} 00:08:20.309 } 00:08:20.309 ] 00:08:20.309 23:42:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:20.309 23:42:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:08:20.309 23:42:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:08:20.309 23:42:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:20.309 23:42:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:20.309 23:42:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:20.309 23:42:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:20.309 23:42:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:20.309 23:42:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:20.309 23:42:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:20.309 23:42:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:20.309 23:42:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:20.309 23:42:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:20.309 23:42:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:20.309 23:42:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:20.309 23:42:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:20.309 23:42:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:20.309 23:42:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:20.309 "name": "Existed_Raid", 00:08:20.309 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:20.309 "strip_size_kb": 0, 00:08:20.309 "state": "configuring", 00:08:20.309 "raid_level": "raid1", 00:08:20.309 "superblock": false, 00:08:20.309 "num_base_bdevs": 3, 00:08:20.309 "num_base_bdevs_discovered": 2, 00:08:20.309 "num_base_bdevs_operational": 3, 00:08:20.309 "base_bdevs_list": [ 00:08:20.309 { 00:08:20.309 "name": "BaseBdev1", 00:08:20.309 "uuid": "a5052e26-9954-47a5-8887-305a2ee7689f", 00:08:20.309 "is_configured": true, 00:08:20.309 "data_offset": 0, 00:08:20.309 "data_size": 65536 00:08:20.309 }, 00:08:20.309 { 00:08:20.309 "name": null, 00:08:20.309 "uuid": "20cfbe28-a743-4f5c-b433-e5ee184a7ed7", 00:08:20.309 "is_configured": false, 00:08:20.309 "data_offset": 0, 00:08:20.309 "data_size": 65536 00:08:20.309 }, 00:08:20.309 { 00:08:20.309 "name": "BaseBdev3", 00:08:20.309 "uuid": "807922d6-be55-45da-88f6-d7afc6ddcfe3", 00:08:20.309 "is_configured": true, 00:08:20.309 "data_offset": 0, 00:08:20.309 "data_size": 65536 00:08:20.309 } 00:08:20.309 ] 00:08:20.309 }' 00:08:20.309 23:42:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:20.309 23:42:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:20.576 23:42:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:20.576 23:42:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:20.576 23:42:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:20.576 23:42:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:08:20.576 23:42:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:20.576 23:42:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:08:20.576 23:42:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:08:20.576 23:42:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:20.576 23:42:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:20.576 [2024-11-26 23:42:08.690230] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:08:20.576 23:42:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:20.576 23:42:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:08:20.576 23:42:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:20.576 23:42:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:20.576 23:42:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:20.576 23:42:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:20.576 23:42:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:20.576 23:42:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:20.576 23:42:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:20.576 23:42:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:20.576 23:42:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:20.853 23:42:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:20.853 23:42:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:20.853 23:42:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:20.853 23:42:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:20.853 23:42:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:20.853 23:42:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:20.853 "name": "Existed_Raid", 00:08:20.853 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:20.853 "strip_size_kb": 0, 00:08:20.853 "state": "configuring", 00:08:20.853 "raid_level": "raid1", 00:08:20.853 "superblock": false, 00:08:20.853 "num_base_bdevs": 3, 00:08:20.853 "num_base_bdevs_discovered": 1, 00:08:20.853 "num_base_bdevs_operational": 3, 00:08:20.853 "base_bdevs_list": [ 00:08:20.853 { 00:08:20.853 "name": "BaseBdev1", 00:08:20.853 "uuid": "a5052e26-9954-47a5-8887-305a2ee7689f", 00:08:20.853 "is_configured": true, 00:08:20.853 "data_offset": 0, 00:08:20.853 "data_size": 65536 00:08:20.853 }, 00:08:20.853 { 00:08:20.853 "name": null, 00:08:20.853 "uuid": "20cfbe28-a743-4f5c-b433-e5ee184a7ed7", 00:08:20.853 "is_configured": false, 00:08:20.853 "data_offset": 0, 00:08:20.853 "data_size": 65536 00:08:20.853 }, 00:08:20.853 { 00:08:20.853 "name": null, 00:08:20.853 "uuid": "807922d6-be55-45da-88f6-d7afc6ddcfe3", 00:08:20.853 "is_configured": false, 00:08:20.853 "data_offset": 0, 00:08:20.853 "data_size": 65536 00:08:20.853 } 00:08:20.853 ] 00:08:20.853 }' 00:08:20.853 23:42:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:20.853 23:42:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:21.127 23:42:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:21.127 23:42:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:08:21.127 23:42:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:21.127 23:42:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:21.127 23:42:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:21.127 23:42:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:08:21.127 23:42:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:08:21.127 23:42:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:21.127 23:42:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:21.127 [2024-11-26 23:42:09.165766] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:21.127 23:42:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:21.127 23:42:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:08:21.127 23:42:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:21.127 23:42:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:21.127 23:42:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:21.127 23:42:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:21.127 23:42:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:21.127 23:42:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:21.127 23:42:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:21.127 23:42:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:21.127 23:42:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:21.127 23:42:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:21.127 23:42:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:21.127 23:42:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:21.127 23:42:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:21.127 23:42:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:21.127 23:42:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:21.127 "name": "Existed_Raid", 00:08:21.127 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:21.127 "strip_size_kb": 0, 00:08:21.127 "state": "configuring", 00:08:21.127 "raid_level": "raid1", 00:08:21.127 "superblock": false, 00:08:21.127 "num_base_bdevs": 3, 00:08:21.127 "num_base_bdevs_discovered": 2, 00:08:21.127 "num_base_bdevs_operational": 3, 00:08:21.127 "base_bdevs_list": [ 00:08:21.127 { 00:08:21.127 "name": "BaseBdev1", 00:08:21.127 "uuid": "a5052e26-9954-47a5-8887-305a2ee7689f", 00:08:21.127 "is_configured": true, 00:08:21.127 "data_offset": 0, 00:08:21.127 "data_size": 65536 00:08:21.127 }, 00:08:21.127 { 00:08:21.127 "name": null, 00:08:21.127 "uuid": "20cfbe28-a743-4f5c-b433-e5ee184a7ed7", 00:08:21.127 "is_configured": false, 00:08:21.127 "data_offset": 0, 00:08:21.127 "data_size": 65536 00:08:21.127 }, 00:08:21.127 { 00:08:21.127 "name": "BaseBdev3", 00:08:21.127 "uuid": "807922d6-be55-45da-88f6-d7afc6ddcfe3", 00:08:21.127 "is_configured": true, 00:08:21.127 "data_offset": 0, 00:08:21.127 "data_size": 65536 00:08:21.127 } 00:08:21.127 ] 00:08:21.127 }' 00:08:21.127 23:42:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:21.127 23:42:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:21.698 23:42:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:08:21.698 23:42:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:21.698 23:42:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:21.698 23:42:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:21.698 23:42:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:21.698 23:42:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:08:21.698 23:42:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:08:21.698 23:42:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:21.698 23:42:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:21.698 [2024-11-26 23:42:09.664938] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:08:21.698 23:42:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:21.698 23:42:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:08:21.698 23:42:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:21.698 23:42:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:21.698 23:42:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:21.698 23:42:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:21.698 23:42:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:21.698 23:42:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:21.698 23:42:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:21.698 23:42:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:21.698 23:42:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:21.698 23:42:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:21.698 23:42:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:21.698 23:42:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:21.698 23:42:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:21.698 23:42:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:21.698 23:42:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:21.698 "name": "Existed_Raid", 00:08:21.698 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:21.698 "strip_size_kb": 0, 00:08:21.698 "state": "configuring", 00:08:21.698 "raid_level": "raid1", 00:08:21.698 "superblock": false, 00:08:21.698 "num_base_bdevs": 3, 00:08:21.698 "num_base_bdevs_discovered": 1, 00:08:21.698 "num_base_bdevs_operational": 3, 00:08:21.698 "base_bdevs_list": [ 00:08:21.698 { 00:08:21.698 "name": null, 00:08:21.698 "uuid": "a5052e26-9954-47a5-8887-305a2ee7689f", 00:08:21.698 "is_configured": false, 00:08:21.698 "data_offset": 0, 00:08:21.698 "data_size": 65536 00:08:21.698 }, 00:08:21.698 { 00:08:21.698 "name": null, 00:08:21.698 "uuid": "20cfbe28-a743-4f5c-b433-e5ee184a7ed7", 00:08:21.698 "is_configured": false, 00:08:21.698 "data_offset": 0, 00:08:21.698 "data_size": 65536 00:08:21.698 }, 00:08:21.698 { 00:08:21.698 "name": "BaseBdev3", 00:08:21.698 "uuid": "807922d6-be55-45da-88f6-d7afc6ddcfe3", 00:08:21.698 "is_configured": true, 00:08:21.698 "data_offset": 0, 00:08:21.698 "data_size": 65536 00:08:21.698 } 00:08:21.698 ] 00:08:21.698 }' 00:08:21.698 23:42:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:21.698 23:42:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:22.269 23:42:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:22.269 23:42:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:08:22.269 23:42:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:22.269 23:42:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:22.269 23:42:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:22.269 23:42:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:08:22.269 23:42:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:08:22.269 23:42:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:22.269 23:42:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:22.269 [2024-11-26 23:42:10.198652] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:22.269 23:42:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:22.269 23:42:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:08:22.269 23:42:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:22.269 23:42:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:22.269 23:42:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:22.269 23:42:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:22.269 23:42:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:22.269 23:42:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:22.269 23:42:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:22.269 23:42:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:22.269 23:42:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:22.269 23:42:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:22.269 23:42:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:22.269 23:42:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:22.269 23:42:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:22.269 23:42:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:22.269 23:42:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:22.269 "name": "Existed_Raid", 00:08:22.269 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:22.269 "strip_size_kb": 0, 00:08:22.269 "state": "configuring", 00:08:22.269 "raid_level": "raid1", 00:08:22.269 "superblock": false, 00:08:22.269 "num_base_bdevs": 3, 00:08:22.269 "num_base_bdevs_discovered": 2, 00:08:22.269 "num_base_bdevs_operational": 3, 00:08:22.269 "base_bdevs_list": [ 00:08:22.269 { 00:08:22.269 "name": null, 00:08:22.269 "uuid": "a5052e26-9954-47a5-8887-305a2ee7689f", 00:08:22.269 "is_configured": false, 00:08:22.269 "data_offset": 0, 00:08:22.269 "data_size": 65536 00:08:22.269 }, 00:08:22.269 { 00:08:22.269 "name": "BaseBdev2", 00:08:22.269 "uuid": "20cfbe28-a743-4f5c-b433-e5ee184a7ed7", 00:08:22.269 "is_configured": true, 00:08:22.269 "data_offset": 0, 00:08:22.269 "data_size": 65536 00:08:22.269 }, 00:08:22.269 { 00:08:22.269 "name": "BaseBdev3", 00:08:22.269 "uuid": "807922d6-be55-45da-88f6-d7afc6ddcfe3", 00:08:22.269 "is_configured": true, 00:08:22.269 "data_offset": 0, 00:08:22.269 "data_size": 65536 00:08:22.269 } 00:08:22.269 ] 00:08:22.269 }' 00:08:22.269 23:42:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:22.269 23:42:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:22.838 23:42:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:08:22.838 23:42:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:22.838 23:42:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:22.838 23:42:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:22.838 23:42:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:22.838 23:42:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:08:22.838 23:42:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:08:22.838 23:42:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:22.838 23:42:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:22.838 23:42:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:22.838 23:42:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:22.838 23:42:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u a5052e26-9954-47a5-8887-305a2ee7689f 00:08:22.838 23:42:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:22.838 23:42:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:22.838 [2024-11-26 23:42:10.741092] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:08:22.838 [2024-11-26 23:42:10.741197] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:08:22.838 [2024-11-26 23:42:10.741223] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:08:22.838 [2024-11-26 23:42:10.741539] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002870 00:08:22.838 [2024-11-26 23:42:10.741702] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:08:22.838 [2024-11-26 23:42:10.741747] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001c80 00:08:22.838 [2024-11-26 23:42:10.741960] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:22.838 NewBaseBdev 00:08:22.838 23:42:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:22.838 23:42:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:08:22.838 23:42:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:08:22.838 23:42:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:22.838 23:42:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:08:22.838 23:42:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:22.838 23:42:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:22.838 23:42:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:22.838 23:42:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:22.838 23:42:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:22.839 23:42:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:22.839 23:42:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:08:22.839 23:42:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:22.839 23:42:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:22.839 [ 00:08:22.839 { 00:08:22.839 "name": "NewBaseBdev", 00:08:22.839 "aliases": [ 00:08:22.839 "a5052e26-9954-47a5-8887-305a2ee7689f" 00:08:22.839 ], 00:08:22.839 "product_name": "Malloc disk", 00:08:22.839 "block_size": 512, 00:08:22.839 "num_blocks": 65536, 00:08:22.839 "uuid": "a5052e26-9954-47a5-8887-305a2ee7689f", 00:08:22.839 "assigned_rate_limits": { 00:08:22.839 "rw_ios_per_sec": 0, 00:08:22.839 "rw_mbytes_per_sec": 0, 00:08:22.839 "r_mbytes_per_sec": 0, 00:08:22.839 "w_mbytes_per_sec": 0 00:08:22.839 }, 00:08:22.839 "claimed": true, 00:08:22.839 "claim_type": "exclusive_write", 00:08:22.839 "zoned": false, 00:08:22.839 "supported_io_types": { 00:08:22.839 "read": true, 00:08:22.839 "write": true, 00:08:22.839 "unmap": true, 00:08:22.839 "flush": true, 00:08:22.839 "reset": true, 00:08:22.839 "nvme_admin": false, 00:08:22.839 "nvme_io": false, 00:08:22.839 "nvme_io_md": false, 00:08:22.839 "write_zeroes": true, 00:08:22.839 "zcopy": true, 00:08:22.839 "get_zone_info": false, 00:08:22.839 "zone_management": false, 00:08:22.839 "zone_append": false, 00:08:22.839 "compare": false, 00:08:22.839 "compare_and_write": false, 00:08:22.839 "abort": true, 00:08:22.839 "seek_hole": false, 00:08:22.839 "seek_data": false, 00:08:22.839 "copy": true, 00:08:22.839 "nvme_iov_md": false 00:08:22.839 }, 00:08:22.839 "memory_domains": [ 00:08:22.839 { 00:08:22.839 "dma_device_id": "system", 00:08:22.839 "dma_device_type": 1 00:08:22.839 }, 00:08:22.839 { 00:08:22.839 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:22.839 "dma_device_type": 2 00:08:22.839 } 00:08:22.839 ], 00:08:22.839 "driver_specific": {} 00:08:22.839 } 00:08:22.839 ] 00:08:22.839 23:42:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:22.839 23:42:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:08:22.839 23:42:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid1 0 3 00:08:22.839 23:42:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:22.839 23:42:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:22.839 23:42:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:22.839 23:42:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:22.839 23:42:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:22.839 23:42:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:22.839 23:42:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:22.839 23:42:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:22.839 23:42:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:22.839 23:42:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:22.839 23:42:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:22.839 23:42:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:22.839 23:42:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:22.839 23:42:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:22.839 23:42:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:22.839 "name": "Existed_Raid", 00:08:22.839 "uuid": "424b8c10-76cf-474c-9abc-2bc72682ec09", 00:08:22.839 "strip_size_kb": 0, 00:08:22.839 "state": "online", 00:08:22.839 "raid_level": "raid1", 00:08:22.839 "superblock": false, 00:08:22.839 "num_base_bdevs": 3, 00:08:22.839 "num_base_bdevs_discovered": 3, 00:08:22.839 "num_base_bdevs_operational": 3, 00:08:22.839 "base_bdevs_list": [ 00:08:22.839 { 00:08:22.839 "name": "NewBaseBdev", 00:08:22.839 "uuid": "a5052e26-9954-47a5-8887-305a2ee7689f", 00:08:22.839 "is_configured": true, 00:08:22.839 "data_offset": 0, 00:08:22.839 "data_size": 65536 00:08:22.839 }, 00:08:22.839 { 00:08:22.839 "name": "BaseBdev2", 00:08:22.839 "uuid": "20cfbe28-a743-4f5c-b433-e5ee184a7ed7", 00:08:22.839 "is_configured": true, 00:08:22.839 "data_offset": 0, 00:08:22.839 "data_size": 65536 00:08:22.839 }, 00:08:22.839 { 00:08:22.839 "name": "BaseBdev3", 00:08:22.839 "uuid": "807922d6-be55-45da-88f6-d7afc6ddcfe3", 00:08:22.839 "is_configured": true, 00:08:22.839 "data_offset": 0, 00:08:22.839 "data_size": 65536 00:08:22.839 } 00:08:22.839 ] 00:08:22.839 }' 00:08:22.839 23:42:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:22.839 23:42:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:23.099 23:42:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:08:23.099 23:42:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:08:23.099 23:42:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:23.099 23:42:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:23.099 23:42:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:08:23.099 23:42:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:23.099 23:42:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:23.099 23:42:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:08:23.099 23:42:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:23.099 23:42:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:23.357 [2024-11-26 23:42:11.228621] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:23.357 23:42:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:23.357 23:42:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:23.357 "name": "Existed_Raid", 00:08:23.357 "aliases": [ 00:08:23.357 "424b8c10-76cf-474c-9abc-2bc72682ec09" 00:08:23.357 ], 00:08:23.357 "product_name": "Raid Volume", 00:08:23.357 "block_size": 512, 00:08:23.357 "num_blocks": 65536, 00:08:23.357 "uuid": "424b8c10-76cf-474c-9abc-2bc72682ec09", 00:08:23.357 "assigned_rate_limits": { 00:08:23.357 "rw_ios_per_sec": 0, 00:08:23.357 "rw_mbytes_per_sec": 0, 00:08:23.357 "r_mbytes_per_sec": 0, 00:08:23.357 "w_mbytes_per_sec": 0 00:08:23.357 }, 00:08:23.357 "claimed": false, 00:08:23.357 "zoned": false, 00:08:23.357 "supported_io_types": { 00:08:23.357 "read": true, 00:08:23.357 "write": true, 00:08:23.357 "unmap": false, 00:08:23.357 "flush": false, 00:08:23.357 "reset": true, 00:08:23.357 "nvme_admin": false, 00:08:23.357 "nvme_io": false, 00:08:23.357 "nvme_io_md": false, 00:08:23.357 "write_zeroes": true, 00:08:23.357 "zcopy": false, 00:08:23.357 "get_zone_info": false, 00:08:23.357 "zone_management": false, 00:08:23.357 "zone_append": false, 00:08:23.357 "compare": false, 00:08:23.357 "compare_and_write": false, 00:08:23.357 "abort": false, 00:08:23.357 "seek_hole": false, 00:08:23.357 "seek_data": false, 00:08:23.357 "copy": false, 00:08:23.357 "nvme_iov_md": false 00:08:23.357 }, 00:08:23.357 "memory_domains": [ 00:08:23.357 { 00:08:23.357 "dma_device_id": "system", 00:08:23.357 "dma_device_type": 1 00:08:23.357 }, 00:08:23.357 { 00:08:23.357 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:23.357 "dma_device_type": 2 00:08:23.357 }, 00:08:23.357 { 00:08:23.357 "dma_device_id": "system", 00:08:23.357 "dma_device_type": 1 00:08:23.357 }, 00:08:23.357 { 00:08:23.357 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:23.357 "dma_device_type": 2 00:08:23.357 }, 00:08:23.357 { 00:08:23.357 "dma_device_id": "system", 00:08:23.357 "dma_device_type": 1 00:08:23.357 }, 00:08:23.357 { 00:08:23.357 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:23.357 "dma_device_type": 2 00:08:23.357 } 00:08:23.357 ], 00:08:23.357 "driver_specific": { 00:08:23.357 "raid": { 00:08:23.357 "uuid": "424b8c10-76cf-474c-9abc-2bc72682ec09", 00:08:23.357 "strip_size_kb": 0, 00:08:23.357 "state": "online", 00:08:23.357 "raid_level": "raid1", 00:08:23.357 "superblock": false, 00:08:23.357 "num_base_bdevs": 3, 00:08:23.357 "num_base_bdevs_discovered": 3, 00:08:23.357 "num_base_bdevs_operational": 3, 00:08:23.357 "base_bdevs_list": [ 00:08:23.357 { 00:08:23.357 "name": "NewBaseBdev", 00:08:23.357 "uuid": "a5052e26-9954-47a5-8887-305a2ee7689f", 00:08:23.357 "is_configured": true, 00:08:23.357 "data_offset": 0, 00:08:23.357 "data_size": 65536 00:08:23.357 }, 00:08:23.357 { 00:08:23.357 "name": "BaseBdev2", 00:08:23.357 "uuid": "20cfbe28-a743-4f5c-b433-e5ee184a7ed7", 00:08:23.357 "is_configured": true, 00:08:23.357 "data_offset": 0, 00:08:23.357 "data_size": 65536 00:08:23.357 }, 00:08:23.357 { 00:08:23.357 "name": "BaseBdev3", 00:08:23.357 "uuid": "807922d6-be55-45da-88f6-d7afc6ddcfe3", 00:08:23.357 "is_configured": true, 00:08:23.357 "data_offset": 0, 00:08:23.357 "data_size": 65536 00:08:23.357 } 00:08:23.357 ] 00:08:23.357 } 00:08:23.357 } 00:08:23.357 }' 00:08:23.357 23:42:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:23.357 23:42:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:08:23.357 BaseBdev2 00:08:23.357 BaseBdev3' 00:08:23.357 23:42:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:23.357 23:42:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:23.357 23:42:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:23.357 23:42:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:08:23.357 23:42:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:23.357 23:42:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:23.357 23:42:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:23.357 23:42:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:23.357 23:42:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:23.357 23:42:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:23.357 23:42:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:23.357 23:42:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:23.357 23:42:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:08:23.357 23:42:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:23.357 23:42:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:23.357 23:42:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:23.357 23:42:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:23.357 23:42:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:23.357 23:42:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:23.357 23:42:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:08:23.357 23:42:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:23.357 23:42:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:23.357 23:42:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:23.357 23:42:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:23.616 23:42:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:23.616 23:42:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:23.616 23:42:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:23.616 23:42:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:23.616 23:42:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:23.616 [2024-11-26 23:42:11.503816] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:23.616 [2024-11-26 23:42:11.503843] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:23.616 [2024-11-26 23:42:11.503915] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:23.616 [2024-11-26 23:42:11.504161] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:23.616 [2024-11-26 23:42:11.504171] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name Existed_Raid, state offline 00:08:23.616 23:42:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:23.616 23:42:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 78162 00:08:23.616 23:42:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # '[' -z 78162 ']' 00:08:23.616 23:42:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@958 -- # kill -0 78162 00:08:23.616 23:42:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # uname 00:08:23.616 23:42:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:08:23.616 23:42:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 78162 00:08:23.616 23:42:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:08:23.616 23:42:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:08:23.616 23:42:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 78162' 00:08:23.616 killing process with pid 78162 00:08:23.616 23:42:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@973 -- # kill 78162 00:08:23.616 [2024-11-26 23:42:11.552973] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:23.616 23:42:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@978 -- # wait 78162 00:08:23.616 [2024-11-26 23:42:11.584832] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:23.876 23:42:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:08:23.876 00:08:23.876 real 0m9.033s 00:08:23.876 user 0m15.528s 00:08:23.876 sys 0m1.756s 00:08:23.876 23:42:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:08:23.876 23:42:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:23.876 ************************************ 00:08:23.876 END TEST raid_state_function_test 00:08:23.876 ************************************ 00:08:23.876 23:42:11 bdev_raid -- bdev/bdev_raid.sh@969 -- # run_test raid_state_function_test_sb raid_state_function_test raid1 3 true 00:08:23.876 23:42:11 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:08:23.876 23:42:11 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:08:23.876 23:42:11 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:23.876 ************************************ 00:08:23.876 START TEST raid_state_function_test_sb 00:08:23.876 ************************************ 00:08:23.876 23:42:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1129 -- # raid_state_function_test raid1 3 true 00:08:23.876 23:42:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=raid1 00:08:23.876 23:42:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=3 00:08:23.876 23:42:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:08:23.876 23:42:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:08:23.876 23:42:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:08:23.876 23:42:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:23.876 23:42:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:08:23.876 23:42:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:23.876 23:42:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:23.876 23:42:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:08:23.876 23:42:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:23.876 23:42:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:23.876 23:42:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:08:23.876 23:42:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:23.876 23:42:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:23.876 23:42:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:08:23.876 23:42:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:08:23.876 23:42:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:08:23.876 23:42:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:08:23.876 23:42:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:08:23.876 23:42:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:08:23.876 23:42:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' raid1 '!=' raid1 ']' 00:08:23.876 23:42:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@219 -- # strip_size=0 00:08:23.876 23:42:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:08:23.876 23:42:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:08:23.876 23:42:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=78766 00:08:23.876 23:42:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:08:23.876 23:42:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 78766' 00:08:23.876 Process raid pid: 78766 00:08:23.876 23:42:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 78766 00:08:23.876 23:42:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@835 -- # '[' -z 78766 ']' 00:08:23.876 23:42:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:23.876 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:23.876 23:42:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@840 -- # local max_retries=100 00:08:23.876 23:42:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:23.876 23:42:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@844 -- # xtrace_disable 00:08:23.876 23:42:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:23.876 [2024-11-26 23:42:11.967115] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:08:23.877 [2024-11-26 23:42:11.967248] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:08:24.136 [2024-11-26 23:42:12.119923] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:24.136 [2024-11-26 23:42:12.144659] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:24.136 [2024-11-26 23:42:12.186665] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:24.136 [2024-11-26 23:42:12.186792] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:24.705 23:42:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:08:24.705 23:42:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@868 -- # return 0 00:08:24.705 23:42:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:24.705 23:42:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:24.705 23:42:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:24.705 [2024-11-26 23:42:12.813224] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:24.705 [2024-11-26 23:42:12.813394] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:24.705 [2024-11-26 23:42:12.813422] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:24.705 [2024-11-26 23:42:12.813452] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:24.705 [2024-11-26 23:42:12.813469] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:08:24.705 [2024-11-26 23:42:12.813497] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:08:24.705 23:42:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:24.705 23:42:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:08:24.705 23:42:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:24.705 23:42:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:24.705 23:42:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:24.705 23:42:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:24.705 23:42:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:24.705 23:42:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:24.705 23:42:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:24.705 23:42:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:24.705 23:42:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:24.705 23:42:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:24.705 23:42:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:24.705 23:42:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:24.705 23:42:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:24.963 23:42:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:24.963 23:42:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:24.963 "name": "Existed_Raid", 00:08:24.963 "uuid": "4d0615e8-fe2f-4de5-8fd3-253eccfc6ee9", 00:08:24.963 "strip_size_kb": 0, 00:08:24.963 "state": "configuring", 00:08:24.963 "raid_level": "raid1", 00:08:24.963 "superblock": true, 00:08:24.963 "num_base_bdevs": 3, 00:08:24.963 "num_base_bdevs_discovered": 0, 00:08:24.963 "num_base_bdevs_operational": 3, 00:08:24.963 "base_bdevs_list": [ 00:08:24.963 { 00:08:24.963 "name": "BaseBdev1", 00:08:24.963 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:24.963 "is_configured": false, 00:08:24.963 "data_offset": 0, 00:08:24.963 "data_size": 0 00:08:24.963 }, 00:08:24.963 { 00:08:24.963 "name": "BaseBdev2", 00:08:24.963 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:24.963 "is_configured": false, 00:08:24.963 "data_offset": 0, 00:08:24.963 "data_size": 0 00:08:24.963 }, 00:08:24.963 { 00:08:24.963 "name": "BaseBdev3", 00:08:24.963 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:24.963 "is_configured": false, 00:08:24.963 "data_offset": 0, 00:08:24.963 "data_size": 0 00:08:24.963 } 00:08:24.963 ] 00:08:24.963 }' 00:08:24.963 23:42:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:24.963 23:42:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:25.220 23:42:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:25.220 23:42:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:25.220 23:42:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:25.220 [2024-11-26 23:42:13.284137] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:25.221 [2024-11-26 23:42:13.284277] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:08:25.221 23:42:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:25.221 23:42:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:25.221 23:42:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:25.221 23:42:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:25.221 [2024-11-26 23:42:13.296125] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:25.221 [2024-11-26 23:42:13.296175] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:25.221 [2024-11-26 23:42:13.296185] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:25.221 [2024-11-26 23:42:13.296195] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:25.221 [2024-11-26 23:42:13.296202] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:08:25.221 [2024-11-26 23:42:13.296211] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:08:25.221 23:42:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:25.221 23:42:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:08:25.221 23:42:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:25.221 23:42:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:25.221 [2024-11-26 23:42:13.324168] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:25.221 BaseBdev1 00:08:25.221 23:42:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:25.221 23:42:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:08:25.221 23:42:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:08:25.221 23:42:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:25.221 23:42:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:08:25.221 23:42:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:25.221 23:42:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:25.221 23:42:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:25.221 23:42:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:25.221 23:42:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:25.221 23:42:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:25.221 23:42:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:08:25.221 23:42:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:25.221 23:42:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:25.221 [ 00:08:25.221 { 00:08:25.221 "name": "BaseBdev1", 00:08:25.221 "aliases": [ 00:08:25.480 "6b1f6948-e827-48a9-baa5-cd1e97fd5603" 00:08:25.480 ], 00:08:25.480 "product_name": "Malloc disk", 00:08:25.480 "block_size": 512, 00:08:25.480 "num_blocks": 65536, 00:08:25.480 "uuid": "6b1f6948-e827-48a9-baa5-cd1e97fd5603", 00:08:25.480 "assigned_rate_limits": { 00:08:25.480 "rw_ios_per_sec": 0, 00:08:25.480 "rw_mbytes_per_sec": 0, 00:08:25.480 "r_mbytes_per_sec": 0, 00:08:25.480 "w_mbytes_per_sec": 0 00:08:25.480 }, 00:08:25.480 "claimed": true, 00:08:25.480 "claim_type": "exclusive_write", 00:08:25.480 "zoned": false, 00:08:25.480 "supported_io_types": { 00:08:25.480 "read": true, 00:08:25.480 "write": true, 00:08:25.480 "unmap": true, 00:08:25.480 "flush": true, 00:08:25.480 "reset": true, 00:08:25.480 "nvme_admin": false, 00:08:25.480 "nvme_io": false, 00:08:25.480 "nvme_io_md": false, 00:08:25.480 "write_zeroes": true, 00:08:25.480 "zcopy": true, 00:08:25.480 "get_zone_info": false, 00:08:25.480 "zone_management": false, 00:08:25.480 "zone_append": false, 00:08:25.480 "compare": false, 00:08:25.480 "compare_and_write": false, 00:08:25.480 "abort": true, 00:08:25.480 "seek_hole": false, 00:08:25.480 "seek_data": false, 00:08:25.480 "copy": true, 00:08:25.480 "nvme_iov_md": false 00:08:25.480 }, 00:08:25.480 "memory_domains": [ 00:08:25.480 { 00:08:25.480 "dma_device_id": "system", 00:08:25.480 "dma_device_type": 1 00:08:25.480 }, 00:08:25.480 { 00:08:25.480 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:25.480 "dma_device_type": 2 00:08:25.480 } 00:08:25.480 ], 00:08:25.480 "driver_specific": {} 00:08:25.480 } 00:08:25.480 ] 00:08:25.480 23:42:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:25.480 23:42:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:08:25.480 23:42:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:08:25.480 23:42:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:25.480 23:42:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:25.480 23:42:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:25.480 23:42:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:25.480 23:42:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:25.480 23:42:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:25.480 23:42:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:25.480 23:42:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:25.480 23:42:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:25.480 23:42:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:25.480 23:42:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:25.480 23:42:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:25.480 23:42:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:25.480 23:42:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:25.480 23:42:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:25.480 "name": "Existed_Raid", 00:08:25.480 "uuid": "df4daabb-fd1a-4b99-aa9d-80aaa23c8813", 00:08:25.480 "strip_size_kb": 0, 00:08:25.480 "state": "configuring", 00:08:25.480 "raid_level": "raid1", 00:08:25.480 "superblock": true, 00:08:25.480 "num_base_bdevs": 3, 00:08:25.480 "num_base_bdevs_discovered": 1, 00:08:25.480 "num_base_bdevs_operational": 3, 00:08:25.480 "base_bdevs_list": [ 00:08:25.480 { 00:08:25.480 "name": "BaseBdev1", 00:08:25.480 "uuid": "6b1f6948-e827-48a9-baa5-cd1e97fd5603", 00:08:25.480 "is_configured": true, 00:08:25.480 "data_offset": 2048, 00:08:25.480 "data_size": 63488 00:08:25.480 }, 00:08:25.480 { 00:08:25.480 "name": "BaseBdev2", 00:08:25.480 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:25.480 "is_configured": false, 00:08:25.480 "data_offset": 0, 00:08:25.480 "data_size": 0 00:08:25.480 }, 00:08:25.480 { 00:08:25.480 "name": "BaseBdev3", 00:08:25.480 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:25.480 "is_configured": false, 00:08:25.480 "data_offset": 0, 00:08:25.480 "data_size": 0 00:08:25.480 } 00:08:25.480 ] 00:08:25.480 }' 00:08:25.480 23:42:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:25.480 23:42:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:25.739 23:42:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:25.739 23:42:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:25.739 23:42:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:25.739 [2024-11-26 23:42:13.803532] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:25.739 [2024-11-26 23:42:13.803694] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:08:25.739 23:42:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:25.739 23:42:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:25.739 23:42:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:25.739 23:42:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:25.739 [2024-11-26 23:42:13.815572] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:25.739 [2024-11-26 23:42:13.818005] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:25.739 [2024-11-26 23:42:13.818052] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:25.739 [2024-11-26 23:42:13.818063] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:08:25.740 [2024-11-26 23:42:13.818074] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:08:25.740 23:42:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:25.740 23:42:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:08:25.740 23:42:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:25.740 23:42:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:08:25.740 23:42:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:25.740 23:42:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:25.740 23:42:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:25.740 23:42:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:25.740 23:42:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:25.740 23:42:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:25.740 23:42:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:25.740 23:42:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:25.740 23:42:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:25.740 23:42:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:25.740 23:42:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:25.740 23:42:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:25.740 23:42:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:25.740 23:42:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:25.740 23:42:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:25.740 "name": "Existed_Raid", 00:08:25.740 "uuid": "32ee297c-d6c2-4630-9515-65c747894357", 00:08:25.740 "strip_size_kb": 0, 00:08:25.740 "state": "configuring", 00:08:25.740 "raid_level": "raid1", 00:08:25.740 "superblock": true, 00:08:25.740 "num_base_bdevs": 3, 00:08:25.740 "num_base_bdevs_discovered": 1, 00:08:25.740 "num_base_bdevs_operational": 3, 00:08:25.740 "base_bdevs_list": [ 00:08:25.740 { 00:08:25.740 "name": "BaseBdev1", 00:08:25.740 "uuid": "6b1f6948-e827-48a9-baa5-cd1e97fd5603", 00:08:25.740 "is_configured": true, 00:08:25.740 "data_offset": 2048, 00:08:25.740 "data_size": 63488 00:08:25.740 }, 00:08:25.740 { 00:08:25.740 "name": "BaseBdev2", 00:08:25.740 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:25.740 "is_configured": false, 00:08:25.740 "data_offset": 0, 00:08:25.740 "data_size": 0 00:08:25.740 }, 00:08:25.740 { 00:08:25.740 "name": "BaseBdev3", 00:08:25.740 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:25.740 "is_configured": false, 00:08:25.740 "data_offset": 0, 00:08:25.740 "data_size": 0 00:08:25.740 } 00:08:25.740 ] 00:08:25.740 }' 00:08:25.740 23:42:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:25.740 23:42:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:26.306 23:42:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:08:26.306 23:42:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:26.306 23:42:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:26.306 [2024-11-26 23:42:14.259640] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:26.306 BaseBdev2 00:08:26.306 23:42:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:26.306 23:42:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:08:26.306 23:42:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:08:26.306 23:42:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:26.306 23:42:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:08:26.306 23:42:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:26.306 23:42:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:26.306 23:42:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:26.306 23:42:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:26.306 23:42:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:26.306 23:42:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:26.306 23:42:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:08:26.306 23:42:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:26.306 23:42:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:26.306 [ 00:08:26.306 { 00:08:26.306 "name": "BaseBdev2", 00:08:26.306 "aliases": [ 00:08:26.306 "9ffcbe45-b53d-4fe5-ab35-dca86f99beef" 00:08:26.306 ], 00:08:26.306 "product_name": "Malloc disk", 00:08:26.306 "block_size": 512, 00:08:26.306 "num_blocks": 65536, 00:08:26.306 "uuid": "9ffcbe45-b53d-4fe5-ab35-dca86f99beef", 00:08:26.306 "assigned_rate_limits": { 00:08:26.306 "rw_ios_per_sec": 0, 00:08:26.306 "rw_mbytes_per_sec": 0, 00:08:26.306 "r_mbytes_per_sec": 0, 00:08:26.306 "w_mbytes_per_sec": 0 00:08:26.306 }, 00:08:26.306 "claimed": true, 00:08:26.306 "claim_type": "exclusive_write", 00:08:26.306 "zoned": false, 00:08:26.306 "supported_io_types": { 00:08:26.306 "read": true, 00:08:26.306 "write": true, 00:08:26.306 "unmap": true, 00:08:26.306 "flush": true, 00:08:26.306 "reset": true, 00:08:26.306 "nvme_admin": false, 00:08:26.306 "nvme_io": false, 00:08:26.306 "nvme_io_md": false, 00:08:26.306 "write_zeroes": true, 00:08:26.306 "zcopy": true, 00:08:26.306 "get_zone_info": false, 00:08:26.306 "zone_management": false, 00:08:26.306 "zone_append": false, 00:08:26.306 "compare": false, 00:08:26.306 "compare_and_write": false, 00:08:26.306 "abort": true, 00:08:26.306 "seek_hole": false, 00:08:26.306 "seek_data": false, 00:08:26.306 "copy": true, 00:08:26.306 "nvme_iov_md": false 00:08:26.306 }, 00:08:26.306 "memory_domains": [ 00:08:26.306 { 00:08:26.306 "dma_device_id": "system", 00:08:26.306 "dma_device_type": 1 00:08:26.306 }, 00:08:26.306 { 00:08:26.306 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:26.306 "dma_device_type": 2 00:08:26.306 } 00:08:26.306 ], 00:08:26.306 "driver_specific": {} 00:08:26.306 } 00:08:26.306 ] 00:08:26.306 23:42:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:26.306 23:42:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:08:26.306 23:42:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:08:26.306 23:42:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:26.306 23:42:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:08:26.306 23:42:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:26.306 23:42:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:26.306 23:42:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:26.306 23:42:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:26.306 23:42:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:26.306 23:42:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:26.306 23:42:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:26.306 23:42:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:26.306 23:42:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:26.306 23:42:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:26.306 23:42:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:26.306 23:42:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:26.306 23:42:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:26.306 23:42:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:26.306 23:42:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:26.306 "name": "Existed_Raid", 00:08:26.306 "uuid": "32ee297c-d6c2-4630-9515-65c747894357", 00:08:26.306 "strip_size_kb": 0, 00:08:26.306 "state": "configuring", 00:08:26.306 "raid_level": "raid1", 00:08:26.306 "superblock": true, 00:08:26.306 "num_base_bdevs": 3, 00:08:26.307 "num_base_bdevs_discovered": 2, 00:08:26.307 "num_base_bdevs_operational": 3, 00:08:26.307 "base_bdevs_list": [ 00:08:26.307 { 00:08:26.307 "name": "BaseBdev1", 00:08:26.307 "uuid": "6b1f6948-e827-48a9-baa5-cd1e97fd5603", 00:08:26.307 "is_configured": true, 00:08:26.307 "data_offset": 2048, 00:08:26.307 "data_size": 63488 00:08:26.307 }, 00:08:26.307 { 00:08:26.307 "name": "BaseBdev2", 00:08:26.307 "uuid": "9ffcbe45-b53d-4fe5-ab35-dca86f99beef", 00:08:26.307 "is_configured": true, 00:08:26.307 "data_offset": 2048, 00:08:26.307 "data_size": 63488 00:08:26.307 }, 00:08:26.307 { 00:08:26.307 "name": "BaseBdev3", 00:08:26.307 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:26.307 "is_configured": false, 00:08:26.307 "data_offset": 0, 00:08:26.307 "data_size": 0 00:08:26.307 } 00:08:26.307 ] 00:08:26.307 }' 00:08:26.307 23:42:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:26.307 23:42:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:26.874 23:42:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:08:26.874 23:42:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:26.874 23:42:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:26.874 [2024-11-26 23:42:14.773614] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:26.874 [2024-11-26 23:42:14.773900] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:08:26.874 [2024-11-26 23:42:14.773927] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:08:26.874 BaseBdev3 00:08:26.874 [2024-11-26 23:42:14.774417] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:08:26.874 [2024-11-26 23:42:14.774649] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:08:26.874 [2024-11-26 23:42:14.774673] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:08:26.874 [2024-11-26 23:42:14.774835] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:26.874 23:42:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:26.874 23:42:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:08:26.874 23:42:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:08:26.874 23:42:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:26.874 23:42:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:08:26.874 23:42:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:26.874 23:42:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:26.874 23:42:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:26.874 23:42:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:26.874 23:42:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:26.874 23:42:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:26.874 23:42:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:08:26.874 23:42:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:26.874 23:42:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:26.874 [ 00:08:26.874 { 00:08:26.874 "name": "BaseBdev3", 00:08:26.874 "aliases": [ 00:08:26.874 "4984b65e-1e8c-4ce4-99de-3a26dff3e7d2" 00:08:26.874 ], 00:08:26.874 "product_name": "Malloc disk", 00:08:26.874 "block_size": 512, 00:08:26.874 "num_blocks": 65536, 00:08:26.874 "uuid": "4984b65e-1e8c-4ce4-99de-3a26dff3e7d2", 00:08:26.874 "assigned_rate_limits": { 00:08:26.874 "rw_ios_per_sec": 0, 00:08:26.874 "rw_mbytes_per_sec": 0, 00:08:26.874 "r_mbytes_per_sec": 0, 00:08:26.874 "w_mbytes_per_sec": 0 00:08:26.874 }, 00:08:26.874 "claimed": true, 00:08:26.874 "claim_type": "exclusive_write", 00:08:26.874 "zoned": false, 00:08:26.874 "supported_io_types": { 00:08:26.874 "read": true, 00:08:26.874 "write": true, 00:08:26.874 "unmap": true, 00:08:26.874 "flush": true, 00:08:26.874 "reset": true, 00:08:26.874 "nvme_admin": false, 00:08:26.874 "nvme_io": false, 00:08:26.874 "nvme_io_md": false, 00:08:26.874 "write_zeroes": true, 00:08:26.874 "zcopy": true, 00:08:26.875 "get_zone_info": false, 00:08:26.875 "zone_management": false, 00:08:26.875 "zone_append": false, 00:08:26.875 "compare": false, 00:08:26.875 "compare_and_write": false, 00:08:26.875 "abort": true, 00:08:26.875 "seek_hole": false, 00:08:26.875 "seek_data": false, 00:08:26.875 "copy": true, 00:08:26.875 "nvme_iov_md": false 00:08:26.875 }, 00:08:26.875 "memory_domains": [ 00:08:26.875 { 00:08:26.875 "dma_device_id": "system", 00:08:26.875 "dma_device_type": 1 00:08:26.875 }, 00:08:26.875 { 00:08:26.875 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:26.875 "dma_device_type": 2 00:08:26.875 } 00:08:26.875 ], 00:08:26.875 "driver_specific": {} 00:08:26.875 } 00:08:26.875 ] 00:08:26.875 23:42:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:26.875 23:42:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:08:26.875 23:42:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:08:26.875 23:42:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:26.875 23:42:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid1 0 3 00:08:26.875 23:42:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:26.875 23:42:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:26.875 23:42:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:26.875 23:42:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:26.875 23:42:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:26.875 23:42:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:26.875 23:42:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:26.875 23:42:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:26.875 23:42:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:26.875 23:42:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:26.875 23:42:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:26.875 23:42:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:26.875 23:42:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:26.875 23:42:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:26.875 23:42:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:26.875 "name": "Existed_Raid", 00:08:26.875 "uuid": "32ee297c-d6c2-4630-9515-65c747894357", 00:08:26.875 "strip_size_kb": 0, 00:08:26.875 "state": "online", 00:08:26.875 "raid_level": "raid1", 00:08:26.875 "superblock": true, 00:08:26.875 "num_base_bdevs": 3, 00:08:26.875 "num_base_bdevs_discovered": 3, 00:08:26.875 "num_base_bdevs_operational": 3, 00:08:26.875 "base_bdevs_list": [ 00:08:26.875 { 00:08:26.875 "name": "BaseBdev1", 00:08:26.875 "uuid": "6b1f6948-e827-48a9-baa5-cd1e97fd5603", 00:08:26.875 "is_configured": true, 00:08:26.875 "data_offset": 2048, 00:08:26.875 "data_size": 63488 00:08:26.875 }, 00:08:26.875 { 00:08:26.875 "name": "BaseBdev2", 00:08:26.875 "uuid": "9ffcbe45-b53d-4fe5-ab35-dca86f99beef", 00:08:26.875 "is_configured": true, 00:08:26.875 "data_offset": 2048, 00:08:26.875 "data_size": 63488 00:08:26.875 }, 00:08:26.875 { 00:08:26.875 "name": "BaseBdev3", 00:08:26.875 "uuid": "4984b65e-1e8c-4ce4-99de-3a26dff3e7d2", 00:08:26.875 "is_configured": true, 00:08:26.875 "data_offset": 2048, 00:08:26.875 "data_size": 63488 00:08:26.875 } 00:08:26.875 ] 00:08:26.875 }' 00:08:26.875 23:42:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:26.875 23:42:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:27.445 23:42:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:08:27.445 23:42:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:08:27.445 23:42:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:27.445 23:42:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:27.445 23:42:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:08:27.445 23:42:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:27.445 23:42:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:08:27.445 23:42:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:27.445 23:42:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:27.445 23:42:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:27.445 [2024-11-26 23:42:15.289182] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:27.445 23:42:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:27.445 23:42:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:27.445 "name": "Existed_Raid", 00:08:27.445 "aliases": [ 00:08:27.445 "32ee297c-d6c2-4630-9515-65c747894357" 00:08:27.445 ], 00:08:27.445 "product_name": "Raid Volume", 00:08:27.445 "block_size": 512, 00:08:27.445 "num_blocks": 63488, 00:08:27.445 "uuid": "32ee297c-d6c2-4630-9515-65c747894357", 00:08:27.445 "assigned_rate_limits": { 00:08:27.445 "rw_ios_per_sec": 0, 00:08:27.445 "rw_mbytes_per_sec": 0, 00:08:27.445 "r_mbytes_per_sec": 0, 00:08:27.445 "w_mbytes_per_sec": 0 00:08:27.445 }, 00:08:27.445 "claimed": false, 00:08:27.445 "zoned": false, 00:08:27.445 "supported_io_types": { 00:08:27.445 "read": true, 00:08:27.445 "write": true, 00:08:27.445 "unmap": false, 00:08:27.445 "flush": false, 00:08:27.445 "reset": true, 00:08:27.445 "nvme_admin": false, 00:08:27.445 "nvme_io": false, 00:08:27.445 "nvme_io_md": false, 00:08:27.445 "write_zeroes": true, 00:08:27.445 "zcopy": false, 00:08:27.445 "get_zone_info": false, 00:08:27.445 "zone_management": false, 00:08:27.445 "zone_append": false, 00:08:27.445 "compare": false, 00:08:27.445 "compare_and_write": false, 00:08:27.445 "abort": false, 00:08:27.445 "seek_hole": false, 00:08:27.445 "seek_data": false, 00:08:27.445 "copy": false, 00:08:27.445 "nvme_iov_md": false 00:08:27.445 }, 00:08:27.445 "memory_domains": [ 00:08:27.445 { 00:08:27.445 "dma_device_id": "system", 00:08:27.445 "dma_device_type": 1 00:08:27.445 }, 00:08:27.445 { 00:08:27.445 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:27.445 "dma_device_type": 2 00:08:27.445 }, 00:08:27.445 { 00:08:27.445 "dma_device_id": "system", 00:08:27.445 "dma_device_type": 1 00:08:27.445 }, 00:08:27.445 { 00:08:27.445 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:27.445 "dma_device_type": 2 00:08:27.445 }, 00:08:27.445 { 00:08:27.445 "dma_device_id": "system", 00:08:27.445 "dma_device_type": 1 00:08:27.445 }, 00:08:27.445 { 00:08:27.445 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:27.445 "dma_device_type": 2 00:08:27.445 } 00:08:27.445 ], 00:08:27.445 "driver_specific": { 00:08:27.445 "raid": { 00:08:27.445 "uuid": "32ee297c-d6c2-4630-9515-65c747894357", 00:08:27.445 "strip_size_kb": 0, 00:08:27.445 "state": "online", 00:08:27.445 "raid_level": "raid1", 00:08:27.445 "superblock": true, 00:08:27.445 "num_base_bdevs": 3, 00:08:27.445 "num_base_bdevs_discovered": 3, 00:08:27.445 "num_base_bdevs_operational": 3, 00:08:27.445 "base_bdevs_list": [ 00:08:27.445 { 00:08:27.445 "name": "BaseBdev1", 00:08:27.445 "uuid": "6b1f6948-e827-48a9-baa5-cd1e97fd5603", 00:08:27.445 "is_configured": true, 00:08:27.445 "data_offset": 2048, 00:08:27.445 "data_size": 63488 00:08:27.445 }, 00:08:27.445 { 00:08:27.445 "name": "BaseBdev2", 00:08:27.445 "uuid": "9ffcbe45-b53d-4fe5-ab35-dca86f99beef", 00:08:27.445 "is_configured": true, 00:08:27.445 "data_offset": 2048, 00:08:27.445 "data_size": 63488 00:08:27.445 }, 00:08:27.445 { 00:08:27.445 "name": "BaseBdev3", 00:08:27.445 "uuid": "4984b65e-1e8c-4ce4-99de-3a26dff3e7d2", 00:08:27.445 "is_configured": true, 00:08:27.445 "data_offset": 2048, 00:08:27.445 "data_size": 63488 00:08:27.445 } 00:08:27.445 ] 00:08:27.445 } 00:08:27.445 } 00:08:27.445 }' 00:08:27.445 23:42:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:27.445 23:42:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:08:27.445 BaseBdev2 00:08:27.445 BaseBdev3' 00:08:27.445 23:42:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:27.445 23:42:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:27.445 23:42:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:27.446 23:42:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:08:27.446 23:42:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:27.446 23:42:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:27.446 23:42:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:27.446 23:42:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:27.446 23:42:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:27.446 23:42:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:27.446 23:42:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:27.446 23:42:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:27.446 23:42:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:08:27.446 23:42:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:27.446 23:42:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:27.446 23:42:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:27.446 23:42:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:27.446 23:42:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:27.446 23:42:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:27.446 23:42:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:08:27.446 23:42:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:27.446 23:42:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:27.446 23:42:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:27.446 23:42:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:27.446 23:42:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:27.446 23:42:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:27.446 23:42:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:08:27.446 23:42:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:27.446 23:42:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:27.446 [2024-11-26 23:42:15.568453] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:08:27.704 23:42:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:27.704 23:42:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:08:27.704 23:42:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy raid1 00:08:27.704 23:42:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:08:27.704 23:42:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@199 -- # return 0 00:08:27.704 23:42:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:08:27.704 23:42:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid1 0 2 00:08:27.704 23:42:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:27.704 23:42:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:27.704 23:42:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:27.704 23:42:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:27.704 23:42:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:27.704 23:42:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:27.704 23:42:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:27.704 23:42:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:27.704 23:42:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:27.704 23:42:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:27.704 23:42:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:27.704 23:42:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:27.704 23:42:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:27.704 23:42:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:27.704 23:42:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:27.704 "name": "Existed_Raid", 00:08:27.704 "uuid": "32ee297c-d6c2-4630-9515-65c747894357", 00:08:27.704 "strip_size_kb": 0, 00:08:27.704 "state": "online", 00:08:27.704 "raid_level": "raid1", 00:08:27.704 "superblock": true, 00:08:27.704 "num_base_bdevs": 3, 00:08:27.704 "num_base_bdevs_discovered": 2, 00:08:27.704 "num_base_bdevs_operational": 2, 00:08:27.704 "base_bdevs_list": [ 00:08:27.704 { 00:08:27.704 "name": null, 00:08:27.704 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:27.704 "is_configured": false, 00:08:27.704 "data_offset": 0, 00:08:27.704 "data_size": 63488 00:08:27.704 }, 00:08:27.704 { 00:08:27.704 "name": "BaseBdev2", 00:08:27.704 "uuid": "9ffcbe45-b53d-4fe5-ab35-dca86f99beef", 00:08:27.704 "is_configured": true, 00:08:27.704 "data_offset": 2048, 00:08:27.704 "data_size": 63488 00:08:27.704 }, 00:08:27.704 { 00:08:27.705 "name": "BaseBdev3", 00:08:27.705 "uuid": "4984b65e-1e8c-4ce4-99de-3a26dff3e7d2", 00:08:27.705 "is_configured": true, 00:08:27.705 "data_offset": 2048, 00:08:27.705 "data_size": 63488 00:08:27.705 } 00:08:27.705 ] 00:08:27.705 }' 00:08:27.705 23:42:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:27.705 23:42:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:27.964 23:42:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:08:27.964 23:42:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:27.964 23:42:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:27.964 23:42:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:08:27.964 23:42:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:27.964 23:42:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:27.964 23:42:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:28.223 23:42:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:08:28.223 23:42:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:08:28.223 23:42:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:08:28.223 23:42:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:28.223 23:42:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:28.223 [2024-11-26 23:42:16.112585] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:08:28.223 23:42:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:28.223 23:42:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:08:28.223 23:42:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:28.223 23:42:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:28.224 23:42:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:28.224 23:42:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:28.224 23:42:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:08:28.224 23:42:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:28.224 23:42:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:08:28.224 23:42:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:08:28.224 23:42:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:08:28.224 23:42:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:28.224 23:42:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:28.224 [2024-11-26 23:42:16.189063] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:08:28.224 [2024-11-26 23:42:16.189187] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:28.224 [2024-11-26 23:42:16.210609] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:28.224 [2024-11-26 23:42:16.210786] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:28.224 [2024-11-26 23:42:16.210835] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:08:28.224 23:42:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:28.224 23:42:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:08:28.224 23:42:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:28.224 23:42:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:28.224 23:42:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:08:28.224 23:42:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:28.224 23:42:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:28.224 23:42:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:28.224 23:42:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:08:28.224 23:42:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:08:28.224 23:42:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 3 -gt 2 ']' 00:08:28.224 23:42:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:08:28.224 23:42:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:08:28.224 23:42:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:08:28.224 23:42:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:28.224 23:42:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:28.224 BaseBdev2 00:08:28.224 23:42:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:28.224 23:42:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:08:28.224 23:42:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:08:28.224 23:42:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:28.224 23:42:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:08:28.224 23:42:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:28.224 23:42:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:28.224 23:42:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:28.224 23:42:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:28.224 23:42:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:28.224 23:42:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:28.224 23:42:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:08:28.224 23:42:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:28.224 23:42:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:28.224 [ 00:08:28.224 { 00:08:28.224 "name": "BaseBdev2", 00:08:28.224 "aliases": [ 00:08:28.224 "443b2228-73b9-4114-9377-84abbf1ddc37" 00:08:28.224 ], 00:08:28.224 "product_name": "Malloc disk", 00:08:28.224 "block_size": 512, 00:08:28.224 "num_blocks": 65536, 00:08:28.224 "uuid": "443b2228-73b9-4114-9377-84abbf1ddc37", 00:08:28.224 "assigned_rate_limits": { 00:08:28.224 "rw_ios_per_sec": 0, 00:08:28.224 "rw_mbytes_per_sec": 0, 00:08:28.224 "r_mbytes_per_sec": 0, 00:08:28.224 "w_mbytes_per_sec": 0 00:08:28.224 }, 00:08:28.224 "claimed": false, 00:08:28.224 "zoned": false, 00:08:28.224 "supported_io_types": { 00:08:28.224 "read": true, 00:08:28.224 "write": true, 00:08:28.224 "unmap": true, 00:08:28.224 "flush": true, 00:08:28.224 "reset": true, 00:08:28.224 "nvme_admin": false, 00:08:28.224 "nvme_io": false, 00:08:28.224 "nvme_io_md": false, 00:08:28.224 "write_zeroes": true, 00:08:28.224 "zcopy": true, 00:08:28.224 "get_zone_info": false, 00:08:28.224 "zone_management": false, 00:08:28.224 "zone_append": false, 00:08:28.224 "compare": false, 00:08:28.224 "compare_and_write": false, 00:08:28.224 "abort": true, 00:08:28.224 "seek_hole": false, 00:08:28.224 "seek_data": false, 00:08:28.224 "copy": true, 00:08:28.224 "nvme_iov_md": false 00:08:28.224 }, 00:08:28.224 "memory_domains": [ 00:08:28.224 { 00:08:28.224 "dma_device_id": "system", 00:08:28.224 "dma_device_type": 1 00:08:28.224 }, 00:08:28.224 { 00:08:28.224 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:28.224 "dma_device_type": 2 00:08:28.224 } 00:08:28.224 ], 00:08:28.224 "driver_specific": {} 00:08:28.224 } 00:08:28.224 ] 00:08:28.224 23:42:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:28.224 23:42:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:08:28.224 23:42:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:08:28.224 23:42:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:08:28.224 23:42:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:08:28.224 23:42:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:28.224 23:42:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:28.224 BaseBdev3 00:08:28.224 23:42:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:28.224 23:42:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:08:28.224 23:42:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:08:28.224 23:42:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:28.224 23:42:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:08:28.224 23:42:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:28.224 23:42:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:28.224 23:42:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:28.224 23:42:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:28.224 23:42:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:28.483 23:42:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:28.483 23:42:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:08:28.483 23:42:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:28.483 23:42:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:28.483 [ 00:08:28.483 { 00:08:28.483 "name": "BaseBdev3", 00:08:28.483 "aliases": [ 00:08:28.483 "32edf3b8-1a10-4bc7-b11f-b064d7d5dc8c" 00:08:28.483 ], 00:08:28.483 "product_name": "Malloc disk", 00:08:28.483 "block_size": 512, 00:08:28.483 "num_blocks": 65536, 00:08:28.483 "uuid": "32edf3b8-1a10-4bc7-b11f-b064d7d5dc8c", 00:08:28.483 "assigned_rate_limits": { 00:08:28.483 "rw_ios_per_sec": 0, 00:08:28.483 "rw_mbytes_per_sec": 0, 00:08:28.483 "r_mbytes_per_sec": 0, 00:08:28.483 "w_mbytes_per_sec": 0 00:08:28.483 }, 00:08:28.483 "claimed": false, 00:08:28.483 "zoned": false, 00:08:28.483 "supported_io_types": { 00:08:28.483 "read": true, 00:08:28.483 "write": true, 00:08:28.483 "unmap": true, 00:08:28.483 "flush": true, 00:08:28.483 "reset": true, 00:08:28.483 "nvme_admin": false, 00:08:28.483 "nvme_io": false, 00:08:28.483 "nvme_io_md": false, 00:08:28.483 "write_zeroes": true, 00:08:28.483 "zcopy": true, 00:08:28.483 "get_zone_info": false, 00:08:28.483 "zone_management": false, 00:08:28.483 "zone_append": false, 00:08:28.483 "compare": false, 00:08:28.483 "compare_and_write": false, 00:08:28.483 "abort": true, 00:08:28.483 "seek_hole": false, 00:08:28.483 "seek_data": false, 00:08:28.483 "copy": true, 00:08:28.483 "nvme_iov_md": false 00:08:28.483 }, 00:08:28.483 "memory_domains": [ 00:08:28.483 { 00:08:28.483 "dma_device_id": "system", 00:08:28.483 "dma_device_type": 1 00:08:28.483 }, 00:08:28.483 { 00:08:28.483 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:28.483 "dma_device_type": 2 00:08:28.483 } 00:08:28.483 ], 00:08:28.483 "driver_specific": {} 00:08:28.483 } 00:08:28.483 ] 00:08:28.484 23:42:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:28.484 23:42:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:08:28.484 23:42:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:08:28.484 23:42:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:08:28.484 23:42:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:28.484 23:42:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:28.484 23:42:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:28.484 [2024-11-26 23:42:16.392388] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:28.484 [2024-11-26 23:42:16.392535] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:28.484 [2024-11-26 23:42:16.392585] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:28.484 [2024-11-26 23:42:16.394836] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:28.484 23:42:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:28.484 23:42:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:08:28.484 23:42:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:28.484 23:42:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:28.484 23:42:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:28.484 23:42:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:28.484 23:42:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:28.484 23:42:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:28.484 23:42:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:28.484 23:42:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:28.484 23:42:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:28.484 23:42:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:28.484 23:42:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:28.484 23:42:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:28.484 23:42:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:28.484 23:42:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:28.484 23:42:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:28.484 "name": "Existed_Raid", 00:08:28.484 "uuid": "9e5ce50c-e993-4565-b864-41a88d7e531e", 00:08:28.484 "strip_size_kb": 0, 00:08:28.484 "state": "configuring", 00:08:28.484 "raid_level": "raid1", 00:08:28.484 "superblock": true, 00:08:28.484 "num_base_bdevs": 3, 00:08:28.484 "num_base_bdevs_discovered": 2, 00:08:28.484 "num_base_bdevs_operational": 3, 00:08:28.484 "base_bdevs_list": [ 00:08:28.484 { 00:08:28.484 "name": "BaseBdev1", 00:08:28.484 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:28.484 "is_configured": false, 00:08:28.484 "data_offset": 0, 00:08:28.484 "data_size": 0 00:08:28.484 }, 00:08:28.484 { 00:08:28.484 "name": "BaseBdev2", 00:08:28.484 "uuid": "443b2228-73b9-4114-9377-84abbf1ddc37", 00:08:28.484 "is_configured": true, 00:08:28.484 "data_offset": 2048, 00:08:28.484 "data_size": 63488 00:08:28.484 }, 00:08:28.484 { 00:08:28.484 "name": "BaseBdev3", 00:08:28.484 "uuid": "32edf3b8-1a10-4bc7-b11f-b064d7d5dc8c", 00:08:28.484 "is_configured": true, 00:08:28.484 "data_offset": 2048, 00:08:28.484 "data_size": 63488 00:08:28.484 } 00:08:28.484 ] 00:08:28.484 }' 00:08:28.484 23:42:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:28.484 23:42:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:28.744 23:42:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:08:28.744 23:42:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:28.744 23:42:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:28.744 [2024-11-26 23:42:16.831652] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:08:28.744 23:42:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:28.744 23:42:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:08:28.744 23:42:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:28.744 23:42:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:28.744 23:42:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:28.744 23:42:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:28.744 23:42:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:28.744 23:42:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:28.744 23:42:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:28.744 23:42:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:28.744 23:42:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:28.744 23:42:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:28.744 23:42:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:28.744 23:42:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:28.744 23:42:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:28.744 23:42:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:28.744 23:42:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:28.744 "name": "Existed_Raid", 00:08:28.744 "uuid": "9e5ce50c-e993-4565-b864-41a88d7e531e", 00:08:28.744 "strip_size_kb": 0, 00:08:28.744 "state": "configuring", 00:08:28.744 "raid_level": "raid1", 00:08:28.744 "superblock": true, 00:08:28.744 "num_base_bdevs": 3, 00:08:28.744 "num_base_bdevs_discovered": 1, 00:08:28.744 "num_base_bdevs_operational": 3, 00:08:28.744 "base_bdevs_list": [ 00:08:28.744 { 00:08:28.744 "name": "BaseBdev1", 00:08:28.744 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:28.744 "is_configured": false, 00:08:28.744 "data_offset": 0, 00:08:28.744 "data_size": 0 00:08:28.744 }, 00:08:28.744 { 00:08:28.744 "name": null, 00:08:28.744 "uuid": "443b2228-73b9-4114-9377-84abbf1ddc37", 00:08:28.744 "is_configured": false, 00:08:28.744 "data_offset": 0, 00:08:28.744 "data_size": 63488 00:08:28.744 }, 00:08:28.744 { 00:08:28.744 "name": "BaseBdev3", 00:08:28.744 "uuid": "32edf3b8-1a10-4bc7-b11f-b064d7d5dc8c", 00:08:28.744 "is_configured": true, 00:08:28.744 "data_offset": 2048, 00:08:28.744 "data_size": 63488 00:08:28.744 } 00:08:28.744 ] 00:08:28.744 }' 00:08:29.002 23:42:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:29.002 23:42:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:29.262 23:42:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:29.262 23:42:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:08:29.262 23:42:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:29.262 23:42:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:29.262 23:42:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:29.262 23:42:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:08:29.262 23:42:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:08:29.262 23:42:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:29.262 23:42:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:29.262 [2024-11-26 23:42:17.339724] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:29.262 BaseBdev1 00:08:29.262 23:42:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:29.262 23:42:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:08:29.262 23:42:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:08:29.262 23:42:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:29.262 23:42:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:08:29.262 23:42:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:29.262 23:42:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:29.262 23:42:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:29.262 23:42:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:29.262 23:42:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:29.262 23:42:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:29.262 23:42:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:08:29.262 23:42:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:29.262 23:42:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:29.262 [ 00:08:29.262 { 00:08:29.262 "name": "BaseBdev1", 00:08:29.262 "aliases": [ 00:08:29.262 "2fc65128-e67b-4d22-a394-5518c8ba4121" 00:08:29.262 ], 00:08:29.262 "product_name": "Malloc disk", 00:08:29.262 "block_size": 512, 00:08:29.262 "num_blocks": 65536, 00:08:29.262 "uuid": "2fc65128-e67b-4d22-a394-5518c8ba4121", 00:08:29.262 "assigned_rate_limits": { 00:08:29.262 "rw_ios_per_sec": 0, 00:08:29.262 "rw_mbytes_per_sec": 0, 00:08:29.262 "r_mbytes_per_sec": 0, 00:08:29.262 "w_mbytes_per_sec": 0 00:08:29.262 }, 00:08:29.262 "claimed": true, 00:08:29.262 "claim_type": "exclusive_write", 00:08:29.262 "zoned": false, 00:08:29.262 "supported_io_types": { 00:08:29.262 "read": true, 00:08:29.262 "write": true, 00:08:29.262 "unmap": true, 00:08:29.262 "flush": true, 00:08:29.262 "reset": true, 00:08:29.262 "nvme_admin": false, 00:08:29.262 "nvme_io": false, 00:08:29.262 "nvme_io_md": false, 00:08:29.262 "write_zeroes": true, 00:08:29.262 "zcopy": true, 00:08:29.262 "get_zone_info": false, 00:08:29.262 "zone_management": false, 00:08:29.262 "zone_append": false, 00:08:29.262 "compare": false, 00:08:29.262 "compare_and_write": false, 00:08:29.262 "abort": true, 00:08:29.262 "seek_hole": false, 00:08:29.262 "seek_data": false, 00:08:29.262 "copy": true, 00:08:29.262 "nvme_iov_md": false 00:08:29.262 }, 00:08:29.262 "memory_domains": [ 00:08:29.262 { 00:08:29.262 "dma_device_id": "system", 00:08:29.262 "dma_device_type": 1 00:08:29.262 }, 00:08:29.262 { 00:08:29.262 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:29.262 "dma_device_type": 2 00:08:29.262 } 00:08:29.262 ], 00:08:29.262 "driver_specific": {} 00:08:29.262 } 00:08:29.262 ] 00:08:29.262 23:42:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:29.262 23:42:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:08:29.262 23:42:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:08:29.262 23:42:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:29.262 23:42:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:29.262 23:42:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:29.262 23:42:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:29.262 23:42:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:29.262 23:42:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:29.262 23:42:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:29.262 23:42:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:29.262 23:42:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:29.262 23:42:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:29.262 23:42:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:29.262 23:42:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:29.262 23:42:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:29.521 23:42:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:29.521 23:42:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:29.521 "name": "Existed_Raid", 00:08:29.521 "uuid": "9e5ce50c-e993-4565-b864-41a88d7e531e", 00:08:29.521 "strip_size_kb": 0, 00:08:29.521 "state": "configuring", 00:08:29.521 "raid_level": "raid1", 00:08:29.521 "superblock": true, 00:08:29.521 "num_base_bdevs": 3, 00:08:29.521 "num_base_bdevs_discovered": 2, 00:08:29.521 "num_base_bdevs_operational": 3, 00:08:29.521 "base_bdevs_list": [ 00:08:29.521 { 00:08:29.521 "name": "BaseBdev1", 00:08:29.521 "uuid": "2fc65128-e67b-4d22-a394-5518c8ba4121", 00:08:29.521 "is_configured": true, 00:08:29.521 "data_offset": 2048, 00:08:29.521 "data_size": 63488 00:08:29.521 }, 00:08:29.521 { 00:08:29.521 "name": null, 00:08:29.521 "uuid": "443b2228-73b9-4114-9377-84abbf1ddc37", 00:08:29.521 "is_configured": false, 00:08:29.521 "data_offset": 0, 00:08:29.521 "data_size": 63488 00:08:29.521 }, 00:08:29.521 { 00:08:29.521 "name": "BaseBdev3", 00:08:29.521 "uuid": "32edf3b8-1a10-4bc7-b11f-b064d7d5dc8c", 00:08:29.521 "is_configured": true, 00:08:29.521 "data_offset": 2048, 00:08:29.521 "data_size": 63488 00:08:29.521 } 00:08:29.521 ] 00:08:29.521 }' 00:08:29.521 23:42:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:29.521 23:42:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:29.781 23:42:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:29.781 23:42:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:08:29.781 23:42:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:29.781 23:42:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:29.781 23:42:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:29.781 23:42:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:08:29.781 23:42:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:08:29.781 23:42:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:29.781 23:42:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:29.781 [2024-11-26 23:42:17.827040] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:08:29.781 23:42:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:29.781 23:42:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:08:29.781 23:42:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:29.781 23:42:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:29.781 23:42:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:29.781 23:42:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:29.781 23:42:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:29.781 23:42:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:29.781 23:42:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:29.781 23:42:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:29.781 23:42:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:29.781 23:42:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:29.781 23:42:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:29.781 23:42:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:29.781 23:42:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:29.781 23:42:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:29.781 23:42:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:29.781 "name": "Existed_Raid", 00:08:29.781 "uuid": "9e5ce50c-e993-4565-b864-41a88d7e531e", 00:08:29.781 "strip_size_kb": 0, 00:08:29.781 "state": "configuring", 00:08:29.781 "raid_level": "raid1", 00:08:29.781 "superblock": true, 00:08:29.781 "num_base_bdevs": 3, 00:08:29.781 "num_base_bdevs_discovered": 1, 00:08:29.781 "num_base_bdevs_operational": 3, 00:08:29.781 "base_bdevs_list": [ 00:08:29.781 { 00:08:29.781 "name": "BaseBdev1", 00:08:29.781 "uuid": "2fc65128-e67b-4d22-a394-5518c8ba4121", 00:08:29.781 "is_configured": true, 00:08:29.781 "data_offset": 2048, 00:08:29.781 "data_size": 63488 00:08:29.781 }, 00:08:29.781 { 00:08:29.781 "name": null, 00:08:29.781 "uuid": "443b2228-73b9-4114-9377-84abbf1ddc37", 00:08:29.781 "is_configured": false, 00:08:29.781 "data_offset": 0, 00:08:29.781 "data_size": 63488 00:08:29.781 }, 00:08:29.781 { 00:08:29.781 "name": null, 00:08:29.781 "uuid": "32edf3b8-1a10-4bc7-b11f-b064d7d5dc8c", 00:08:29.781 "is_configured": false, 00:08:29.781 "data_offset": 0, 00:08:29.781 "data_size": 63488 00:08:29.781 } 00:08:29.781 ] 00:08:29.781 }' 00:08:29.781 23:42:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:29.781 23:42:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:30.348 23:42:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:08:30.348 23:42:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:30.348 23:42:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:30.348 23:42:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:30.348 23:42:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:30.348 23:42:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:08:30.348 23:42:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:08:30.348 23:42:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:30.348 23:42:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:30.348 [2024-11-26 23:42:18.322319] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:30.349 23:42:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:30.349 23:42:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:08:30.349 23:42:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:30.349 23:42:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:30.349 23:42:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:30.349 23:42:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:30.349 23:42:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:30.349 23:42:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:30.349 23:42:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:30.349 23:42:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:30.349 23:42:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:30.349 23:42:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:30.349 23:42:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:30.349 23:42:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:30.349 23:42:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:30.349 23:42:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:30.349 23:42:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:30.349 "name": "Existed_Raid", 00:08:30.349 "uuid": "9e5ce50c-e993-4565-b864-41a88d7e531e", 00:08:30.349 "strip_size_kb": 0, 00:08:30.349 "state": "configuring", 00:08:30.349 "raid_level": "raid1", 00:08:30.349 "superblock": true, 00:08:30.349 "num_base_bdevs": 3, 00:08:30.349 "num_base_bdevs_discovered": 2, 00:08:30.349 "num_base_bdevs_operational": 3, 00:08:30.349 "base_bdevs_list": [ 00:08:30.349 { 00:08:30.349 "name": "BaseBdev1", 00:08:30.349 "uuid": "2fc65128-e67b-4d22-a394-5518c8ba4121", 00:08:30.349 "is_configured": true, 00:08:30.349 "data_offset": 2048, 00:08:30.349 "data_size": 63488 00:08:30.349 }, 00:08:30.349 { 00:08:30.349 "name": null, 00:08:30.349 "uuid": "443b2228-73b9-4114-9377-84abbf1ddc37", 00:08:30.349 "is_configured": false, 00:08:30.349 "data_offset": 0, 00:08:30.349 "data_size": 63488 00:08:30.349 }, 00:08:30.349 { 00:08:30.349 "name": "BaseBdev3", 00:08:30.349 "uuid": "32edf3b8-1a10-4bc7-b11f-b064d7d5dc8c", 00:08:30.349 "is_configured": true, 00:08:30.349 "data_offset": 2048, 00:08:30.349 "data_size": 63488 00:08:30.349 } 00:08:30.349 ] 00:08:30.349 }' 00:08:30.349 23:42:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:30.349 23:42:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:30.916 23:42:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:30.916 23:42:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:08:30.916 23:42:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:30.916 23:42:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:30.916 23:42:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:30.916 23:42:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:08:30.916 23:42:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:08:30.916 23:42:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:30.916 23:42:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:30.916 [2024-11-26 23:42:18.809680] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:08:30.916 23:42:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:30.916 23:42:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:08:30.916 23:42:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:30.916 23:42:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:30.916 23:42:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:30.916 23:42:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:30.916 23:42:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:30.916 23:42:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:30.916 23:42:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:30.916 23:42:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:30.916 23:42:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:30.916 23:42:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:30.916 23:42:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:30.916 23:42:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:30.916 23:42:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:30.916 23:42:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:30.916 23:42:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:30.916 "name": "Existed_Raid", 00:08:30.916 "uuid": "9e5ce50c-e993-4565-b864-41a88d7e531e", 00:08:30.916 "strip_size_kb": 0, 00:08:30.916 "state": "configuring", 00:08:30.916 "raid_level": "raid1", 00:08:30.916 "superblock": true, 00:08:30.916 "num_base_bdevs": 3, 00:08:30.916 "num_base_bdevs_discovered": 1, 00:08:30.916 "num_base_bdevs_operational": 3, 00:08:30.916 "base_bdevs_list": [ 00:08:30.916 { 00:08:30.916 "name": null, 00:08:30.916 "uuid": "2fc65128-e67b-4d22-a394-5518c8ba4121", 00:08:30.916 "is_configured": false, 00:08:30.916 "data_offset": 0, 00:08:30.916 "data_size": 63488 00:08:30.916 }, 00:08:30.916 { 00:08:30.916 "name": null, 00:08:30.916 "uuid": "443b2228-73b9-4114-9377-84abbf1ddc37", 00:08:30.916 "is_configured": false, 00:08:30.916 "data_offset": 0, 00:08:30.916 "data_size": 63488 00:08:30.916 }, 00:08:30.916 { 00:08:30.916 "name": "BaseBdev3", 00:08:30.916 "uuid": "32edf3b8-1a10-4bc7-b11f-b064d7d5dc8c", 00:08:30.916 "is_configured": true, 00:08:30.916 "data_offset": 2048, 00:08:30.916 "data_size": 63488 00:08:30.916 } 00:08:30.916 ] 00:08:30.916 }' 00:08:30.916 23:42:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:30.916 23:42:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:31.175 23:42:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:08:31.175 23:42:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:31.175 23:42:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:31.175 23:42:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:31.175 23:42:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:31.175 23:42:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:08:31.175 23:42:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:08:31.175 23:42:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:31.175 23:42:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:31.175 [2024-11-26 23:42:19.249184] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:31.175 23:42:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:31.175 23:42:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:08:31.175 23:42:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:31.175 23:42:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:31.175 23:42:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:31.175 23:42:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:31.175 23:42:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:31.175 23:42:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:31.175 23:42:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:31.175 23:42:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:31.175 23:42:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:31.175 23:42:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:31.175 23:42:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:31.175 23:42:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:31.175 23:42:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:31.175 23:42:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:31.433 23:42:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:31.433 "name": "Existed_Raid", 00:08:31.433 "uuid": "9e5ce50c-e993-4565-b864-41a88d7e531e", 00:08:31.433 "strip_size_kb": 0, 00:08:31.433 "state": "configuring", 00:08:31.433 "raid_level": "raid1", 00:08:31.433 "superblock": true, 00:08:31.433 "num_base_bdevs": 3, 00:08:31.433 "num_base_bdevs_discovered": 2, 00:08:31.433 "num_base_bdevs_operational": 3, 00:08:31.433 "base_bdevs_list": [ 00:08:31.433 { 00:08:31.433 "name": null, 00:08:31.433 "uuid": "2fc65128-e67b-4d22-a394-5518c8ba4121", 00:08:31.433 "is_configured": false, 00:08:31.433 "data_offset": 0, 00:08:31.433 "data_size": 63488 00:08:31.433 }, 00:08:31.433 { 00:08:31.433 "name": "BaseBdev2", 00:08:31.433 "uuid": "443b2228-73b9-4114-9377-84abbf1ddc37", 00:08:31.433 "is_configured": true, 00:08:31.433 "data_offset": 2048, 00:08:31.433 "data_size": 63488 00:08:31.433 }, 00:08:31.433 { 00:08:31.433 "name": "BaseBdev3", 00:08:31.433 "uuid": "32edf3b8-1a10-4bc7-b11f-b064d7d5dc8c", 00:08:31.433 "is_configured": true, 00:08:31.433 "data_offset": 2048, 00:08:31.433 "data_size": 63488 00:08:31.433 } 00:08:31.433 ] 00:08:31.433 }' 00:08:31.433 23:42:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:31.433 23:42:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:31.700 23:42:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:31.700 23:42:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:08:31.700 23:42:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:31.700 23:42:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:31.700 23:42:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:31.700 23:42:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:08:31.700 23:42:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:08:31.700 23:42:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:31.700 23:42:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:31.700 23:42:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:31.700 23:42:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:31.700 23:42:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 2fc65128-e67b-4d22-a394-5518c8ba4121 00:08:31.700 23:42:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:31.700 23:42:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:31.700 [2024-11-26 23:42:19.812958] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:08:31.700 [2024-11-26 23:42:19.813156] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:08:31.700 [2024-11-26 23:42:19.813170] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:08:31.700 [2024-11-26 23:42:19.813479] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002870 00:08:31.700 NewBaseBdev 00:08:31.700 [2024-11-26 23:42:19.813613] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:08:31.700 [2024-11-26 23:42:19.813737] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001c80 00:08:31.700 [2024-11-26 23:42:19.813913] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:31.700 23:42:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:31.700 23:42:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:08:31.700 23:42:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:08:31.700 23:42:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:31.700 23:42:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:08:31.700 23:42:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:31.700 23:42:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:31.700 23:42:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:31.700 23:42:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:31.700 23:42:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:31.700 23:42:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:31.701 23:42:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:08:31.701 23:42:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:31.701 23:42:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:31.963 [ 00:08:31.963 { 00:08:31.963 "name": "NewBaseBdev", 00:08:31.963 "aliases": [ 00:08:31.963 "2fc65128-e67b-4d22-a394-5518c8ba4121" 00:08:31.963 ], 00:08:31.963 "product_name": "Malloc disk", 00:08:31.963 "block_size": 512, 00:08:31.963 "num_blocks": 65536, 00:08:31.963 "uuid": "2fc65128-e67b-4d22-a394-5518c8ba4121", 00:08:31.963 "assigned_rate_limits": { 00:08:31.963 "rw_ios_per_sec": 0, 00:08:31.963 "rw_mbytes_per_sec": 0, 00:08:31.963 "r_mbytes_per_sec": 0, 00:08:31.963 "w_mbytes_per_sec": 0 00:08:31.963 }, 00:08:31.963 "claimed": true, 00:08:31.963 "claim_type": "exclusive_write", 00:08:31.963 "zoned": false, 00:08:31.963 "supported_io_types": { 00:08:31.963 "read": true, 00:08:31.963 "write": true, 00:08:31.963 "unmap": true, 00:08:31.963 "flush": true, 00:08:31.963 "reset": true, 00:08:31.963 "nvme_admin": false, 00:08:31.963 "nvme_io": false, 00:08:31.963 "nvme_io_md": false, 00:08:31.963 "write_zeroes": true, 00:08:31.963 "zcopy": true, 00:08:31.963 "get_zone_info": false, 00:08:31.963 "zone_management": false, 00:08:31.963 "zone_append": false, 00:08:31.963 "compare": false, 00:08:31.963 "compare_and_write": false, 00:08:31.963 "abort": true, 00:08:31.963 "seek_hole": false, 00:08:31.963 "seek_data": false, 00:08:31.963 "copy": true, 00:08:31.963 "nvme_iov_md": false 00:08:31.963 }, 00:08:31.963 "memory_domains": [ 00:08:31.963 { 00:08:31.963 "dma_device_id": "system", 00:08:31.963 "dma_device_type": 1 00:08:31.963 }, 00:08:31.963 { 00:08:31.963 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:31.963 "dma_device_type": 2 00:08:31.963 } 00:08:31.963 ], 00:08:31.963 "driver_specific": {} 00:08:31.963 } 00:08:31.963 ] 00:08:31.963 23:42:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:31.963 23:42:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:08:31.963 23:42:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid1 0 3 00:08:31.963 23:42:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:31.963 23:42:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:31.963 23:42:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:31.963 23:42:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:31.963 23:42:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:31.963 23:42:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:31.963 23:42:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:31.963 23:42:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:31.963 23:42:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:31.963 23:42:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:31.963 23:42:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:31.963 23:42:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:31.963 23:42:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:31.963 23:42:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:31.963 23:42:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:31.963 "name": "Existed_Raid", 00:08:31.963 "uuid": "9e5ce50c-e993-4565-b864-41a88d7e531e", 00:08:31.963 "strip_size_kb": 0, 00:08:31.963 "state": "online", 00:08:31.963 "raid_level": "raid1", 00:08:31.963 "superblock": true, 00:08:31.963 "num_base_bdevs": 3, 00:08:31.963 "num_base_bdevs_discovered": 3, 00:08:31.963 "num_base_bdevs_operational": 3, 00:08:31.963 "base_bdevs_list": [ 00:08:31.963 { 00:08:31.963 "name": "NewBaseBdev", 00:08:31.963 "uuid": "2fc65128-e67b-4d22-a394-5518c8ba4121", 00:08:31.963 "is_configured": true, 00:08:31.963 "data_offset": 2048, 00:08:31.963 "data_size": 63488 00:08:31.963 }, 00:08:31.963 { 00:08:31.963 "name": "BaseBdev2", 00:08:31.963 "uuid": "443b2228-73b9-4114-9377-84abbf1ddc37", 00:08:31.963 "is_configured": true, 00:08:31.963 "data_offset": 2048, 00:08:31.963 "data_size": 63488 00:08:31.963 }, 00:08:31.963 { 00:08:31.963 "name": "BaseBdev3", 00:08:31.963 "uuid": "32edf3b8-1a10-4bc7-b11f-b064d7d5dc8c", 00:08:31.963 "is_configured": true, 00:08:31.963 "data_offset": 2048, 00:08:31.963 "data_size": 63488 00:08:31.963 } 00:08:31.963 ] 00:08:31.963 }' 00:08:31.963 23:42:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:31.963 23:42:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:32.224 23:42:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:08:32.224 23:42:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:08:32.224 23:42:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:32.224 23:42:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:32.224 23:42:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:08:32.224 23:42:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:32.224 23:42:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:08:32.224 23:42:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:32.224 23:42:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:32.224 23:42:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:32.224 [2024-11-26 23:42:20.272690] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:32.224 23:42:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:32.224 23:42:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:32.224 "name": "Existed_Raid", 00:08:32.224 "aliases": [ 00:08:32.224 "9e5ce50c-e993-4565-b864-41a88d7e531e" 00:08:32.224 ], 00:08:32.224 "product_name": "Raid Volume", 00:08:32.224 "block_size": 512, 00:08:32.224 "num_blocks": 63488, 00:08:32.224 "uuid": "9e5ce50c-e993-4565-b864-41a88d7e531e", 00:08:32.224 "assigned_rate_limits": { 00:08:32.224 "rw_ios_per_sec": 0, 00:08:32.224 "rw_mbytes_per_sec": 0, 00:08:32.224 "r_mbytes_per_sec": 0, 00:08:32.224 "w_mbytes_per_sec": 0 00:08:32.224 }, 00:08:32.224 "claimed": false, 00:08:32.224 "zoned": false, 00:08:32.224 "supported_io_types": { 00:08:32.224 "read": true, 00:08:32.224 "write": true, 00:08:32.224 "unmap": false, 00:08:32.224 "flush": false, 00:08:32.224 "reset": true, 00:08:32.224 "nvme_admin": false, 00:08:32.224 "nvme_io": false, 00:08:32.224 "nvme_io_md": false, 00:08:32.224 "write_zeroes": true, 00:08:32.224 "zcopy": false, 00:08:32.224 "get_zone_info": false, 00:08:32.224 "zone_management": false, 00:08:32.224 "zone_append": false, 00:08:32.224 "compare": false, 00:08:32.224 "compare_and_write": false, 00:08:32.224 "abort": false, 00:08:32.224 "seek_hole": false, 00:08:32.224 "seek_data": false, 00:08:32.224 "copy": false, 00:08:32.224 "nvme_iov_md": false 00:08:32.224 }, 00:08:32.224 "memory_domains": [ 00:08:32.224 { 00:08:32.224 "dma_device_id": "system", 00:08:32.224 "dma_device_type": 1 00:08:32.224 }, 00:08:32.224 { 00:08:32.224 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:32.224 "dma_device_type": 2 00:08:32.224 }, 00:08:32.224 { 00:08:32.224 "dma_device_id": "system", 00:08:32.224 "dma_device_type": 1 00:08:32.224 }, 00:08:32.224 { 00:08:32.224 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:32.224 "dma_device_type": 2 00:08:32.224 }, 00:08:32.224 { 00:08:32.224 "dma_device_id": "system", 00:08:32.224 "dma_device_type": 1 00:08:32.224 }, 00:08:32.224 { 00:08:32.224 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:32.224 "dma_device_type": 2 00:08:32.224 } 00:08:32.224 ], 00:08:32.224 "driver_specific": { 00:08:32.224 "raid": { 00:08:32.224 "uuid": "9e5ce50c-e993-4565-b864-41a88d7e531e", 00:08:32.224 "strip_size_kb": 0, 00:08:32.224 "state": "online", 00:08:32.224 "raid_level": "raid1", 00:08:32.224 "superblock": true, 00:08:32.224 "num_base_bdevs": 3, 00:08:32.224 "num_base_bdevs_discovered": 3, 00:08:32.224 "num_base_bdevs_operational": 3, 00:08:32.224 "base_bdevs_list": [ 00:08:32.224 { 00:08:32.224 "name": "NewBaseBdev", 00:08:32.224 "uuid": "2fc65128-e67b-4d22-a394-5518c8ba4121", 00:08:32.224 "is_configured": true, 00:08:32.224 "data_offset": 2048, 00:08:32.224 "data_size": 63488 00:08:32.224 }, 00:08:32.224 { 00:08:32.224 "name": "BaseBdev2", 00:08:32.224 "uuid": "443b2228-73b9-4114-9377-84abbf1ddc37", 00:08:32.224 "is_configured": true, 00:08:32.224 "data_offset": 2048, 00:08:32.224 "data_size": 63488 00:08:32.224 }, 00:08:32.224 { 00:08:32.224 "name": "BaseBdev3", 00:08:32.224 "uuid": "32edf3b8-1a10-4bc7-b11f-b064d7d5dc8c", 00:08:32.224 "is_configured": true, 00:08:32.224 "data_offset": 2048, 00:08:32.224 "data_size": 63488 00:08:32.224 } 00:08:32.224 ] 00:08:32.224 } 00:08:32.224 } 00:08:32.224 }' 00:08:32.224 23:42:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:32.224 23:42:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:08:32.224 BaseBdev2 00:08:32.224 BaseBdev3' 00:08:32.224 23:42:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:32.485 23:42:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:32.485 23:42:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:32.485 23:42:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:08:32.485 23:42:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:32.485 23:42:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:32.485 23:42:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:32.485 23:42:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:32.485 23:42:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:32.485 23:42:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:32.485 23:42:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:32.485 23:42:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:32.485 23:42:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:08:32.485 23:42:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:32.485 23:42:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:32.485 23:42:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:32.485 23:42:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:32.485 23:42:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:32.485 23:42:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:32.485 23:42:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:08:32.485 23:42:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:32.485 23:42:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:32.485 23:42:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:32.485 23:42:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:32.485 23:42:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:32.485 23:42:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:32.485 23:42:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:32.485 23:42:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:32.485 23:42:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:32.485 [2024-11-26 23:42:20.531808] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:32.485 [2024-11-26 23:42:20.531881] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:32.485 [2024-11-26 23:42:20.531976] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:32.485 [2024-11-26 23:42:20.532254] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:32.485 [2024-11-26 23:42:20.532266] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name Existed_Raid, state offline 00:08:32.485 23:42:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:32.485 23:42:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 78766 00:08:32.485 23:42:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # '[' -z 78766 ']' 00:08:32.485 23:42:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@958 -- # kill -0 78766 00:08:32.485 23:42:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # uname 00:08:32.485 23:42:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:08:32.485 23:42:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 78766 00:08:32.485 23:42:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:08:32.485 23:42:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:08:32.485 23:42:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@972 -- # echo 'killing process with pid 78766' 00:08:32.485 killing process with pid 78766 00:08:32.485 23:42:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@973 -- # kill 78766 00:08:32.485 [2024-11-26 23:42:20.583031] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:32.485 23:42:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@978 -- # wait 78766 00:08:32.745 [2024-11-26 23:42:20.615255] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:32.745 23:42:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:08:32.745 00:08:32.745 real 0m8.951s 00:08:32.745 user 0m15.253s 00:08:32.745 sys 0m1.832s 00:08:32.745 23:42:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1130 -- # xtrace_disable 00:08:32.745 23:42:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:32.745 ************************************ 00:08:32.745 END TEST raid_state_function_test_sb 00:08:32.745 ************************************ 00:08:33.016 23:42:20 bdev_raid -- bdev/bdev_raid.sh@970 -- # run_test raid_superblock_test raid_superblock_test raid1 3 00:08:33.016 23:42:20 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:08:33.016 23:42:20 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:08:33.016 23:42:20 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:33.016 ************************************ 00:08:33.016 START TEST raid_superblock_test 00:08:33.016 ************************************ 00:08:33.016 23:42:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1129 -- # raid_superblock_test raid1 3 00:08:33.016 23:42:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=raid1 00:08:33.016 23:42:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=3 00:08:33.017 23:42:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:08:33.017 23:42:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:08:33.017 23:42:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:08:33.017 23:42:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:08:33.017 23:42:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:08:33.017 23:42:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:08:33.017 23:42:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:08:33.017 23:42:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:08:33.017 23:42:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:08:33.017 23:42:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:08:33.017 23:42:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:08:33.017 23:42:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' raid1 '!=' raid1 ']' 00:08:33.017 23:42:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@408 -- # strip_size=0 00:08:33.017 23:42:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=79370 00:08:33.017 23:42:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:08:33.017 23:42:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 79370 00:08:33.017 23:42:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@835 -- # '[' -z 79370 ']' 00:08:33.017 23:42:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:33.017 23:42:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:08:33.017 23:42:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:33.017 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:33.017 23:42:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:08:33.017 23:42:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:33.017 [2024-11-26 23:42:20.979678] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:08:33.017 [2024-11-26 23:42:20.980285] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid79370 ] 00:08:33.017 [2024-11-26 23:42:21.137618] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:33.279 [2024-11-26 23:42:21.166326] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:33.279 [2024-11-26 23:42:21.210040] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:33.279 [2024-11-26 23:42:21.210150] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:33.851 23:42:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:08:33.851 23:42:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@868 -- # return 0 00:08:33.851 23:42:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:08:33.851 23:42:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:08:33.851 23:42:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:08:33.851 23:42:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:08:33.851 23:42:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:08:33.851 23:42:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:08:33.851 23:42:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:08:33.851 23:42:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:08:33.851 23:42:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:08:33.851 23:42:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:33.851 23:42:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:33.851 malloc1 00:08:33.851 23:42:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:33.851 23:42:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:08:33.851 23:42:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:33.851 23:42:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:33.851 [2024-11-26 23:42:21.830304] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:08:33.851 [2024-11-26 23:42:21.830461] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:33.851 [2024-11-26 23:42:21.830502] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:08:33.851 [2024-11-26 23:42:21.830571] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:33.851 [2024-11-26 23:42:21.832754] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:33.851 [2024-11-26 23:42:21.832826] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:08:33.851 pt1 00:08:33.851 23:42:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:33.851 23:42:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:08:33.851 23:42:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:08:33.851 23:42:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:08:33.851 23:42:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:08:33.851 23:42:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:08:33.851 23:42:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:08:33.851 23:42:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:08:33.851 23:42:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:08:33.851 23:42:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:08:33.851 23:42:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:33.851 23:42:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:33.851 malloc2 00:08:33.851 23:42:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:33.851 23:42:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:08:33.851 23:42:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:33.851 23:42:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:33.851 [2024-11-26 23:42:21.858862] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:08:33.851 [2024-11-26 23:42:21.858937] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:33.851 [2024-11-26 23:42:21.858955] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:08:33.851 [2024-11-26 23:42:21.858966] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:33.851 [2024-11-26 23:42:21.861139] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:33.851 [2024-11-26 23:42:21.861210] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:08:33.851 pt2 00:08:33.851 23:42:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:33.851 23:42:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:08:33.851 23:42:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:08:33.851 23:42:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc3 00:08:33.851 23:42:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt3 00:08:33.851 23:42:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000003 00:08:33.851 23:42:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:08:33.851 23:42:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:08:33.851 23:42:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:08:33.851 23:42:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc3 00:08:33.851 23:42:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:33.851 23:42:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:33.851 malloc3 00:08:33.851 23:42:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:33.851 23:42:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:08:33.851 23:42:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:33.851 23:42:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:33.851 [2024-11-26 23:42:21.887742] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:08:33.851 [2024-11-26 23:42:21.887799] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:33.851 [2024-11-26 23:42:21.887819] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:08:33.851 [2024-11-26 23:42:21.887829] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:33.851 [2024-11-26 23:42:21.889966] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:33.851 [2024-11-26 23:42:21.890004] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:08:33.851 pt3 00:08:33.851 23:42:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:33.851 23:42:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:08:33.851 23:42:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:08:33.851 23:42:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''pt1 pt2 pt3'\''' -n raid_bdev1 -s 00:08:33.851 23:42:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:33.851 23:42:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:33.851 [2024-11-26 23:42:21.899795] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:08:33.851 [2024-11-26 23:42:21.901659] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:08:33.851 [2024-11-26 23:42:21.901773] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:08:33.851 [2024-11-26 23:42:21.901932] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:08:33.851 [2024-11-26 23:42:21.901946] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:08:33.851 [2024-11-26 23:42:21.902223] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:08:33.851 [2024-11-26 23:42:21.902379] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:08:33.851 [2024-11-26 23:42:21.902393] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:08:33.851 [2024-11-26 23:42:21.902524] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:33.851 23:42:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:33.851 23:42:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:08:33.851 23:42:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:33.851 23:42:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:33.851 23:42:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:33.851 23:42:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:33.851 23:42:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:33.851 23:42:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:33.851 23:42:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:33.851 23:42:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:33.851 23:42:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:33.851 23:42:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:33.851 23:42:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:33.851 23:42:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:33.851 23:42:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:33.852 23:42:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:33.852 23:42:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:33.852 "name": "raid_bdev1", 00:08:33.852 "uuid": "d6f5f119-aa2b-49d1-a498-29fc01c847a1", 00:08:33.852 "strip_size_kb": 0, 00:08:33.852 "state": "online", 00:08:33.852 "raid_level": "raid1", 00:08:33.852 "superblock": true, 00:08:33.852 "num_base_bdevs": 3, 00:08:33.852 "num_base_bdevs_discovered": 3, 00:08:33.852 "num_base_bdevs_operational": 3, 00:08:33.852 "base_bdevs_list": [ 00:08:33.852 { 00:08:33.852 "name": "pt1", 00:08:33.852 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:33.852 "is_configured": true, 00:08:33.852 "data_offset": 2048, 00:08:33.852 "data_size": 63488 00:08:33.852 }, 00:08:33.852 { 00:08:33.852 "name": "pt2", 00:08:33.852 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:33.852 "is_configured": true, 00:08:33.852 "data_offset": 2048, 00:08:33.852 "data_size": 63488 00:08:33.852 }, 00:08:33.852 { 00:08:33.852 "name": "pt3", 00:08:33.852 "uuid": "00000000-0000-0000-0000-000000000003", 00:08:33.852 "is_configured": true, 00:08:33.852 "data_offset": 2048, 00:08:33.852 "data_size": 63488 00:08:33.852 } 00:08:33.852 ] 00:08:33.852 }' 00:08:33.852 23:42:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:33.852 23:42:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:34.421 23:42:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:08:34.421 23:42:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:08:34.421 23:42:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:34.421 23:42:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:34.421 23:42:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:08:34.421 23:42:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:34.421 23:42:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:08:34.421 23:42:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:34.421 23:42:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:34.421 23:42:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:34.421 [2024-11-26 23:42:22.347377] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:34.421 23:42:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:34.421 23:42:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:34.421 "name": "raid_bdev1", 00:08:34.421 "aliases": [ 00:08:34.421 "d6f5f119-aa2b-49d1-a498-29fc01c847a1" 00:08:34.421 ], 00:08:34.421 "product_name": "Raid Volume", 00:08:34.421 "block_size": 512, 00:08:34.421 "num_blocks": 63488, 00:08:34.421 "uuid": "d6f5f119-aa2b-49d1-a498-29fc01c847a1", 00:08:34.421 "assigned_rate_limits": { 00:08:34.421 "rw_ios_per_sec": 0, 00:08:34.421 "rw_mbytes_per_sec": 0, 00:08:34.421 "r_mbytes_per_sec": 0, 00:08:34.421 "w_mbytes_per_sec": 0 00:08:34.421 }, 00:08:34.421 "claimed": false, 00:08:34.421 "zoned": false, 00:08:34.421 "supported_io_types": { 00:08:34.421 "read": true, 00:08:34.421 "write": true, 00:08:34.421 "unmap": false, 00:08:34.421 "flush": false, 00:08:34.421 "reset": true, 00:08:34.421 "nvme_admin": false, 00:08:34.421 "nvme_io": false, 00:08:34.421 "nvme_io_md": false, 00:08:34.421 "write_zeroes": true, 00:08:34.421 "zcopy": false, 00:08:34.421 "get_zone_info": false, 00:08:34.421 "zone_management": false, 00:08:34.421 "zone_append": false, 00:08:34.421 "compare": false, 00:08:34.421 "compare_and_write": false, 00:08:34.421 "abort": false, 00:08:34.421 "seek_hole": false, 00:08:34.421 "seek_data": false, 00:08:34.421 "copy": false, 00:08:34.421 "nvme_iov_md": false 00:08:34.421 }, 00:08:34.421 "memory_domains": [ 00:08:34.421 { 00:08:34.421 "dma_device_id": "system", 00:08:34.421 "dma_device_type": 1 00:08:34.421 }, 00:08:34.421 { 00:08:34.421 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:34.421 "dma_device_type": 2 00:08:34.421 }, 00:08:34.421 { 00:08:34.421 "dma_device_id": "system", 00:08:34.421 "dma_device_type": 1 00:08:34.421 }, 00:08:34.421 { 00:08:34.421 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:34.421 "dma_device_type": 2 00:08:34.421 }, 00:08:34.421 { 00:08:34.421 "dma_device_id": "system", 00:08:34.421 "dma_device_type": 1 00:08:34.421 }, 00:08:34.421 { 00:08:34.421 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:34.421 "dma_device_type": 2 00:08:34.421 } 00:08:34.421 ], 00:08:34.421 "driver_specific": { 00:08:34.421 "raid": { 00:08:34.421 "uuid": "d6f5f119-aa2b-49d1-a498-29fc01c847a1", 00:08:34.421 "strip_size_kb": 0, 00:08:34.421 "state": "online", 00:08:34.421 "raid_level": "raid1", 00:08:34.421 "superblock": true, 00:08:34.421 "num_base_bdevs": 3, 00:08:34.421 "num_base_bdevs_discovered": 3, 00:08:34.421 "num_base_bdevs_operational": 3, 00:08:34.421 "base_bdevs_list": [ 00:08:34.421 { 00:08:34.421 "name": "pt1", 00:08:34.421 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:34.421 "is_configured": true, 00:08:34.421 "data_offset": 2048, 00:08:34.421 "data_size": 63488 00:08:34.421 }, 00:08:34.421 { 00:08:34.421 "name": "pt2", 00:08:34.421 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:34.421 "is_configured": true, 00:08:34.421 "data_offset": 2048, 00:08:34.421 "data_size": 63488 00:08:34.421 }, 00:08:34.421 { 00:08:34.421 "name": "pt3", 00:08:34.421 "uuid": "00000000-0000-0000-0000-000000000003", 00:08:34.421 "is_configured": true, 00:08:34.421 "data_offset": 2048, 00:08:34.421 "data_size": 63488 00:08:34.421 } 00:08:34.421 ] 00:08:34.421 } 00:08:34.421 } 00:08:34.421 }' 00:08:34.422 23:42:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:34.422 23:42:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:08:34.422 pt2 00:08:34.422 pt3' 00:08:34.422 23:42:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:34.422 23:42:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:34.422 23:42:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:34.422 23:42:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:34.422 23:42:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:08:34.422 23:42:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:34.422 23:42:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:34.422 23:42:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:34.422 23:42:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:34.422 23:42:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:34.422 23:42:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:34.422 23:42:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:34.422 23:42:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:08:34.422 23:42:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:34.422 23:42:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:34.681 23:42:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:34.681 23:42:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:34.681 23:42:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:34.681 23:42:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:34.681 23:42:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:08:34.681 23:42:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:34.681 23:42:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:34.681 23:42:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:34.681 23:42:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:34.681 23:42:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:34.681 23:42:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:34.681 23:42:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:08:34.681 23:42:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:34.681 23:42:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:34.681 23:42:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:08:34.681 [2024-11-26 23:42:22.646724] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:34.681 23:42:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:34.681 23:42:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=d6f5f119-aa2b-49d1-a498-29fc01c847a1 00:08:34.681 23:42:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z d6f5f119-aa2b-49d1-a498-29fc01c847a1 ']' 00:08:34.681 23:42:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:08:34.681 23:42:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:34.681 23:42:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:34.681 [2024-11-26 23:42:22.698431] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:08:34.681 [2024-11-26 23:42:22.698494] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:34.681 [2024-11-26 23:42:22.698596] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:34.681 [2024-11-26 23:42:22.698692] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:34.681 [2024-11-26 23:42:22.698739] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:08:34.681 23:42:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:34.681 23:42:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:08:34.681 23:42:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:34.681 23:42:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:34.681 23:42:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:34.681 23:42:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:34.681 23:42:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:08:34.681 23:42:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:08:34.681 23:42:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:08:34.681 23:42:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:08:34.681 23:42:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:34.681 23:42:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:34.681 23:42:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:34.681 23:42:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:08:34.681 23:42:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:08:34.681 23:42:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:34.681 23:42:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:34.681 23:42:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:34.681 23:42:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:08:34.681 23:42:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt3 00:08:34.681 23:42:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:34.681 23:42:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:34.681 23:42:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:34.681 23:42:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:08:34.681 23:42:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:08:34.681 23:42:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:34.681 23:42:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:34.941 23:42:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:34.941 23:42:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:08:34.941 23:42:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:08:34.941 23:42:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@652 -- # local es=0 00:08:34.941 23:42:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:08:34.941 23:42:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:08:34.941 23:42:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:08:34.942 23:42:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:08:34.942 23:42:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:08:34.942 23:42:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:08:34.942 23:42:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:34.942 23:42:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:34.942 [2024-11-26 23:42:22.842209] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:08:34.942 [2024-11-26 23:42:22.844012] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:08:34.942 [2024-11-26 23:42:22.844058] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc3 is claimed 00:08:34.942 [2024-11-26 23:42:22.844106] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:08:34.942 [2024-11-26 23:42:22.844160] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:08:34.942 [2024-11-26 23:42:22.844180] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc3 00:08:34.942 [2024-11-26 23:42:22.844192] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:08:34.942 [2024-11-26 23:42:22.844201] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state configuring 00:08:34.942 request: 00:08:34.942 { 00:08:34.942 "name": "raid_bdev1", 00:08:34.942 "raid_level": "raid1", 00:08:34.942 "base_bdevs": [ 00:08:34.942 "malloc1", 00:08:34.942 "malloc2", 00:08:34.942 "malloc3" 00:08:34.942 ], 00:08:34.942 "superblock": false, 00:08:34.942 "method": "bdev_raid_create", 00:08:34.942 "req_id": 1 00:08:34.942 } 00:08:34.942 Got JSON-RPC error response 00:08:34.942 response: 00:08:34.942 { 00:08:34.942 "code": -17, 00:08:34.942 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:08:34.942 } 00:08:34.942 23:42:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:08:34.942 23:42:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # es=1 00:08:34.942 23:42:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:08:34.942 23:42:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:08:34.942 23:42:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:08:34.942 23:42:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:34.942 23:42:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:08:34.942 23:42:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:34.942 23:42:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:34.942 23:42:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:34.942 23:42:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:08:34.942 23:42:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:08:34.942 23:42:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:08:34.942 23:42:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:34.942 23:42:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:34.942 [2024-11-26 23:42:22.910057] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:08:34.942 [2024-11-26 23:42:22.910153] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:34.942 [2024-11-26 23:42:22.910188] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:08:34.942 [2024-11-26 23:42:22.910217] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:34.942 [2024-11-26 23:42:22.912296] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:34.942 [2024-11-26 23:42:22.912375] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:08:34.942 [2024-11-26 23:42:22.912459] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:08:34.942 [2024-11-26 23:42:22.912525] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:08:34.942 pt1 00:08:34.942 23:42:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:34.942 23:42:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 3 00:08:34.942 23:42:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:34.942 23:42:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:34.942 23:42:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:34.942 23:42:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:34.942 23:42:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:34.942 23:42:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:34.942 23:42:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:34.942 23:42:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:34.942 23:42:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:34.942 23:42:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:34.942 23:42:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:34.942 23:42:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:34.942 23:42:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:34.942 23:42:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:34.942 23:42:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:34.942 "name": "raid_bdev1", 00:08:34.942 "uuid": "d6f5f119-aa2b-49d1-a498-29fc01c847a1", 00:08:34.942 "strip_size_kb": 0, 00:08:34.942 "state": "configuring", 00:08:34.942 "raid_level": "raid1", 00:08:34.942 "superblock": true, 00:08:34.942 "num_base_bdevs": 3, 00:08:34.942 "num_base_bdevs_discovered": 1, 00:08:34.942 "num_base_bdevs_operational": 3, 00:08:34.942 "base_bdevs_list": [ 00:08:34.942 { 00:08:34.942 "name": "pt1", 00:08:34.942 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:34.942 "is_configured": true, 00:08:34.942 "data_offset": 2048, 00:08:34.942 "data_size": 63488 00:08:34.942 }, 00:08:34.942 { 00:08:34.942 "name": null, 00:08:34.942 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:34.942 "is_configured": false, 00:08:34.942 "data_offset": 2048, 00:08:34.942 "data_size": 63488 00:08:34.942 }, 00:08:34.942 { 00:08:34.942 "name": null, 00:08:34.942 "uuid": "00000000-0000-0000-0000-000000000003", 00:08:34.942 "is_configured": false, 00:08:34.942 "data_offset": 2048, 00:08:34.942 "data_size": 63488 00:08:34.942 } 00:08:34.942 ] 00:08:34.942 }' 00:08:34.942 23:42:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:34.942 23:42:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:35.202 23:42:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 3 -gt 2 ']' 00:08:35.202 23:42:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@472 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:08:35.202 23:42:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:35.202 23:42:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:35.202 [2024-11-26 23:42:23.313424] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:08:35.202 [2024-11-26 23:42:23.313487] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:35.202 [2024-11-26 23:42:23.313508] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:08:35.202 [2024-11-26 23:42:23.313521] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:35.202 [2024-11-26 23:42:23.313908] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:35.202 [2024-11-26 23:42:23.313927] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:08:35.202 [2024-11-26 23:42:23.313995] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:08:35.202 [2024-11-26 23:42:23.314027] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:08:35.202 pt2 00:08:35.202 23:42:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:35.202 23:42:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@473 -- # rpc_cmd bdev_passthru_delete pt2 00:08:35.202 23:42:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:35.202 23:42:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:35.202 [2024-11-26 23:42:23.325424] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt2 00:08:35.461 23:42:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:35.461 23:42:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@474 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 3 00:08:35.461 23:42:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:35.461 23:42:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:35.461 23:42:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:35.461 23:42:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:35.461 23:42:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:35.461 23:42:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:35.461 23:42:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:35.461 23:42:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:35.461 23:42:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:35.461 23:42:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:35.461 23:42:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:35.461 23:42:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:35.461 23:42:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:35.461 23:42:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:35.461 23:42:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:35.461 "name": "raid_bdev1", 00:08:35.461 "uuid": "d6f5f119-aa2b-49d1-a498-29fc01c847a1", 00:08:35.461 "strip_size_kb": 0, 00:08:35.461 "state": "configuring", 00:08:35.461 "raid_level": "raid1", 00:08:35.461 "superblock": true, 00:08:35.461 "num_base_bdevs": 3, 00:08:35.461 "num_base_bdevs_discovered": 1, 00:08:35.461 "num_base_bdevs_operational": 3, 00:08:35.461 "base_bdevs_list": [ 00:08:35.461 { 00:08:35.461 "name": "pt1", 00:08:35.461 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:35.461 "is_configured": true, 00:08:35.461 "data_offset": 2048, 00:08:35.461 "data_size": 63488 00:08:35.461 }, 00:08:35.461 { 00:08:35.461 "name": null, 00:08:35.461 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:35.461 "is_configured": false, 00:08:35.461 "data_offset": 0, 00:08:35.461 "data_size": 63488 00:08:35.461 }, 00:08:35.461 { 00:08:35.461 "name": null, 00:08:35.461 "uuid": "00000000-0000-0000-0000-000000000003", 00:08:35.461 "is_configured": false, 00:08:35.461 "data_offset": 2048, 00:08:35.461 "data_size": 63488 00:08:35.461 } 00:08:35.461 ] 00:08:35.461 }' 00:08:35.461 23:42:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:35.461 23:42:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:35.722 23:42:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:08:35.722 23:42:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:08:35.722 23:42:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:08:35.722 23:42:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:35.722 23:42:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:35.722 [2024-11-26 23:42:23.724688] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:08:35.722 [2024-11-26 23:42:23.724795] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:35.722 [2024-11-26 23:42:23.724832] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:08:35.722 [2024-11-26 23:42:23.724859] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:35.722 [2024-11-26 23:42:23.725299] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:35.722 [2024-11-26 23:42:23.725364] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:08:35.722 [2024-11-26 23:42:23.725469] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:08:35.722 [2024-11-26 23:42:23.725519] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:08:35.722 pt2 00:08:35.722 23:42:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:35.722 23:42:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:08:35.722 23:42:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:08:35.722 23:42:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:08:35.722 23:42:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:35.722 23:42:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:35.722 [2024-11-26 23:42:23.736659] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:08:35.722 [2024-11-26 23:42:23.736739] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:35.722 [2024-11-26 23:42:23.736772] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:08:35.722 [2024-11-26 23:42:23.736797] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:35.722 [2024-11-26 23:42:23.737116] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:35.722 [2024-11-26 23:42:23.737167] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:08:35.722 [2024-11-26 23:42:23.737247] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:08:35.722 [2024-11-26 23:42:23.737291] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:08:35.722 [2024-11-26 23:42:23.737422] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:08:35.722 [2024-11-26 23:42:23.737462] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:08:35.722 [2024-11-26 23:42:23.737700] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:08:35.722 [2024-11-26 23:42:23.737841] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:08:35.722 [2024-11-26 23:42:23.737881] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:08:35.722 [2024-11-26 23:42:23.738008] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:35.722 pt3 00:08:35.722 23:42:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:35.722 23:42:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:08:35.722 23:42:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:08:35.722 23:42:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:08:35.722 23:42:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:35.722 23:42:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:35.722 23:42:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:35.722 23:42:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:35.722 23:42:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:35.722 23:42:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:35.722 23:42:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:35.722 23:42:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:35.722 23:42:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:35.722 23:42:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:35.722 23:42:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:35.722 23:42:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:35.722 23:42:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:35.722 23:42:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:35.722 23:42:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:35.722 "name": "raid_bdev1", 00:08:35.722 "uuid": "d6f5f119-aa2b-49d1-a498-29fc01c847a1", 00:08:35.722 "strip_size_kb": 0, 00:08:35.722 "state": "online", 00:08:35.722 "raid_level": "raid1", 00:08:35.722 "superblock": true, 00:08:35.722 "num_base_bdevs": 3, 00:08:35.722 "num_base_bdevs_discovered": 3, 00:08:35.722 "num_base_bdevs_operational": 3, 00:08:35.722 "base_bdevs_list": [ 00:08:35.722 { 00:08:35.722 "name": "pt1", 00:08:35.722 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:35.722 "is_configured": true, 00:08:35.722 "data_offset": 2048, 00:08:35.722 "data_size": 63488 00:08:35.722 }, 00:08:35.722 { 00:08:35.722 "name": "pt2", 00:08:35.722 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:35.722 "is_configured": true, 00:08:35.722 "data_offset": 2048, 00:08:35.722 "data_size": 63488 00:08:35.722 }, 00:08:35.722 { 00:08:35.722 "name": "pt3", 00:08:35.722 "uuid": "00000000-0000-0000-0000-000000000003", 00:08:35.722 "is_configured": true, 00:08:35.722 "data_offset": 2048, 00:08:35.722 "data_size": 63488 00:08:35.722 } 00:08:35.722 ] 00:08:35.722 }' 00:08:35.722 23:42:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:35.722 23:42:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:36.341 23:42:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:08:36.341 23:42:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:08:36.341 23:42:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:36.341 23:42:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:36.341 23:42:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:08:36.341 23:42:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:36.341 23:42:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:08:36.341 23:42:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:36.341 23:42:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:36.341 23:42:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:36.341 [2024-11-26 23:42:24.176188] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:36.341 23:42:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:36.341 23:42:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:36.341 "name": "raid_bdev1", 00:08:36.341 "aliases": [ 00:08:36.341 "d6f5f119-aa2b-49d1-a498-29fc01c847a1" 00:08:36.341 ], 00:08:36.341 "product_name": "Raid Volume", 00:08:36.341 "block_size": 512, 00:08:36.341 "num_blocks": 63488, 00:08:36.341 "uuid": "d6f5f119-aa2b-49d1-a498-29fc01c847a1", 00:08:36.341 "assigned_rate_limits": { 00:08:36.341 "rw_ios_per_sec": 0, 00:08:36.341 "rw_mbytes_per_sec": 0, 00:08:36.341 "r_mbytes_per_sec": 0, 00:08:36.341 "w_mbytes_per_sec": 0 00:08:36.341 }, 00:08:36.341 "claimed": false, 00:08:36.341 "zoned": false, 00:08:36.341 "supported_io_types": { 00:08:36.341 "read": true, 00:08:36.341 "write": true, 00:08:36.341 "unmap": false, 00:08:36.341 "flush": false, 00:08:36.341 "reset": true, 00:08:36.341 "nvme_admin": false, 00:08:36.341 "nvme_io": false, 00:08:36.341 "nvme_io_md": false, 00:08:36.341 "write_zeroes": true, 00:08:36.341 "zcopy": false, 00:08:36.341 "get_zone_info": false, 00:08:36.341 "zone_management": false, 00:08:36.341 "zone_append": false, 00:08:36.341 "compare": false, 00:08:36.341 "compare_and_write": false, 00:08:36.341 "abort": false, 00:08:36.341 "seek_hole": false, 00:08:36.341 "seek_data": false, 00:08:36.341 "copy": false, 00:08:36.341 "nvme_iov_md": false 00:08:36.341 }, 00:08:36.341 "memory_domains": [ 00:08:36.341 { 00:08:36.341 "dma_device_id": "system", 00:08:36.341 "dma_device_type": 1 00:08:36.341 }, 00:08:36.341 { 00:08:36.341 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:36.341 "dma_device_type": 2 00:08:36.341 }, 00:08:36.341 { 00:08:36.341 "dma_device_id": "system", 00:08:36.341 "dma_device_type": 1 00:08:36.341 }, 00:08:36.341 { 00:08:36.341 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:36.341 "dma_device_type": 2 00:08:36.341 }, 00:08:36.341 { 00:08:36.341 "dma_device_id": "system", 00:08:36.341 "dma_device_type": 1 00:08:36.341 }, 00:08:36.341 { 00:08:36.341 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:36.341 "dma_device_type": 2 00:08:36.341 } 00:08:36.341 ], 00:08:36.341 "driver_specific": { 00:08:36.342 "raid": { 00:08:36.342 "uuid": "d6f5f119-aa2b-49d1-a498-29fc01c847a1", 00:08:36.342 "strip_size_kb": 0, 00:08:36.342 "state": "online", 00:08:36.342 "raid_level": "raid1", 00:08:36.342 "superblock": true, 00:08:36.342 "num_base_bdevs": 3, 00:08:36.342 "num_base_bdevs_discovered": 3, 00:08:36.342 "num_base_bdevs_operational": 3, 00:08:36.342 "base_bdevs_list": [ 00:08:36.342 { 00:08:36.342 "name": "pt1", 00:08:36.342 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:36.342 "is_configured": true, 00:08:36.342 "data_offset": 2048, 00:08:36.342 "data_size": 63488 00:08:36.342 }, 00:08:36.342 { 00:08:36.342 "name": "pt2", 00:08:36.342 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:36.342 "is_configured": true, 00:08:36.342 "data_offset": 2048, 00:08:36.342 "data_size": 63488 00:08:36.342 }, 00:08:36.342 { 00:08:36.342 "name": "pt3", 00:08:36.342 "uuid": "00000000-0000-0000-0000-000000000003", 00:08:36.342 "is_configured": true, 00:08:36.342 "data_offset": 2048, 00:08:36.342 "data_size": 63488 00:08:36.342 } 00:08:36.342 ] 00:08:36.342 } 00:08:36.342 } 00:08:36.342 }' 00:08:36.342 23:42:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:36.342 23:42:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:08:36.342 pt2 00:08:36.342 pt3' 00:08:36.342 23:42:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:36.342 23:42:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:36.342 23:42:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:36.342 23:42:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:36.342 23:42:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:08:36.342 23:42:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:36.342 23:42:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:36.342 23:42:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:36.342 23:42:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:36.342 23:42:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:36.342 23:42:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:36.342 23:42:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:36.342 23:42:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:08:36.342 23:42:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:36.342 23:42:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:36.342 23:42:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:36.342 23:42:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:36.342 23:42:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:36.342 23:42:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:36.342 23:42:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:08:36.342 23:42:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:36.342 23:42:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:36.342 23:42:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:36.342 23:42:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:36.342 23:42:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:36.342 23:42:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:36.602 23:42:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:08:36.602 23:42:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:08:36.602 23:42:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:36.602 23:42:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:36.602 [2024-11-26 23:42:24.479652] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:36.602 23:42:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:36.602 23:42:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' d6f5f119-aa2b-49d1-a498-29fc01c847a1 '!=' d6f5f119-aa2b-49d1-a498-29fc01c847a1 ']' 00:08:36.602 23:42:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy raid1 00:08:36.602 23:42:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:08:36.602 23:42:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@199 -- # return 0 00:08:36.602 23:42:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@493 -- # rpc_cmd bdev_passthru_delete pt1 00:08:36.602 23:42:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:36.602 23:42:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:36.602 [2024-11-26 23:42:24.523362] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt1 00:08:36.602 23:42:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:36.602 23:42:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@496 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:08:36.602 23:42:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:36.602 23:42:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:36.602 23:42:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:36.602 23:42:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:36.603 23:42:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:36.603 23:42:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:36.603 23:42:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:36.603 23:42:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:36.603 23:42:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:36.603 23:42:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:36.603 23:42:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:36.603 23:42:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:36.603 23:42:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:36.603 23:42:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:36.603 23:42:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:36.603 "name": "raid_bdev1", 00:08:36.603 "uuid": "d6f5f119-aa2b-49d1-a498-29fc01c847a1", 00:08:36.603 "strip_size_kb": 0, 00:08:36.603 "state": "online", 00:08:36.603 "raid_level": "raid1", 00:08:36.603 "superblock": true, 00:08:36.603 "num_base_bdevs": 3, 00:08:36.603 "num_base_bdevs_discovered": 2, 00:08:36.603 "num_base_bdevs_operational": 2, 00:08:36.603 "base_bdevs_list": [ 00:08:36.603 { 00:08:36.603 "name": null, 00:08:36.603 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:36.603 "is_configured": false, 00:08:36.603 "data_offset": 0, 00:08:36.603 "data_size": 63488 00:08:36.603 }, 00:08:36.603 { 00:08:36.603 "name": "pt2", 00:08:36.603 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:36.603 "is_configured": true, 00:08:36.603 "data_offset": 2048, 00:08:36.603 "data_size": 63488 00:08:36.603 }, 00:08:36.603 { 00:08:36.603 "name": "pt3", 00:08:36.603 "uuid": "00000000-0000-0000-0000-000000000003", 00:08:36.603 "is_configured": true, 00:08:36.603 "data_offset": 2048, 00:08:36.603 "data_size": 63488 00:08:36.603 } 00:08:36.603 ] 00:08:36.603 }' 00:08:36.603 23:42:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:36.603 23:42:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:36.863 23:42:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@499 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:08:36.863 23:42:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:36.863 23:42:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:36.863 [2024-11-26 23:42:24.974567] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:08:36.863 [2024-11-26 23:42:24.974662] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:36.863 [2024-11-26 23:42:24.974759] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:36.863 [2024-11-26 23:42:24.974835] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:36.863 [2024-11-26 23:42:24.974878] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:08:36.863 23:42:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:36.863 23:42:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:36.863 23:42:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # jq -r '.[]' 00:08:36.863 23:42:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:36.863 23:42:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:36.863 23:42:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:37.127 23:42:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # raid_bdev= 00:08:37.127 23:42:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@501 -- # '[' -n '' ']' 00:08:37.127 23:42:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i = 1 )) 00:08:37.127 23:42:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:08:37.127 23:42:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt2 00:08:37.127 23:42:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:37.127 23:42:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:37.127 23:42:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:37.127 23:42:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:08:37.127 23:42:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:08:37.127 23:42:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt3 00:08:37.127 23:42:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:37.127 23:42:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:37.127 23:42:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:37.127 23:42:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:08:37.127 23:42:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:08:37.127 23:42:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i = 1 )) 00:08:37.127 23:42:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:08:37.127 23:42:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@512 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:08:37.127 23:42:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:37.127 23:42:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:37.127 [2024-11-26 23:42:25.042446] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:08:37.127 [2024-11-26 23:42:25.042537] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:37.127 [2024-11-26 23:42:25.042589] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009980 00:08:37.127 [2024-11-26 23:42:25.042617] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:37.127 [2024-11-26 23:42:25.044796] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:37.127 [2024-11-26 23:42:25.044863] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:08:37.127 [2024-11-26 23:42:25.044958] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:08:37.127 [2024-11-26 23:42:25.045006] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:08:37.127 pt2 00:08:37.127 23:42:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:37.127 23:42:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@515 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 2 00:08:37.127 23:42:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:37.127 23:42:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:37.127 23:42:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:37.127 23:42:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:37.127 23:42:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:37.127 23:42:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:37.127 23:42:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:37.127 23:42:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:37.127 23:42:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:37.127 23:42:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:37.127 23:42:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:37.127 23:42:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:37.127 23:42:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:37.127 23:42:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:37.127 23:42:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:37.127 "name": "raid_bdev1", 00:08:37.127 "uuid": "d6f5f119-aa2b-49d1-a498-29fc01c847a1", 00:08:37.127 "strip_size_kb": 0, 00:08:37.127 "state": "configuring", 00:08:37.127 "raid_level": "raid1", 00:08:37.127 "superblock": true, 00:08:37.127 "num_base_bdevs": 3, 00:08:37.127 "num_base_bdevs_discovered": 1, 00:08:37.127 "num_base_bdevs_operational": 2, 00:08:37.127 "base_bdevs_list": [ 00:08:37.127 { 00:08:37.127 "name": null, 00:08:37.127 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:37.127 "is_configured": false, 00:08:37.127 "data_offset": 2048, 00:08:37.127 "data_size": 63488 00:08:37.127 }, 00:08:37.127 { 00:08:37.127 "name": "pt2", 00:08:37.127 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:37.127 "is_configured": true, 00:08:37.127 "data_offset": 2048, 00:08:37.127 "data_size": 63488 00:08:37.127 }, 00:08:37.127 { 00:08:37.127 "name": null, 00:08:37.127 "uuid": "00000000-0000-0000-0000-000000000003", 00:08:37.127 "is_configured": false, 00:08:37.127 "data_offset": 2048, 00:08:37.127 "data_size": 63488 00:08:37.127 } 00:08:37.127 ] 00:08:37.127 }' 00:08:37.127 23:42:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:37.127 23:42:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:37.696 23:42:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i++ )) 00:08:37.696 23:42:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:08:37.696 23:42:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@519 -- # i=2 00:08:37.696 23:42:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@520 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:08:37.696 23:42:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:37.696 23:42:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:37.696 [2024-11-26 23:42:25.533692] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:08:37.696 [2024-11-26 23:42:25.533764] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:37.696 [2024-11-26 23:42:25.533788] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009f80 00:08:37.696 [2024-11-26 23:42:25.533797] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:37.696 [2024-11-26 23:42:25.534206] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:37.696 [2024-11-26 23:42:25.534223] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:08:37.696 [2024-11-26 23:42:25.534301] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:08:37.696 [2024-11-26 23:42:25.534321] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:08:37.696 [2024-11-26 23:42:25.534433] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:08:37.696 [2024-11-26 23:42:25.534442] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:08:37.696 [2024-11-26 23:42:25.534686] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:08:37.696 [2024-11-26 23:42:25.534814] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:08:37.696 [2024-11-26 23:42:25.534827] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001c80 00:08:37.696 [2024-11-26 23:42:25.534929] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:37.696 pt3 00:08:37.696 23:42:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:37.696 23:42:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@523 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:08:37.696 23:42:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:37.696 23:42:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:37.696 23:42:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:37.696 23:42:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:37.696 23:42:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:37.696 23:42:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:37.696 23:42:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:37.696 23:42:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:37.697 23:42:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:37.697 23:42:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:37.697 23:42:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:37.697 23:42:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:37.697 23:42:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:37.697 23:42:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:37.697 23:42:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:37.697 "name": "raid_bdev1", 00:08:37.697 "uuid": "d6f5f119-aa2b-49d1-a498-29fc01c847a1", 00:08:37.697 "strip_size_kb": 0, 00:08:37.697 "state": "online", 00:08:37.697 "raid_level": "raid1", 00:08:37.697 "superblock": true, 00:08:37.697 "num_base_bdevs": 3, 00:08:37.697 "num_base_bdevs_discovered": 2, 00:08:37.697 "num_base_bdevs_operational": 2, 00:08:37.697 "base_bdevs_list": [ 00:08:37.697 { 00:08:37.697 "name": null, 00:08:37.697 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:37.697 "is_configured": false, 00:08:37.697 "data_offset": 2048, 00:08:37.697 "data_size": 63488 00:08:37.697 }, 00:08:37.697 { 00:08:37.697 "name": "pt2", 00:08:37.697 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:37.697 "is_configured": true, 00:08:37.697 "data_offset": 2048, 00:08:37.697 "data_size": 63488 00:08:37.697 }, 00:08:37.697 { 00:08:37.697 "name": "pt3", 00:08:37.697 "uuid": "00000000-0000-0000-0000-000000000003", 00:08:37.697 "is_configured": true, 00:08:37.697 "data_offset": 2048, 00:08:37.697 "data_size": 63488 00:08:37.697 } 00:08:37.697 ] 00:08:37.697 }' 00:08:37.697 23:42:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:37.697 23:42:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:37.957 23:42:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@526 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:08:37.957 23:42:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:37.957 23:42:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:37.957 [2024-11-26 23:42:25.980933] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:08:37.957 [2024-11-26 23:42:25.980963] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:37.957 [2024-11-26 23:42:25.981041] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:37.957 [2024-11-26 23:42:25.981098] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:37.958 [2024-11-26 23:42:25.981109] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name raid_bdev1, state offline 00:08:37.958 23:42:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:37.958 23:42:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # jq -r '.[]' 00:08:37.958 23:42:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:37.958 23:42:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:37.958 23:42:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:37.958 23:42:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:37.958 23:42:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # raid_bdev= 00:08:37.958 23:42:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@528 -- # '[' -n '' ']' 00:08:37.958 23:42:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@532 -- # '[' 3 -gt 2 ']' 00:08:37.958 23:42:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@534 -- # i=2 00:08:37.958 23:42:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@535 -- # rpc_cmd bdev_passthru_delete pt3 00:08:37.958 23:42:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:37.958 23:42:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:37.958 23:42:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:37.958 23:42:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@540 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:08:37.958 23:42:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:37.958 23:42:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:37.958 [2024-11-26 23:42:26.060795] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:08:37.958 [2024-11-26 23:42:26.060899] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:37.958 [2024-11-26 23:42:26.060923] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:08:37.958 [2024-11-26 23:42:26.060934] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:37.958 [2024-11-26 23:42:26.063130] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:37.958 [2024-11-26 23:42:26.063211] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:08:37.958 [2024-11-26 23:42:26.063306] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:08:37.958 [2024-11-26 23:42:26.063355] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:08:37.958 [2024-11-26 23:42:26.063489] bdev_raid.c:3685:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev pt2 (4) greater than existing raid bdev raid_bdev1 (2) 00:08:37.958 [2024-11-26 23:42:26.063504] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:08:37.958 [2024-11-26 23:42:26.063519] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002000 name raid_bdev1, state configuring 00:08:37.958 [2024-11-26 23:42:26.063552] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:08:37.958 pt1 00:08:37.958 23:42:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:37.958 23:42:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@542 -- # '[' 3 -gt 2 ']' 00:08:37.958 23:42:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@545 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 2 00:08:37.958 23:42:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:37.958 23:42:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:37.958 23:42:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:37.958 23:42:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:37.958 23:42:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:37.958 23:42:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:37.958 23:42:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:37.958 23:42:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:37.958 23:42:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:37.958 23:42:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:37.958 23:42:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:37.958 23:42:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:37.958 23:42:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:38.218 23:42:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:38.218 23:42:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:38.218 "name": "raid_bdev1", 00:08:38.218 "uuid": "d6f5f119-aa2b-49d1-a498-29fc01c847a1", 00:08:38.218 "strip_size_kb": 0, 00:08:38.218 "state": "configuring", 00:08:38.218 "raid_level": "raid1", 00:08:38.218 "superblock": true, 00:08:38.218 "num_base_bdevs": 3, 00:08:38.218 "num_base_bdevs_discovered": 1, 00:08:38.218 "num_base_bdevs_operational": 2, 00:08:38.218 "base_bdevs_list": [ 00:08:38.218 { 00:08:38.218 "name": null, 00:08:38.218 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:38.218 "is_configured": false, 00:08:38.218 "data_offset": 2048, 00:08:38.218 "data_size": 63488 00:08:38.218 }, 00:08:38.218 { 00:08:38.218 "name": "pt2", 00:08:38.218 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:38.218 "is_configured": true, 00:08:38.218 "data_offset": 2048, 00:08:38.218 "data_size": 63488 00:08:38.218 }, 00:08:38.218 { 00:08:38.218 "name": null, 00:08:38.218 "uuid": "00000000-0000-0000-0000-000000000003", 00:08:38.218 "is_configured": false, 00:08:38.218 "data_offset": 2048, 00:08:38.218 "data_size": 63488 00:08:38.218 } 00:08:38.218 ] 00:08:38.218 }' 00:08:38.218 23:42:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:38.218 23:42:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:38.478 23:42:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@546 -- # rpc_cmd bdev_raid_get_bdevs configuring 00:08:38.478 23:42:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@546 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:08:38.478 23:42:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:38.478 23:42:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:38.478 23:42:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:38.750 23:42:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@546 -- # [[ false == \f\a\l\s\e ]] 00:08:38.751 23:42:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@549 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:08:38.751 23:42:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:38.751 23:42:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:38.751 [2024-11-26 23:42:26.627812] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:08:38.751 [2024-11-26 23:42:26.627916] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:38.751 [2024-11-26 23:42:26.627950] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a880 00:08:38.751 [2024-11-26 23:42:26.627980] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:38.751 [2024-11-26 23:42:26.628413] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:38.751 [2024-11-26 23:42:26.628475] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:08:38.751 [2024-11-26 23:42:26.628574] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:08:38.751 [2024-11-26 23:42:26.628627] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:08:38.751 [2024-11-26 23:42:26.628739] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000002380 00:08:38.751 [2024-11-26 23:42:26.628779] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:08:38.751 [2024-11-26 23:42:26.629027] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000026d0 00:08:38.751 [2024-11-26 23:42:26.629184] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000002380 00:08:38.751 [2024-11-26 23:42:26.629222] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000002380 00:08:38.751 [2024-11-26 23:42:26.629372] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:38.751 pt3 00:08:38.751 23:42:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:38.751 23:42:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@554 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:08:38.751 23:42:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:38.751 23:42:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:38.751 23:42:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:38.751 23:42:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:38.751 23:42:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:38.751 23:42:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:38.751 23:42:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:38.751 23:42:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:38.751 23:42:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:38.751 23:42:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:38.751 23:42:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:38.752 23:42:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:38.752 23:42:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:38.752 23:42:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:38.752 23:42:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:38.752 "name": "raid_bdev1", 00:08:38.752 "uuid": "d6f5f119-aa2b-49d1-a498-29fc01c847a1", 00:08:38.752 "strip_size_kb": 0, 00:08:38.752 "state": "online", 00:08:38.752 "raid_level": "raid1", 00:08:38.752 "superblock": true, 00:08:38.752 "num_base_bdevs": 3, 00:08:38.752 "num_base_bdevs_discovered": 2, 00:08:38.752 "num_base_bdevs_operational": 2, 00:08:38.752 "base_bdevs_list": [ 00:08:38.752 { 00:08:38.752 "name": null, 00:08:38.752 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:38.752 "is_configured": false, 00:08:38.752 "data_offset": 2048, 00:08:38.752 "data_size": 63488 00:08:38.752 }, 00:08:38.752 { 00:08:38.752 "name": "pt2", 00:08:38.752 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:38.752 "is_configured": true, 00:08:38.752 "data_offset": 2048, 00:08:38.752 "data_size": 63488 00:08:38.752 }, 00:08:38.752 { 00:08:38.752 "name": "pt3", 00:08:38.752 "uuid": "00000000-0000-0000-0000-000000000003", 00:08:38.752 "is_configured": true, 00:08:38.752 "data_offset": 2048, 00:08:38.752 "data_size": 63488 00:08:38.752 } 00:08:38.752 ] 00:08:38.752 }' 00:08:38.752 23:42:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:38.752 23:42:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:39.017 23:42:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@555 -- # rpc_cmd bdev_raid_get_bdevs online 00:08:39.017 23:42:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@555 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:08:39.017 23:42:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:39.017 23:42:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:39.018 23:42:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:39.018 23:42:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@555 -- # [[ false == \f\a\l\s\e ]] 00:08:39.018 23:42:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@558 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:08:39.018 23:42:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@558 -- # jq -r '.[] | .uuid' 00:08:39.018 23:42:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:39.018 23:42:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:39.018 [2024-11-26 23:42:27.087386] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:39.018 23:42:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:39.018 23:42:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@558 -- # '[' d6f5f119-aa2b-49d1-a498-29fc01c847a1 '!=' d6f5f119-aa2b-49d1-a498-29fc01c847a1 ']' 00:08:39.018 23:42:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 79370 00:08:39.018 23:42:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # '[' -z 79370 ']' 00:08:39.018 23:42:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@958 -- # kill -0 79370 00:08:39.018 23:42:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # uname 00:08:39.018 23:42:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:08:39.018 23:42:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 79370 00:08:39.276 23:42:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:08:39.276 23:42:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:08:39.276 23:42:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 79370' 00:08:39.276 killing process with pid 79370 00:08:39.276 23:42:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@973 -- # kill 79370 00:08:39.276 [2024-11-26 23:42:27.159229] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:39.276 [2024-11-26 23:42:27.159405] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:39.276 23:42:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@978 -- # wait 79370 00:08:39.276 [2024-11-26 23:42:27.159476] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:39.276 [2024-11-26 23:42:27.159486] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002380 name raid_bdev1, state offline 00:08:39.276 [2024-11-26 23:42:27.193894] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:39.537 23:42:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:08:39.537 00:08:39.537 real 0m6.515s 00:08:39.537 user 0m11.020s 00:08:39.537 sys 0m1.276s 00:08:39.537 23:42:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:08:39.537 23:42:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:39.537 ************************************ 00:08:39.537 END TEST raid_superblock_test 00:08:39.537 ************************************ 00:08:39.537 23:42:27 bdev_raid -- bdev/bdev_raid.sh@971 -- # run_test raid_read_error_test raid_io_error_test raid1 3 read 00:08:39.537 23:42:27 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:08:39.537 23:42:27 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:08:39.537 23:42:27 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:39.537 ************************************ 00:08:39.537 START TEST raid_read_error_test 00:08:39.537 ************************************ 00:08:39.537 23:42:27 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test raid1 3 read 00:08:39.537 23:42:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid1 00:08:39.537 23:42:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=3 00:08:39.537 23:42:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=read 00:08:39.537 23:42:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:08:39.537 23:42:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:39.537 23:42:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:08:39.537 23:42:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:08:39.537 23:42:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:39.537 23:42:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:08:39.537 23:42:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:08:39.537 23:42:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:39.537 23:42:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:08:39.537 23:42:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:08:39.537 23:42:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:39.537 23:42:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:08:39.537 23:42:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:08:39.537 23:42:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:08:39.537 23:42:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:08:39.537 23:42:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:08:39.537 23:42:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:08:39.537 23:42:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:08:39.537 23:42:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid1 '!=' raid1 ']' 00:08:39.537 23:42:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@804 -- # strip_size=0 00:08:39.537 23:42:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:08:39.537 23:42:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.j4tTDtxpD4 00:08:39.537 23:42:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=79805 00:08:39.537 23:42:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 79805 00:08:39.537 23:42:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:08:39.537 23:42:27 bdev_raid.raid_read_error_test -- common/autotest_common.sh@835 -- # '[' -z 79805 ']' 00:08:39.537 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:39.537 23:42:27 bdev_raid.raid_read_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:39.537 23:42:27 bdev_raid.raid_read_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:08:39.537 23:42:27 bdev_raid.raid_read_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:39.537 23:42:27 bdev_raid.raid_read_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:08:39.537 23:42:27 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:39.537 [2024-11-26 23:42:27.581977] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:08:39.537 [2024-11-26 23:42:27.582186] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid79805 ] 00:08:39.797 [2024-11-26 23:42:27.738996] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:39.797 [2024-11-26 23:42:27.766714] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:39.797 [2024-11-26 23:42:27.810188] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:39.797 [2024-11-26 23:42:27.810223] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:40.366 23:42:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:08:40.366 23:42:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@868 -- # return 0 00:08:40.366 23:42:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:08:40.366 23:42:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:08:40.367 23:42:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:40.367 23:42:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:40.367 BaseBdev1_malloc 00:08:40.367 23:42:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:40.367 23:42:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:08:40.367 23:42:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:40.367 23:42:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:40.367 true 00:08:40.367 23:42:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:40.367 23:42:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:08:40.367 23:42:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:40.367 23:42:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:40.367 [2024-11-26 23:42:28.441896] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:08:40.367 [2024-11-26 23:42:28.441953] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:40.367 [2024-11-26 23:42:28.441980] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006980 00:08:40.367 [2024-11-26 23:42:28.441990] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:40.367 [2024-11-26 23:42:28.444162] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:40.367 [2024-11-26 23:42:28.444200] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:08:40.367 BaseBdev1 00:08:40.367 23:42:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:40.367 23:42:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:08:40.367 23:42:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:08:40.367 23:42:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:40.367 23:42:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:40.367 BaseBdev2_malloc 00:08:40.367 23:42:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:40.367 23:42:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:08:40.367 23:42:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:40.367 23:42:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:40.367 true 00:08:40.367 23:42:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:40.367 23:42:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:08:40.367 23:42:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:40.367 23:42:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:40.367 [2024-11-26 23:42:28.482793] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:08:40.367 [2024-11-26 23:42:28.482842] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:40.367 [2024-11-26 23:42:28.482861] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:08:40.367 [2024-11-26 23:42:28.482877] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:40.367 [2024-11-26 23:42:28.485017] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:40.367 [2024-11-26 23:42:28.485105] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:08:40.367 BaseBdev2 00:08:40.367 23:42:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:40.367 23:42:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:08:40.367 23:42:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:08:40.367 23:42:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:40.367 23:42:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:40.627 BaseBdev3_malloc 00:08:40.627 23:42:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:40.627 23:42:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:08:40.627 23:42:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:40.627 23:42:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:40.627 true 00:08:40.627 23:42:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:40.627 23:42:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:08:40.627 23:42:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:40.627 23:42:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:40.627 [2024-11-26 23:42:28.523478] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:08:40.627 [2024-11-26 23:42:28.523523] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:40.627 [2024-11-26 23:42:28.523541] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008780 00:08:40.627 [2024-11-26 23:42:28.523550] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:40.627 [2024-11-26 23:42:28.525557] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:40.627 [2024-11-26 23:42:28.525592] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:08:40.627 BaseBdev3 00:08:40.627 23:42:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:40.627 23:42:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n raid_bdev1 -s 00:08:40.627 23:42:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:40.627 23:42:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:40.627 [2024-11-26 23:42:28.535514] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:40.627 [2024-11-26 23:42:28.537305] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:40.627 [2024-11-26 23:42:28.537391] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:40.627 [2024-11-26 23:42:28.537569] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:08:40.627 [2024-11-26 23:42:28.537583] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:08:40.627 [2024-11-26 23:42:28.537845] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002bb0 00:08:40.627 [2024-11-26 23:42:28.537993] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:08:40.627 [2024-11-26 23:42:28.538001] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001c80 00:08:40.627 [2024-11-26 23:42:28.538111] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:40.627 23:42:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:40.627 23:42:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:08:40.627 23:42:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:40.627 23:42:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:40.627 23:42:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:40.627 23:42:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:40.627 23:42:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:40.627 23:42:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:40.627 23:42:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:40.627 23:42:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:40.627 23:42:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:40.627 23:42:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:40.627 23:42:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:40.627 23:42:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:40.627 23:42:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:40.627 23:42:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:40.627 23:42:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:40.627 "name": "raid_bdev1", 00:08:40.627 "uuid": "f460e6d5-112b-4d61-9c6e-bb64cbfa4466", 00:08:40.627 "strip_size_kb": 0, 00:08:40.627 "state": "online", 00:08:40.627 "raid_level": "raid1", 00:08:40.627 "superblock": true, 00:08:40.627 "num_base_bdevs": 3, 00:08:40.627 "num_base_bdevs_discovered": 3, 00:08:40.627 "num_base_bdevs_operational": 3, 00:08:40.627 "base_bdevs_list": [ 00:08:40.627 { 00:08:40.627 "name": "BaseBdev1", 00:08:40.627 "uuid": "a9fac88d-8ab9-59cc-a8c1-d43c85358b6d", 00:08:40.627 "is_configured": true, 00:08:40.627 "data_offset": 2048, 00:08:40.627 "data_size": 63488 00:08:40.627 }, 00:08:40.627 { 00:08:40.627 "name": "BaseBdev2", 00:08:40.627 "uuid": "975cf39a-4f89-5278-a409-dc6cb12f9c91", 00:08:40.627 "is_configured": true, 00:08:40.627 "data_offset": 2048, 00:08:40.627 "data_size": 63488 00:08:40.627 }, 00:08:40.627 { 00:08:40.627 "name": "BaseBdev3", 00:08:40.627 "uuid": "0672a031-868c-5cc3-864f-2ec34c1ea007", 00:08:40.627 "is_configured": true, 00:08:40.627 "data_offset": 2048, 00:08:40.627 "data_size": 63488 00:08:40.627 } 00:08:40.627 ] 00:08:40.627 }' 00:08:40.627 23:42:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:40.627 23:42:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:40.887 23:42:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:08:40.887 23:42:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:08:41.147 [2024-11-26 23:42:29.075094] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002d50 00:08:42.091 23:42:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:08:42.091 23:42:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:42.091 23:42:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:42.091 23:42:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:42.091 23:42:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:08:42.092 23:42:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid1 = \r\a\i\d\1 ]] 00:08:42.092 23:42:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ read = \w\r\i\t\e ]] 00:08:42.092 23:42:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=3 00:08:42.092 23:42:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:08:42.092 23:42:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:42.092 23:42:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:42.092 23:42:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:42.092 23:42:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:42.092 23:42:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:42.092 23:42:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:42.092 23:42:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:42.092 23:42:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:42.092 23:42:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:42.092 23:42:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:42.092 23:42:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:42.092 23:42:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:42.092 23:42:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:42.092 23:42:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:42.092 23:42:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:42.092 "name": "raid_bdev1", 00:08:42.092 "uuid": "f460e6d5-112b-4d61-9c6e-bb64cbfa4466", 00:08:42.092 "strip_size_kb": 0, 00:08:42.092 "state": "online", 00:08:42.092 "raid_level": "raid1", 00:08:42.092 "superblock": true, 00:08:42.092 "num_base_bdevs": 3, 00:08:42.092 "num_base_bdevs_discovered": 3, 00:08:42.092 "num_base_bdevs_operational": 3, 00:08:42.092 "base_bdevs_list": [ 00:08:42.092 { 00:08:42.092 "name": "BaseBdev1", 00:08:42.092 "uuid": "a9fac88d-8ab9-59cc-a8c1-d43c85358b6d", 00:08:42.092 "is_configured": true, 00:08:42.092 "data_offset": 2048, 00:08:42.092 "data_size": 63488 00:08:42.092 }, 00:08:42.092 { 00:08:42.092 "name": "BaseBdev2", 00:08:42.092 "uuid": "975cf39a-4f89-5278-a409-dc6cb12f9c91", 00:08:42.092 "is_configured": true, 00:08:42.092 "data_offset": 2048, 00:08:42.092 "data_size": 63488 00:08:42.092 }, 00:08:42.092 { 00:08:42.092 "name": "BaseBdev3", 00:08:42.092 "uuid": "0672a031-868c-5cc3-864f-2ec34c1ea007", 00:08:42.092 "is_configured": true, 00:08:42.092 "data_offset": 2048, 00:08:42.092 "data_size": 63488 00:08:42.092 } 00:08:42.092 ] 00:08:42.092 }' 00:08:42.092 23:42:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:42.092 23:42:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:42.352 23:42:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:08:42.352 23:42:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:42.352 23:42:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:42.352 [2024-11-26 23:42:30.472617] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:08:42.352 [2024-11-26 23:42:30.472720] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:42.352 [2024-11-26 23:42:30.475461] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:42.352 [2024-11-26 23:42:30.475562] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:42.352 [2024-11-26 23:42:30.475679] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:42.352 [2024-11-26 23:42:30.475734] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name raid_bdev1, state offline 00:08:42.352 { 00:08:42.352 "results": [ 00:08:42.352 { 00:08:42.352 "job": "raid_bdev1", 00:08:42.352 "core_mask": "0x1", 00:08:42.352 "workload": "randrw", 00:08:42.352 "percentage": 50, 00:08:42.352 "status": "finished", 00:08:42.352 "queue_depth": 1, 00:08:42.352 "io_size": 131072, 00:08:42.352 "runtime": 1.398468, 00:08:42.352 "iops": 14186.953151591599, 00:08:42.352 "mibps": 1773.3691439489498, 00:08:42.352 "io_failed": 0, 00:08:42.352 "io_timeout": 0, 00:08:42.352 "avg_latency_us": 67.86202845471193, 00:08:42.352 "min_latency_us": 22.358078602620086, 00:08:42.352 "max_latency_us": 1395.1441048034935 00:08:42.352 } 00:08:42.352 ], 00:08:42.352 "core_count": 1 00:08:42.352 } 00:08:42.352 23:42:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:42.352 23:42:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 79805 00:08:42.352 23:42:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # '[' -z 79805 ']' 00:08:42.352 23:42:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@958 -- # kill -0 79805 00:08:42.612 23:42:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # uname 00:08:42.612 23:42:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:08:42.612 23:42:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 79805 00:08:42.612 23:42:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:08:42.612 killing process with pid 79805 00:08:42.612 23:42:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:08:42.612 23:42:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 79805' 00:08:42.612 23:42:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@973 -- # kill 79805 00:08:42.612 [2024-11-26 23:42:30.522178] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:42.612 23:42:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@978 -- # wait 79805 00:08:42.612 [2024-11-26 23:42:30.548148] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:42.885 23:42:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.j4tTDtxpD4 00:08:42.885 23:42:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:08:42.885 23:42:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:08:42.885 23:42:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.00 00:08:42.885 23:42:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid1 00:08:42.885 ************************************ 00:08:42.885 END TEST raid_read_error_test 00:08:42.885 ************************************ 00:08:42.885 23:42:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:08:42.885 23:42:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@199 -- # return 0 00:08:42.885 23:42:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@847 -- # [[ 0.00 = \0\.\0\0 ]] 00:08:42.885 00:08:42.885 real 0m3.278s 00:08:42.885 user 0m4.195s 00:08:42.885 sys 0m0.519s 00:08:42.885 23:42:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:08:42.885 23:42:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:42.885 23:42:30 bdev_raid -- bdev/bdev_raid.sh@972 -- # run_test raid_write_error_test raid_io_error_test raid1 3 write 00:08:42.885 23:42:30 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:08:42.885 23:42:30 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:08:42.885 23:42:30 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:42.885 ************************************ 00:08:42.885 START TEST raid_write_error_test 00:08:42.885 ************************************ 00:08:42.885 23:42:30 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test raid1 3 write 00:08:42.885 23:42:30 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid1 00:08:42.886 23:42:30 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=3 00:08:42.886 23:42:30 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=write 00:08:42.886 23:42:30 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:08:42.886 23:42:30 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:42.886 23:42:30 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:08:42.886 23:42:30 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:08:42.886 23:42:30 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:42.886 23:42:30 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:08:42.886 23:42:30 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:08:42.886 23:42:30 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:42.886 23:42:30 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:08:42.886 23:42:30 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:08:42.886 23:42:30 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:42.886 23:42:30 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:08:42.886 23:42:30 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:08:42.886 23:42:30 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:08:42.886 23:42:30 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:08:42.886 23:42:30 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:08:42.886 23:42:30 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:08:42.886 23:42:30 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:08:42.886 23:42:30 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid1 '!=' raid1 ']' 00:08:42.886 23:42:30 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@804 -- # strip_size=0 00:08:42.886 23:42:30 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:08:42.886 23:42:30 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.9w7OvwTcJf 00:08:42.886 23:42:30 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=79938 00:08:42.887 23:42:30 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:08:42.887 23:42:30 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 79938 00:08:42.887 23:42:30 bdev_raid.raid_write_error_test -- common/autotest_common.sh@835 -- # '[' -z 79938 ']' 00:08:42.887 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:42.887 23:42:30 bdev_raid.raid_write_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:42.887 23:42:30 bdev_raid.raid_write_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:08:42.887 23:42:30 bdev_raid.raid_write_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:42.887 23:42:30 bdev_raid.raid_write_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:08:42.887 23:42:30 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:42.887 [2024-11-26 23:42:30.929026] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:08:42.887 [2024-11-26 23:42:30.929230] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid79938 ] 00:08:43.152 [2024-11-26 23:42:31.085038] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:43.152 [2024-11-26 23:42:31.110658] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:43.152 [2024-11-26 23:42:31.154190] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:43.152 [2024-11-26 23:42:31.154225] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:43.722 23:42:31 bdev_raid.raid_write_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:08:43.722 23:42:31 bdev_raid.raid_write_error_test -- common/autotest_common.sh@868 -- # return 0 00:08:43.722 23:42:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:08:43.722 23:42:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:08:43.722 23:42:31 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:43.722 23:42:31 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:43.722 BaseBdev1_malloc 00:08:43.722 23:42:31 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:43.722 23:42:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:08:43.722 23:42:31 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:43.722 23:42:31 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:43.722 true 00:08:43.722 23:42:31 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:43.722 23:42:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:08:43.722 23:42:31 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:43.722 23:42:31 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:43.722 [2024-11-26 23:42:31.785863] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:08:43.722 [2024-11-26 23:42:31.785922] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:43.722 [2024-11-26 23:42:31.785964] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006980 00:08:43.722 [2024-11-26 23:42:31.785973] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:43.722 [2024-11-26 23:42:31.788107] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:43.722 [2024-11-26 23:42:31.788140] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:08:43.722 BaseBdev1 00:08:43.722 23:42:31 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:43.722 23:42:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:08:43.722 23:42:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:08:43.722 23:42:31 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:43.722 23:42:31 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:43.722 BaseBdev2_malloc 00:08:43.722 23:42:31 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:43.722 23:42:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:08:43.722 23:42:31 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:43.722 23:42:31 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:43.722 true 00:08:43.722 23:42:31 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:43.722 23:42:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:08:43.722 23:42:31 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:43.722 23:42:31 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:43.722 [2024-11-26 23:42:31.826450] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:08:43.722 [2024-11-26 23:42:31.826494] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:43.722 [2024-11-26 23:42:31.826510] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:08:43.722 [2024-11-26 23:42:31.826526] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:43.722 [2024-11-26 23:42:31.828551] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:43.722 [2024-11-26 23:42:31.828588] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:08:43.722 BaseBdev2 00:08:43.722 23:42:31 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:43.722 23:42:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:08:43.722 23:42:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:08:43.722 23:42:31 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:43.722 23:42:31 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:43.722 BaseBdev3_malloc 00:08:43.722 23:42:31 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:43.722 23:42:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:08:43.722 23:42:31 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:43.722 23:42:31 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:43.982 true 00:08:43.982 23:42:31 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:43.982 23:42:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:08:43.982 23:42:31 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:43.982 23:42:31 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:43.982 [2024-11-26 23:42:31.866911] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:08:43.982 [2024-11-26 23:42:31.866956] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:43.982 [2024-11-26 23:42:31.866975] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008780 00:08:43.982 [2024-11-26 23:42:31.866983] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:43.982 [2024-11-26 23:42:31.868985] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:43.982 [2024-11-26 23:42:31.869019] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:08:43.982 BaseBdev3 00:08:43.983 23:42:31 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:43.983 23:42:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n raid_bdev1 -s 00:08:43.983 23:42:31 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:43.983 23:42:31 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:43.983 [2024-11-26 23:42:31.878935] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:43.983 [2024-11-26 23:42:31.880802] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:43.983 [2024-11-26 23:42:31.880871] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:43.983 [2024-11-26 23:42:31.881036] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:08:43.983 [2024-11-26 23:42:31.881055] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:08:43.983 [2024-11-26 23:42:31.881286] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002bb0 00:08:43.983 [2024-11-26 23:42:31.881458] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:08:43.983 [2024-11-26 23:42:31.881476] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001c80 00:08:43.983 [2024-11-26 23:42:31.881627] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:43.983 23:42:31 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:43.983 23:42:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:08:43.983 23:42:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:43.983 23:42:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:43.983 23:42:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:43.983 23:42:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:43.983 23:42:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:43.983 23:42:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:43.983 23:42:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:43.983 23:42:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:43.983 23:42:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:43.983 23:42:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:43.983 23:42:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:43.983 23:42:31 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:43.983 23:42:31 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:43.983 23:42:31 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:43.983 23:42:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:43.983 "name": "raid_bdev1", 00:08:43.983 "uuid": "c8812250-8331-4bba-9939-49070a4726ca", 00:08:43.983 "strip_size_kb": 0, 00:08:43.983 "state": "online", 00:08:43.983 "raid_level": "raid1", 00:08:43.983 "superblock": true, 00:08:43.983 "num_base_bdevs": 3, 00:08:43.983 "num_base_bdevs_discovered": 3, 00:08:43.983 "num_base_bdevs_operational": 3, 00:08:43.983 "base_bdevs_list": [ 00:08:43.983 { 00:08:43.983 "name": "BaseBdev1", 00:08:43.983 "uuid": "3ebdbb5e-1cc9-5464-8eb0-da5c02254cca", 00:08:43.983 "is_configured": true, 00:08:43.983 "data_offset": 2048, 00:08:43.983 "data_size": 63488 00:08:43.983 }, 00:08:43.983 { 00:08:43.983 "name": "BaseBdev2", 00:08:43.983 "uuid": "e4cdba85-c806-5515-87fa-4d3173f2252a", 00:08:43.983 "is_configured": true, 00:08:43.983 "data_offset": 2048, 00:08:43.983 "data_size": 63488 00:08:43.983 }, 00:08:43.983 { 00:08:43.983 "name": "BaseBdev3", 00:08:43.983 "uuid": "0d411c20-6bf9-5e6d-8331-5f664a8a990d", 00:08:43.983 "is_configured": true, 00:08:43.983 "data_offset": 2048, 00:08:43.983 "data_size": 63488 00:08:43.983 } 00:08:43.983 ] 00:08:43.983 }' 00:08:43.983 23:42:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:43.983 23:42:31 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:44.241 23:42:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:08:44.241 23:42:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:08:44.500 [2024-11-26 23:42:32.410444] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002d50 00:08:45.437 23:42:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:08:45.437 23:42:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:45.437 23:42:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:45.437 [2024-11-26 23:42:33.345700] bdev_raid.c:2276:_raid_bdev_fail_base_bdev: *NOTICE*: Failing base bdev in slot 0 ('BaseBdev1') of raid bdev 'raid_bdev1' 00:08:45.437 [2024-11-26 23:42:33.345845] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:08:45.437 [2024-11-26 23:42:33.346095] bdev_raid.c:1974:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 0 raid_ch: 0x60d000002d50 00:08:45.437 23:42:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:45.437 23:42:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:08:45.437 23:42:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid1 = \r\a\i\d\1 ]] 00:08:45.437 23:42:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ write = \w\r\i\t\e ]] 00:08:45.437 23:42:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@833 -- # expected_num_base_bdevs=2 00:08:45.437 23:42:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:08:45.437 23:42:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:45.437 23:42:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:45.437 23:42:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:45.437 23:42:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:45.437 23:42:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:45.437 23:42:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:45.437 23:42:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:45.437 23:42:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:45.437 23:42:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:45.437 23:42:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:45.437 23:42:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:45.437 23:42:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:45.437 23:42:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:45.437 23:42:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:45.437 23:42:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:45.437 "name": "raid_bdev1", 00:08:45.437 "uuid": "c8812250-8331-4bba-9939-49070a4726ca", 00:08:45.437 "strip_size_kb": 0, 00:08:45.437 "state": "online", 00:08:45.437 "raid_level": "raid1", 00:08:45.437 "superblock": true, 00:08:45.437 "num_base_bdevs": 3, 00:08:45.437 "num_base_bdevs_discovered": 2, 00:08:45.437 "num_base_bdevs_operational": 2, 00:08:45.437 "base_bdevs_list": [ 00:08:45.437 { 00:08:45.437 "name": null, 00:08:45.437 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:45.437 "is_configured": false, 00:08:45.437 "data_offset": 0, 00:08:45.437 "data_size": 63488 00:08:45.437 }, 00:08:45.437 { 00:08:45.437 "name": "BaseBdev2", 00:08:45.437 "uuid": "e4cdba85-c806-5515-87fa-4d3173f2252a", 00:08:45.437 "is_configured": true, 00:08:45.437 "data_offset": 2048, 00:08:45.437 "data_size": 63488 00:08:45.437 }, 00:08:45.437 { 00:08:45.437 "name": "BaseBdev3", 00:08:45.437 "uuid": "0d411c20-6bf9-5e6d-8331-5f664a8a990d", 00:08:45.437 "is_configured": true, 00:08:45.437 "data_offset": 2048, 00:08:45.437 "data_size": 63488 00:08:45.437 } 00:08:45.437 ] 00:08:45.437 }' 00:08:45.437 23:42:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:45.437 23:42:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:46.006 23:42:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:08:46.006 23:42:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:46.006 23:42:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:46.006 [2024-11-26 23:42:33.836098] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:08:46.006 [2024-11-26 23:42:33.836186] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:46.006 [2024-11-26 23:42:33.838706] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:46.006 [2024-11-26 23:42:33.838802] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:46.006 [2024-11-26 23:42:33.838912] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:46.006 [2024-11-26 23:42:33.838973] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name raid_bdev1, state offline 00:08:46.006 { 00:08:46.006 "results": [ 00:08:46.006 { 00:08:46.006 "job": "raid_bdev1", 00:08:46.006 "core_mask": "0x1", 00:08:46.006 "workload": "randrw", 00:08:46.006 "percentage": 50, 00:08:46.006 "status": "finished", 00:08:46.006 "queue_depth": 1, 00:08:46.006 "io_size": 131072, 00:08:46.006 "runtime": 1.426608, 00:08:46.006 "iops": 16138.98141605823, 00:08:46.006 "mibps": 2017.3726770072788, 00:08:46.006 "io_failed": 0, 00:08:46.006 "io_timeout": 0, 00:08:46.006 "avg_latency_us": 59.330029648196984, 00:08:46.006 "min_latency_us": 22.46986899563319, 00:08:46.006 "max_latency_us": 1480.9991266375546 00:08:46.006 } 00:08:46.006 ], 00:08:46.006 "core_count": 1 00:08:46.006 } 00:08:46.006 23:42:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:46.006 23:42:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 79938 00:08:46.006 23:42:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # '[' -z 79938 ']' 00:08:46.006 23:42:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@958 -- # kill -0 79938 00:08:46.006 23:42:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # uname 00:08:46.006 23:42:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:08:46.006 23:42:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 79938 00:08:46.006 23:42:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:08:46.006 23:42:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:08:46.006 23:42:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 79938' 00:08:46.006 killing process with pid 79938 00:08:46.006 23:42:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@973 -- # kill 79938 00:08:46.006 [2024-11-26 23:42:33.885250] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:46.006 23:42:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@978 -- # wait 79938 00:08:46.006 [2024-11-26 23:42:33.910422] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:46.006 23:42:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.9w7OvwTcJf 00:08:46.006 23:42:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:08:46.006 23:42:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:08:46.006 23:42:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.00 00:08:46.006 23:42:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid1 00:08:46.006 23:42:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:08:46.006 23:42:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@199 -- # return 0 00:08:46.006 23:42:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@847 -- # [[ 0.00 = \0\.\0\0 ]] 00:08:46.006 ************************************ 00:08:46.006 END TEST raid_write_error_test 00:08:46.006 ************************************ 00:08:46.006 00:08:46.006 real 0m3.297s 00:08:46.006 user 0m4.206s 00:08:46.006 sys 0m0.529s 00:08:46.006 23:42:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:08:46.006 23:42:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:46.265 23:42:34 bdev_raid -- bdev/bdev_raid.sh@966 -- # for n in {2..4} 00:08:46.265 23:42:34 bdev_raid -- bdev/bdev_raid.sh@967 -- # for level in raid0 concat raid1 00:08:46.265 23:42:34 bdev_raid -- bdev/bdev_raid.sh@968 -- # run_test raid_state_function_test raid_state_function_test raid0 4 false 00:08:46.265 23:42:34 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:08:46.265 23:42:34 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:08:46.265 23:42:34 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:46.265 ************************************ 00:08:46.265 START TEST raid_state_function_test 00:08:46.265 ************************************ 00:08:46.265 23:42:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1129 -- # raid_state_function_test raid0 4 false 00:08:46.265 23:42:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=raid0 00:08:46.265 23:42:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=4 00:08:46.265 23:42:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:08:46.265 23:42:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:08:46.265 23:42:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:08:46.265 23:42:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:46.265 23:42:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:08:46.265 23:42:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:46.265 23:42:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:46.265 23:42:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:08:46.265 23:42:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:46.265 23:42:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:46.265 23:42:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:08:46.265 23:42:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:46.265 23:42:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:46.265 23:42:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev4 00:08:46.265 23:42:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:46.265 23:42:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:46.265 23:42:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:08:46.265 23:42:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:08:46.265 23:42:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:08:46.265 23:42:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:08:46.266 23:42:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:08:46.266 23:42:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:08:46.266 23:42:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' raid0 '!=' raid1 ']' 00:08:46.266 23:42:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:08:46.266 23:42:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:08:46.266 23:42:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:08:46.266 23:42:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:08:46.266 23:42:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=80066 00:08:46.266 23:42:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:08:46.266 23:42:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 80066' 00:08:46.266 Process raid pid: 80066 00:08:46.266 23:42:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 80066 00:08:46.266 23:42:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@835 -- # '[' -z 80066 ']' 00:08:46.266 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:46.266 23:42:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:46.266 23:42:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:08:46.266 23:42:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:46.266 23:42:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:08:46.266 23:42:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:46.266 [2024-11-26 23:42:34.288442] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:08:46.266 [2024-11-26 23:42:34.288565] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:08:46.532 [2024-11-26 23:42:34.443140] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:46.532 [2024-11-26 23:42:34.468896] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:46.532 [2024-11-26 23:42:34.511775] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:46.532 [2024-11-26 23:42:34.511902] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:47.111 23:42:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:08:47.111 23:42:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@868 -- # return 0 00:08:47.111 23:42:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:08:47.111 23:42:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:47.111 23:42:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:47.111 [2024-11-26 23:42:35.138976] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:47.111 [2024-11-26 23:42:35.139037] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:47.111 [2024-11-26 23:42:35.139049] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:47.111 [2024-11-26 23:42:35.139060] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:47.111 [2024-11-26 23:42:35.139067] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:08:47.111 [2024-11-26 23:42:35.139079] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:08:47.111 [2024-11-26 23:42:35.139086] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:08:47.111 [2024-11-26 23:42:35.139096] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:08:47.111 23:42:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:47.111 23:42:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:08:47.111 23:42:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:47.111 23:42:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:47.111 23:42:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:47.111 23:42:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:47.111 23:42:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:08:47.111 23:42:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:47.111 23:42:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:47.111 23:42:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:47.111 23:42:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:47.111 23:42:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:47.111 23:42:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:47.111 23:42:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:47.111 23:42:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:47.111 23:42:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:47.112 23:42:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:47.112 "name": "Existed_Raid", 00:08:47.112 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:47.112 "strip_size_kb": 64, 00:08:47.112 "state": "configuring", 00:08:47.112 "raid_level": "raid0", 00:08:47.112 "superblock": false, 00:08:47.112 "num_base_bdevs": 4, 00:08:47.112 "num_base_bdevs_discovered": 0, 00:08:47.112 "num_base_bdevs_operational": 4, 00:08:47.112 "base_bdevs_list": [ 00:08:47.112 { 00:08:47.112 "name": "BaseBdev1", 00:08:47.112 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:47.112 "is_configured": false, 00:08:47.112 "data_offset": 0, 00:08:47.112 "data_size": 0 00:08:47.112 }, 00:08:47.112 { 00:08:47.112 "name": "BaseBdev2", 00:08:47.112 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:47.112 "is_configured": false, 00:08:47.112 "data_offset": 0, 00:08:47.112 "data_size": 0 00:08:47.112 }, 00:08:47.112 { 00:08:47.112 "name": "BaseBdev3", 00:08:47.112 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:47.112 "is_configured": false, 00:08:47.112 "data_offset": 0, 00:08:47.112 "data_size": 0 00:08:47.112 }, 00:08:47.112 { 00:08:47.112 "name": "BaseBdev4", 00:08:47.112 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:47.112 "is_configured": false, 00:08:47.112 "data_offset": 0, 00:08:47.112 "data_size": 0 00:08:47.112 } 00:08:47.112 ] 00:08:47.112 }' 00:08:47.112 23:42:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:47.112 23:42:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:47.682 23:42:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:47.682 23:42:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:47.682 23:42:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:47.682 [2024-11-26 23:42:35.542233] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:47.682 [2024-11-26 23:42:35.542278] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:08:47.682 23:42:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:47.682 23:42:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:08:47.682 23:42:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:47.682 23:42:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:47.682 [2024-11-26 23:42:35.554240] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:47.682 [2024-11-26 23:42:35.554291] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:47.682 [2024-11-26 23:42:35.554300] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:47.682 [2024-11-26 23:42:35.554326] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:47.682 [2024-11-26 23:42:35.554333] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:08:47.682 [2024-11-26 23:42:35.554342] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:08:47.682 [2024-11-26 23:42:35.554348] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:08:47.682 [2024-11-26 23:42:35.554371] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:08:47.682 23:42:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:47.682 23:42:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:08:47.682 23:42:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:47.682 23:42:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:47.682 [2024-11-26 23:42:35.575236] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:47.682 BaseBdev1 00:08:47.682 23:42:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:47.682 23:42:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:08:47.682 23:42:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:08:47.682 23:42:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:47.682 23:42:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:08:47.682 23:42:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:47.682 23:42:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:47.682 23:42:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:47.682 23:42:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:47.682 23:42:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:47.682 23:42:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:47.682 23:42:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:08:47.682 23:42:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:47.682 23:42:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:47.682 [ 00:08:47.682 { 00:08:47.682 "name": "BaseBdev1", 00:08:47.682 "aliases": [ 00:08:47.682 "222a6d5b-e61b-4edb-aedc-509c5946c419" 00:08:47.682 ], 00:08:47.682 "product_name": "Malloc disk", 00:08:47.682 "block_size": 512, 00:08:47.682 "num_blocks": 65536, 00:08:47.682 "uuid": "222a6d5b-e61b-4edb-aedc-509c5946c419", 00:08:47.682 "assigned_rate_limits": { 00:08:47.682 "rw_ios_per_sec": 0, 00:08:47.682 "rw_mbytes_per_sec": 0, 00:08:47.682 "r_mbytes_per_sec": 0, 00:08:47.682 "w_mbytes_per_sec": 0 00:08:47.682 }, 00:08:47.682 "claimed": true, 00:08:47.682 "claim_type": "exclusive_write", 00:08:47.682 "zoned": false, 00:08:47.682 "supported_io_types": { 00:08:47.682 "read": true, 00:08:47.682 "write": true, 00:08:47.682 "unmap": true, 00:08:47.682 "flush": true, 00:08:47.682 "reset": true, 00:08:47.682 "nvme_admin": false, 00:08:47.682 "nvme_io": false, 00:08:47.682 "nvme_io_md": false, 00:08:47.682 "write_zeroes": true, 00:08:47.682 "zcopy": true, 00:08:47.682 "get_zone_info": false, 00:08:47.682 "zone_management": false, 00:08:47.682 "zone_append": false, 00:08:47.682 "compare": false, 00:08:47.682 "compare_and_write": false, 00:08:47.682 "abort": true, 00:08:47.682 "seek_hole": false, 00:08:47.682 "seek_data": false, 00:08:47.682 "copy": true, 00:08:47.682 "nvme_iov_md": false 00:08:47.682 }, 00:08:47.682 "memory_domains": [ 00:08:47.682 { 00:08:47.682 "dma_device_id": "system", 00:08:47.682 "dma_device_type": 1 00:08:47.682 }, 00:08:47.682 { 00:08:47.682 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:47.682 "dma_device_type": 2 00:08:47.682 } 00:08:47.682 ], 00:08:47.682 "driver_specific": {} 00:08:47.682 } 00:08:47.682 ] 00:08:47.682 23:42:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:47.682 23:42:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:08:47.682 23:42:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:08:47.682 23:42:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:47.682 23:42:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:47.682 23:42:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:47.682 23:42:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:47.682 23:42:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:08:47.682 23:42:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:47.682 23:42:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:47.682 23:42:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:47.682 23:42:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:47.683 23:42:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:47.683 23:42:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:47.683 23:42:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:47.683 23:42:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:47.683 23:42:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:47.683 23:42:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:47.683 "name": "Existed_Raid", 00:08:47.683 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:47.683 "strip_size_kb": 64, 00:08:47.683 "state": "configuring", 00:08:47.683 "raid_level": "raid0", 00:08:47.683 "superblock": false, 00:08:47.683 "num_base_bdevs": 4, 00:08:47.683 "num_base_bdevs_discovered": 1, 00:08:47.683 "num_base_bdevs_operational": 4, 00:08:47.683 "base_bdevs_list": [ 00:08:47.683 { 00:08:47.683 "name": "BaseBdev1", 00:08:47.683 "uuid": "222a6d5b-e61b-4edb-aedc-509c5946c419", 00:08:47.683 "is_configured": true, 00:08:47.683 "data_offset": 0, 00:08:47.683 "data_size": 65536 00:08:47.683 }, 00:08:47.683 { 00:08:47.683 "name": "BaseBdev2", 00:08:47.683 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:47.683 "is_configured": false, 00:08:47.683 "data_offset": 0, 00:08:47.683 "data_size": 0 00:08:47.683 }, 00:08:47.683 { 00:08:47.683 "name": "BaseBdev3", 00:08:47.683 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:47.683 "is_configured": false, 00:08:47.683 "data_offset": 0, 00:08:47.683 "data_size": 0 00:08:47.683 }, 00:08:47.683 { 00:08:47.683 "name": "BaseBdev4", 00:08:47.683 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:47.683 "is_configured": false, 00:08:47.683 "data_offset": 0, 00:08:47.683 "data_size": 0 00:08:47.683 } 00:08:47.683 ] 00:08:47.683 }' 00:08:47.683 23:42:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:47.683 23:42:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:47.944 23:42:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:47.944 23:42:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:47.944 23:42:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:47.944 [2024-11-26 23:42:36.034491] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:47.944 [2024-11-26 23:42:36.034551] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:08:47.944 23:42:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:47.944 23:42:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:08:47.944 23:42:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:47.944 23:42:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:47.944 [2024-11-26 23:42:36.042520] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:47.944 [2024-11-26 23:42:36.044456] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:47.944 [2024-11-26 23:42:36.044502] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:47.944 [2024-11-26 23:42:36.044537] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:08:47.944 [2024-11-26 23:42:36.044561] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:08:47.944 [2024-11-26 23:42:36.044587] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:08:47.944 [2024-11-26 23:42:36.044611] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:08:47.944 23:42:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:47.944 23:42:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:08:47.944 23:42:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:47.944 23:42:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:08:47.944 23:42:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:47.944 23:42:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:47.944 23:42:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:47.944 23:42:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:47.944 23:42:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:08:47.944 23:42:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:47.944 23:42:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:47.944 23:42:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:47.944 23:42:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:47.944 23:42:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:47.944 23:42:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:47.944 23:42:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:47.944 23:42:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:47.944 23:42:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:48.202 23:42:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:48.202 "name": "Existed_Raid", 00:08:48.202 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:48.202 "strip_size_kb": 64, 00:08:48.202 "state": "configuring", 00:08:48.202 "raid_level": "raid0", 00:08:48.202 "superblock": false, 00:08:48.202 "num_base_bdevs": 4, 00:08:48.202 "num_base_bdevs_discovered": 1, 00:08:48.202 "num_base_bdevs_operational": 4, 00:08:48.202 "base_bdevs_list": [ 00:08:48.202 { 00:08:48.202 "name": "BaseBdev1", 00:08:48.202 "uuid": "222a6d5b-e61b-4edb-aedc-509c5946c419", 00:08:48.202 "is_configured": true, 00:08:48.202 "data_offset": 0, 00:08:48.202 "data_size": 65536 00:08:48.202 }, 00:08:48.202 { 00:08:48.202 "name": "BaseBdev2", 00:08:48.202 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:48.202 "is_configured": false, 00:08:48.202 "data_offset": 0, 00:08:48.202 "data_size": 0 00:08:48.202 }, 00:08:48.202 { 00:08:48.202 "name": "BaseBdev3", 00:08:48.202 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:48.202 "is_configured": false, 00:08:48.202 "data_offset": 0, 00:08:48.202 "data_size": 0 00:08:48.202 }, 00:08:48.202 { 00:08:48.202 "name": "BaseBdev4", 00:08:48.202 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:48.202 "is_configured": false, 00:08:48.202 "data_offset": 0, 00:08:48.202 "data_size": 0 00:08:48.202 } 00:08:48.202 ] 00:08:48.202 }' 00:08:48.202 23:42:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:48.203 23:42:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:48.463 23:42:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:08:48.463 23:42:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:48.463 23:42:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:48.463 [2024-11-26 23:42:36.508793] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:48.463 BaseBdev2 00:08:48.463 23:42:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:48.463 23:42:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:08:48.463 23:42:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:08:48.463 23:42:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:48.463 23:42:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:08:48.463 23:42:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:48.463 23:42:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:48.463 23:42:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:48.463 23:42:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:48.463 23:42:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:48.463 23:42:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:48.463 23:42:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:08:48.463 23:42:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:48.463 23:42:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:48.463 [ 00:08:48.463 { 00:08:48.463 "name": "BaseBdev2", 00:08:48.463 "aliases": [ 00:08:48.463 "fbb90953-43dc-46f5-a0d4-d11c3ffc464f" 00:08:48.463 ], 00:08:48.463 "product_name": "Malloc disk", 00:08:48.463 "block_size": 512, 00:08:48.463 "num_blocks": 65536, 00:08:48.463 "uuid": "fbb90953-43dc-46f5-a0d4-d11c3ffc464f", 00:08:48.463 "assigned_rate_limits": { 00:08:48.463 "rw_ios_per_sec": 0, 00:08:48.463 "rw_mbytes_per_sec": 0, 00:08:48.463 "r_mbytes_per_sec": 0, 00:08:48.463 "w_mbytes_per_sec": 0 00:08:48.463 }, 00:08:48.463 "claimed": true, 00:08:48.463 "claim_type": "exclusive_write", 00:08:48.463 "zoned": false, 00:08:48.463 "supported_io_types": { 00:08:48.463 "read": true, 00:08:48.463 "write": true, 00:08:48.463 "unmap": true, 00:08:48.463 "flush": true, 00:08:48.463 "reset": true, 00:08:48.463 "nvme_admin": false, 00:08:48.463 "nvme_io": false, 00:08:48.463 "nvme_io_md": false, 00:08:48.463 "write_zeroes": true, 00:08:48.463 "zcopy": true, 00:08:48.463 "get_zone_info": false, 00:08:48.463 "zone_management": false, 00:08:48.463 "zone_append": false, 00:08:48.463 "compare": false, 00:08:48.463 "compare_and_write": false, 00:08:48.463 "abort": true, 00:08:48.463 "seek_hole": false, 00:08:48.463 "seek_data": false, 00:08:48.463 "copy": true, 00:08:48.463 "nvme_iov_md": false 00:08:48.463 }, 00:08:48.463 "memory_domains": [ 00:08:48.463 { 00:08:48.463 "dma_device_id": "system", 00:08:48.463 "dma_device_type": 1 00:08:48.463 }, 00:08:48.463 { 00:08:48.463 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:48.463 "dma_device_type": 2 00:08:48.463 } 00:08:48.463 ], 00:08:48.463 "driver_specific": {} 00:08:48.463 } 00:08:48.463 ] 00:08:48.463 23:42:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:48.463 23:42:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:08:48.463 23:42:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:08:48.463 23:42:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:48.463 23:42:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:08:48.463 23:42:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:48.463 23:42:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:48.463 23:42:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:48.463 23:42:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:48.463 23:42:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:08:48.463 23:42:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:48.463 23:42:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:48.463 23:42:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:48.463 23:42:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:48.463 23:42:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:48.463 23:42:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:48.463 23:42:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:48.463 23:42:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:48.463 23:42:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:48.723 23:42:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:48.723 "name": "Existed_Raid", 00:08:48.723 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:48.723 "strip_size_kb": 64, 00:08:48.723 "state": "configuring", 00:08:48.723 "raid_level": "raid0", 00:08:48.723 "superblock": false, 00:08:48.723 "num_base_bdevs": 4, 00:08:48.723 "num_base_bdevs_discovered": 2, 00:08:48.723 "num_base_bdevs_operational": 4, 00:08:48.723 "base_bdevs_list": [ 00:08:48.723 { 00:08:48.723 "name": "BaseBdev1", 00:08:48.723 "uuid": "222a6d5b-e61b-4edb-aedc-509c5946c419", 00:08:48.723 "is_configured": true, 00:08:48.723 "data_offset": 0, 00:08:48.723 "data_size": 65536 00:08:48.723 }, 00:08:48.723 { 00:08:48.723 "name": "BaseBdev2", 00:08:48.723 "uuid": "fbb90953-43dc-46f5-a0d4-d11c3ffc464f", 00:08:48.723 "is_configured": true, 00:08:48.723 "data_offset": 0, 00:08:48.723 "data_size": 65536 00:08:48.723 }, 00:08:48.723 { 00:08:48.723 "name": "BaseBdev3", 00:08:48.723 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:48.723 "is_configured": false, 00:08:48.723 "data_offset": 0, 00:08:48.723 "data_size": 0 00:08:48.723 }, 00:08:48.723 { 00:08:48.723 "name": "BaseBdev4", 00:08:48.723 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:48.723 "is_configured": false, 00:08:48.723 "data_offset": 0, 00:08:48.723 "data_size": 0 00:08:48.723 } 00:08:48.723 ] 00:08:48.723 }' 00:08:48.723 23:42:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:48.723 23:42:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:48.983 23:42:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:08:48.983 23:42:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:48.983 23:42:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:48.983 [2024-11-26 23:42:37.018253] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:48.983 BaseBdev3 00:08:48.983 23:42:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:48.983 23:42:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:08:48.983 23:42:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:08:48.983 23:42:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:48.983 23:42:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:08:48.983 23:42:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:48.983 23:42:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:48.983 23:42:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:48.983 23:42:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:48.983 23:42:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:48.983 23:42:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:48.983 23:42:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:08:48.983 23:42:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:48.983 23:42:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:48.983 [ 00:08:48.983 { 00:08:48.983 "name": "BaseBdev3", 00:08:48.983 "aliases": [ 00:08:48.983 "67393eb3-1124-4718-93ef-d97763d4c543" 00:08:48.983 ], 00:08:48.983 "product_name": "Malloc disk", 00:08:48.983 "block_size": 512, 00:08:48.983 "num_blocks": 65536, 00:08:48.983 "uuid": "67393eb3-1124-4718-93ef-d97763d4c543", 00:08:48.983 "assigned_rate_limits": { 00:08:48.983 "rw_ios_per_sec": 0, 00:08:48.983 "rw_mbytes_per_sec": 0, 00:08:48.983 "r_mbytes_per_sec": 0, 00:08:48.983 "w_mbytes_per_sec": 0 00:08:48.983 }, 00:08:48.983 "claimed": true, 00:08:48.983 "claim_type": "exclusive_write", 00:08:48.983 "zoned": false, 00:08:48.983 "supported_io_types": { 00:08:48.983 "read": true, 00:08:48.983 "write": true, 00:08:48.983 "unmap": true, 00:08:48.983 "flush": true, 00:08:48.983 "reset": true, 00:08:48.983 "nvme_admin": false, 00:08:48.983 "nvme_io": false, 00:08:48.983 "nvme_io_md": false, 00:08:48.983 "write_zeroes": true, 00:08:48.983 "zcopy": true, 00:08:48.983 "get_zone_info": false, 00:08:48.983 "zone_management": false, 00:08:48.983 "zone_append": false, 00:08:48.983 "compare": false, 00:08:48.983 "compare_and_write": false, 00:08:48.983 "abort": true, 00:08:48.983 "seek_hole": false, 00:08:48.983 "seek_data": false, 00:08:48.983 "copy": true, 00:08:48.983 "nvme_iov_md": false 00:08:48.983 }, 00:08:48.983 "memory_domains": [ 00:08:48.983 { 00:08:48.983 "dma_device_id": "system", 00:08:48.983 "dma_device_type": 1 00:08:48.983 }, 00:08:48.983 { 00:08:48.983 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:48.983 "dma_device_type": 2 00:08:48.983 } 00:08:48.983 ], 00:08:48.983 "driver_specific": {} 00:08:48.983 } 00:08:48.983 ] 00:08:48.983 23:42:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:48.983 23:42:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:08:48.983 23:42:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:08:48.983 23:42:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:48.983 23:42:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:08:48.983 23:42:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:48.983 23:42:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:48.983 23:42:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:48.983 23:42:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:48.983 23:42:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:08:48.983 23:42:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:48.983 23:42:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:48.983 23:42:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:48.983 23:42:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:48.983 23:42:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:48.983 23:42:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:48.983 23:42:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:48.983 23:42:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:48.983 23:42:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:48.983 23:42:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:48.983 "name": "Existed_Raid", 00:08:48.983 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:48.983 "strip_size_kb": 64, 00:08:48.983 "state": "configuring", 00:08:48.983 "raid_level": "raid0", 00:08:48.983 "superblock": false, 00:08:48.983 "num_base_bdevs": 4, 00:08:48.983 "num_base_bdevs_discovered": 3, 00:08:48.983 "num_base_bdevs_operational": 4, 00:08:48.983 "base_bdevs_list": [ 00:08:48.983 { 00:08:48.983 "name": "BaseBdev1", 00:08:48.983 "uuid": "222a6d5b-e61b-4edb-aedc-509c5946c419", 00:08:48.983 "is_configured": true, 00:08:48.983 "data_offset": 0, 00:08:48.983 "data_size": 65536 00:08:48.983 }, 00:08:48.983 { 00:08:48.983 "name": "BaseBdev2", 00:08:48.983 "uuid": "fbb90953-43dc-46f5-a0d4-d11c3ffc464f", 00:08:48.983 "is_configured": true, 00:08:48.983 "data_offset": 0, 00:08:48.983 "data_size": 65536 00:08:48.983 }, 00:08:48.983 { 00:08:48.983 "name": "BaseBdev3", 00:08:48.983 "uuid": "67393eb3-1124-4718-93ef-d97763d4c543", 00:08:48.983 "is_configured": true, 00:08:48.983 "data_offset": 0, 00:08:48.983 "data_size": 65536 00:08:48.983 }, 00:08:48.983 { 00:08:48.983 "name": "BaseBdev4", 00:08:48.983 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:48.983 "is_configured": false, 00:08:48.983 "data_offset": 0, 00:08:48.983 "data_size": 0 00:08:48.983 } 00:08:48.983 ] 00:08:48.983 }' 00:08:49.243 23:42:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:49.243 23:42:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:49.501 23:42:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:08:49.501 23:42:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:49.501 23:42:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:49.501 [2024-11-26 23:42:37.508624] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:08:49.501 [2024-11-26 23:42:37.508755] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:08:49.501 [2024-11-26 23:42:37.508770] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 262144, blocklen 512 00:08:49.501 [2024-11-26 23:42:37.509090] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:08:49.501 [2024-11-26 23:42:37.509245] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:08:49.501 [2024-11-26 23:42:37.509258] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:08:49.501 [2024-11-26 23:42:37.509485] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:49.501 BaseBdev4 00:08:49.501 23:42:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:49.501 23:42:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev4 00:08:49.501 23:42:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:08:49.501 23:42:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:49.501 23:42:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:08:49.501 23:42:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:49.501 23:42:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:49.501 23:42:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:49.501 23:42:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:49.501 23:42:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:49.501 23:42:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:49.501 23:42:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:08:49.501 23:42:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:49.501 23:42:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:49.501 [ 00:08:49.501 { 00:08:49.501 "name": "BaseBdev4", 00:08:49.501 "aliases": [ 00:08:49.501 "971436c1-0382-49bd-9685-e9bab5ea4d96" 00:08:49.501 ], 00:08:49.501 "product_name": "Malloc disk", 00:08:49.501 "block_size": 512, 00:08:49.501 "num_blocks": 65536, 00:08:49.501 "uuid": "971436c1-0382-49bd-9685-e9bab5ea4d96", 00:08:49.501 "assigned_rate_limits": { 00:08:49.501 "rw_ios_per_sec": 0, 00:08:49.501 "rw_mbytes_per_sec": 0, 00:08:49.501 "r_mbytes_per_sec": 0, 00:08:49.501 "w_mbytes_per_sec": 0 00:08:49.501 }, 00:08:49.501 "claimed": true, 00:08:49.501 "claim_type": "exclusive_write", 00:08:49.501 "zoned": false, 00:08:49.501 "supported_io_types": { 00:08:49.501 "read": true, 00:08:49.501 "write": true, 00:08:49.501 "unmap": true, 00:08:49.501 "flush": true, 00:08:49.501 "reset": true, 00:08:49.501 "nvme_admin": false, 00:08:49.501 "nvme_io": false, 00:08:49.501 "nvme_io_md": false, 00:08:49.501 "write_zeroes": true, 00:08:49.501 "zcopy": true, 00:08:49.501 "get_zone_info": false, 00:08:49.501 "zone_management": false, 00:08:49.501 "zone_append": false, 00:08:49.501 "compare": false, 00:08:49.501 "compare_and_write": false, 00:08:49.501 "abort": true, 00:08:49.501 "seek_hole": false, 00:08:49.501 "seek_data": false, 00:08:49.501 "copy": true, 00:08:49.501 "nvme_iov_md": false 00:08:49.501 }, 00:08:49.501 "memory_domains": [ 00:08:49.501 { 00:08:49.501 "dma_device_id": "system", 00:08:49.501 "dma_device_type": 1 00:08:49.501 }, 00:08:49.501 { 00:08:49.501 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:49.501 "dma_device_type": 2 00:08:49.501 } 00:08:49.501 ], 00:08:49.501 "driver_specific": {} 00:08:49.501 } 00:08:49.501 ] 00:08:49.501 23:42:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:49.501 23:42:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:08:49.501 23:42:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:08:49.501 23:42:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:49.501 23:42:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid0 64 4 00:08:49.501 23:42:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:49.501 23:42:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:49.501 23:42:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:49.501 23:42:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:49.501 23:42:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:08:49.501 23:42:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:49.501 23:42:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:49.501 23:42:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:49.501 23:42:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:49.501 23:42:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:49.501 23:42:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:49.501 23:42:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:49.501 23:42:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:49.501 23:42:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:49.501 23:42:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:49.501 "name": "Existed_Raid", 00:08:49.501 "uuid": "87b3fc29-c933-4411-b205-2d349cdacac1", 00:08:49.501 "strip_size_kb": 64, 00:08:49.501 "state": "online", 00:08:49.501 "raid_level": "raid0", 00:08:49.501 "superblock": false, 00:08:49.501 "num_base_bdevs": 4, 00:08:49.501 "num_base_bdevs_discovered": 4, 00:08:49.501 "num_base_bdevs_operational": 4, 00:08:49.501 "base_bdevs_list": [ 00:08:49.501 { 00:08:49.501 "name": "BaseBdev1", 00:08:49.501 "uuid": "222a6d5b-e61b-4edb-aedc-509c5946c419", 00:08:49.501 "is_configured": true, 00:08:49.501 "data_offset": 0, 00:08:49.501 "data_size": 65536 00:08:49.501 }, 00:08:49.501 { 00:08:49.501 "name": "BaseBdev2", 00:08:49.501 "uuid": "fbb90953-43dc-46f5-a0d4-d11c3ffc464f", 00:08:49.501 "is_configured": true, 00:08:49.501 "data_offset": 0, 00:08:49.501 "data_size": 65536 00:08:49.501 }, 00:08:49.501 { 00:08:49.501 "name": "BaseBdev3", 00:08:49.501 "uuid": "67393eb3-1124-4718-93ef-d97763d4c543", 00:08:49.501 "is_configured": true, 00:08:49.501 "data_offset": 0, 00:08:49.501 "data_size": 65536 00:08:49.501 }, 00:08:49.501 { 00:08:49.501 "name": "BaseBdev4", 00:08:49.501 "uuid": "971436c1-0382-49bd-9685-e9bab5ea4d96", 00:08:49.501 "is_configured": true, 00:08:49.501 "data_offset": 0, 00:08:49.501 "data_size": 65536 00:08:49.501 } 00:08:49.501 ] 00:08:49.501 }' 00:08:49.501 23:42:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:49.501 23:42:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:50.070 23:42:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:08:50.070 23:42:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:08:50.070 23:42:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:50.070 23:42:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:50.070 23:42:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:08:50.070 23:42:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:50.070 23:42:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:50.070 23:42:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:08:50.070 23:42:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:50.070 23:42:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:50.070 [2024-11-26 23:42:38.024130] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:50.070 23:42:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:50.070 23:42:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:50.070 "name": "Existed_Raid", 00:08:50.070 "aliases": [ 00:08:50.070 "87b3fc29-c933-4411-b205-2d349cdacac1" 00:08:50.070 ], 00:08:50.070 "product_name": "Raid Volume", 00:08:50.070 "block_size": 512, 00:08:50.070 "num_blocks": 262144, 00:08:50.070 "uuid": "87b3fc29-c933-4411-b205-2d349cdacac1", 00:08:50.070 "assigned_rate_limits": { 00:08:50.070 "rw_ios_per_sec": 0, 00:08:50.070 "rw_mbytes_per_sec": 0, 00:08:50.070 "r_mbytes_per_sec": 0, 00:08:50.070 "w_mbytes_per_sec": 0 00:08:50.070 }, 00:08:50.070 "claimed": false, 00:08:50.070 "zoned": false, 00:08:50.070 "supported_io_types": { 00:08:50.070 "read": true, 00:08:50.070 "write": true, 00:08:50.070 "unmap": true, 00:08:50.070 "flush": true, 00:08:50.070 "reset": true, 00:08:50.070 "nvme_admin": false, 00:08:50.070 "nvme_io": false, 00:08:50.070 "nvme_io_md": false, 00:08:50.070 "write_zeroes": true, 00:08:50.070 "zcopy": false, 00:08:50.070 "get_zone_info": false, 00:08:50.070 "zone_management": false, 00:08:50.070 "zone_append": false, 00:08:50.070 "compare": false, 00:08:50.070 "compare_and_write": false, 00:08:50.070 "abort": false, 00:08:50.070 "seek_hole": false, 00:08:50.070 "seek_data": false, 00:08:50.070 "copy": false, 00:08:50.070 "nvme_iov_md": false 00:08:50.070 }, 00:08:50.070 "memory_domains": [ 00:08:50.070 { 00:08:50.070 "dma_device_id": "system", 00:08:50.070 "dma_device_type": 1 00:08:50.070 }, 00:08:50.070 { 00:08:50.070 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:50.070 "dma_device_type": 2 00:08:50.070 }, 00:08:50.070 { 00:08:50.070 "dma_device_id": "system", 00:08:50.070 "dma_device_type": 1 00:08:50.070 }, 00:08:50.070 { 00:08:50.070 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:50.070 "dma_device_type": 2 00:08:50.070 }, 00:08:50.070 { 00:08:50.070 "dma_device_id": "system", 00:08:50.070 "dma_device_type": 1 00:08:50.070 }, 00:08:50.070 { 00:08:50.070 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:50.070 "dma_device_type": 2 00:08:50.070 }, 00:08:50.070 { 00:08:50.070 "dma_device_id": "system", 00:08:50.070 "dma_device_type": 1 00:08:50.070 }, 00:08:50.070 { 00:08:50.070 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:50.070 "dma_device_type": 2 00:08:50.070 } 00:08:50.070 ], 00:08:50.070 "driver_specific": { 00:08:50.070 "raid": { 00:08:50.070 "uuid": "87b3fc29-c933-4411-b205-2d349cdacac1", 00:08:50.070 "strip_size_kb": 64, 00:08:50.070 "state": "online", 00:08:50.070 "raid_level": "raid0", 00:08:50.070 "superblock": false, 00:08:50.070 "num_base_bdevs": 4, 00:08:50.070 "num_base_bdevs_discovered": 4, 00:08:50.070 "num_base_bdevs_operational": 4, 00:08:50.070 "base_bdevs_list": [ 00:08:50.070 { 00:08:50.070 "name": "BaseBdev1", 00:08:50.070 "uuid": "222a6d5b-e61b-4edb-aedc-509c5946c419", 00:08:50.070 "is_configured": true, 00:08:50.070 "data_offset": 0, 00:08:50.071 "data_size": 65536 00:08:50.071 }, 00:08:50.071 { 00:08:50.071 "name": "BaseBdev2", 00:08:50.071 "uuid": "fbb90953-43dc-46f5-a0d4-d11c3ffc464f", 00:08:50.071 "is_configured": true, 00:08:50.071 "data_offset": 0, 00:08:50.071 "data_size": 65536 00:08:50.071 }, 00:08:50.071 { 00:08:50.071 "name": "BaseBdev3", 00:08:50.071 "uuid": "67393eb3-1124-4718-93ef-d97763d4c543", 00:08:50.071 "is_configured": true, 00:08:50.071 "data_offset": 0, 00:08:50.071 "data_size": 65536 00:08:50.071 }, 00:08:50.071 { 00:08:50.071 "name": "BaseBdev4", 00:08:50.071 "uuid": "971436c1-0382-49bd-9685-e9bab5ea4d96", 00:08:50.071 "is_configured": true, 00:08:50.071 "data_offset": 0, 00:08:50.071 "data_size": 65536 00:08:50.071 } 00:08:50.071 ] 00:08:50.071 } 00:08:50.071 } 00:08:50.071 }' 00:08:50.071 23:42:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:50.071 23:42:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:08:50.071 BaseBdev2 00:08:50.071 BaseBdev3 00:08:50.071 BaseBdev4' 00:08:50.071 23:42:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:50.071 23:42:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:50.071 23:42:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:50.071 23:42:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:08:50.071 23:42:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:50.071 23:42:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:50.071 23:42:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:50.071 23:42:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:50.071 23:42:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:50.071 23:42:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:50.071 23:42:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:50.071 23:42:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:08:50.071 23:42:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:50.071 23:42:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:50.071 23:42:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:50.330 23:42:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:50.330 23:42:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:50.330 23:42:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:50.330 23:42:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:50.330 23:42:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:08:50.330 23:42:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:50.330 23:42:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:50.330 23:42:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:50.330 23:42:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:50.330 23:42:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:50.330 23:42:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:50.330 23:42:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:50.330 23:42:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:08:50.330 23:42:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:50.330 23:42:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:50.330 23:42:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:50.330 23:42:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:50.330 23:42:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:50.330 23:42:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:50.330 23:42:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:08:50.330 23:42:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:50.330 23:42:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:50.330 [2024-11-26 23:42:38.343290] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:08:50.330 [2024-11-26 23:42:38.343382] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:50.330 [2024-11-26 23:42:38.343447] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:50.330 23:42:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:50.330 23:42:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:08:50.330 23:42:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy raid0 00:08:50.330 23:42:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:08:50.330 23:42:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # return 1 00:08:50.330 23:42:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:08:50.330 23:42:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline raid0 64 3 00:08:50.330 23:42:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:50.330 23:42:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:08:50.330 23:42:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:50.330 23:42:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:50.330 23:42:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:50.330 23:42:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:50.330 23:42:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:50.330 23:42:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:50.330 23:42:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:50.330 23:42:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:50.330 23:42:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:50.330 23:42:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:50.330 23:42:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:50.330 23:42:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:50.330 23:42:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:50.330 "name": "Existed_Raid", 00:08:50.330 "uuid": "87b3fc29-c933-4411-b205-2d349cdacac1", 00:08:50.330 "strip_size_kb": 64, 00:08:50.330 "state": "offline", 00:08:50.330 "raid_level": "raid0", 00:08:50.330 "superblock": false, 00:08:50.330 "num_base_bdevs": 4, 00:08:50.330 "num_base_bdevs_discovered": 3, 00:08:50.330 "num_base_bdevs_operational": 3, 00:08:50.330 "base_bdevs_list": [ 00:08:50.330 { 00:08:50.330 "name": null, 00:08:50.330 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:50.330 "is_configured": false, 00:08:50.330 "data_offset": 0, 00:08:50.330 "data_size": 65536 00:08:50.330 }, 00:08:50.330 { 00:08:50.330 "name": "BaseBdev2", 00:08:50.330 "uuid": "fbb90953-43dc-46f5-a0d4-d11c3ffc464f", 00:08:50.330 "is_configured": true, 00:08:50.330 "data_offset": 0, 00:08:50.330 "data_size": 65536 00:08:50.330 }, 00:08:50.330 { 00:08:50.330 "name": "BaseBdev3", 00:08:50.330 "uuid": "67393eb3-1124-4718-93ef-d97763d4c543", 00:08:50.330 "is_configured": true, 00:08:50.330 "data_offset": 0, 00:08:50.330 "data_size": 65536 00:08:50.330 }, 00:08:50.330 { 00:08:50.331 "name": "BaseBdev4", 00:08:50.331 "uuid": "971436c1-0382-49bd-9685-e9bab5ea4d96", 00:08:50.331 "is_configured": true, 00:08:50.331 "data_offset": 0, 00:08:50.331 "data_size": 65536 00:08:50.331 } 00:08:50.331 ] 00:08:50.331 }' 00:08:50.331 23:42:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:50.331 23:42:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:50.900 23:42:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:08:50.900 23:42:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:50.900 23:42:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:08:50.900 23:42:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:50.900 23:42:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:50.900 23:42:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:50.900 23:42:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:50.900 23:42:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:08:50.900 23:42:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:08:50.900 23:42:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:08:50.900 23:42:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:50.900 23:42:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:50.900 [2024-11-26 23:42:38.857736] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:08:50.900 23:42:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:50.900 23:42:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:08:50.900 23:42:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:50.900 23:42:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:08:50.900 23:42:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:50.900 23:42:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:50.900 23:42:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:50.900 23:42:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:50.900 23:42:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:08:50.900 23:42:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:08:50.900 23:42:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:08:50.900 23:42:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:50.900 23:42:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:50.900 [2024-11-26 23:42:38.928751] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:08:50.900 23:42:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:50.900 23:42:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:08:50.900 23:42:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:50.900 23:42:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:08:50.900 23:42:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:50.900 23:42:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:50.900 23:42:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:50.900 23:42:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:50.900 23:42:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:08:50.900 23:42:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:08:50.900 23:42:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev4 00:08:50.900 23:42:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:50.900 23:42:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:50.900 [2024-11-26 23:42:38.995836] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev4 00:08:50.900 [2024-11-26 23:42:38.995928] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:08:50.900 23:42:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:50.900 23:42:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:08:50.900 23:42:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:50.900 23:42:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:08:50.900 23:42:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:50.900 23:42:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:50.900 23:42:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:50.900 23:42:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:51.160 23:42:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:08:51.160 23:42:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:08:51.160 23:42:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 4 -gt 2 ']' 00:08:51.160 23:42:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:08:51.160 23:42:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:08:51.160 23:42:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:08:51.160 23:42:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:51.160 23:42:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:51.160 BaseBdev2 00:08:51.160 23:42:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:51.160 23:42:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:08:51.160 23:42:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:08:51.160 23:42:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:51.160 23:42:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:08:51.160 23:42:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:51.160 23:42:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:51.160 23:42:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:51.160 23:42:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:51.160 23:42:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:51.160 23:42:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:51.160 23:42:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:08:51.160 23:42:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:51.160 23:42:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:51.160 [ 00:08:51.160 { 00:08:51.160 "name": "BaseBdev2", 00:08:51.160 "aliases": [ 00:08:51.160 "f214a47e-6afc-47f2-b397-a265571a3fd2" 00:08:51.160 ], 00:08:51.160 "product_name": "Malloc disk", 00:08:51.160 "block_size": 512, 00:08:51.160 "num_blocks": 65536, 00:08:51.160 "uuid": "f214a47e-6afc-47f2-b397-a265571a3fd2", 00:08:51.160 "assigned_rate_limits": { 00:08:51.160 "rw_ios_per_sec": 0, 00:08:51.160 "rw_mbytes_per_sec": 0, 00:08:51.160 "r_mbytes_per_sec": 0, 00:08:51.160 "w_mbytes_per_sec": 0 00:08:51.160 }, 00:08:51.160 "claimed": false, 00:08:51.160 "zoned": false, 00:08:51.160 "supported_io_types": { 00:08:51.160 "read": true, 00:08:51.160 "write": true, 00:08:51.160 "unmap": true, 00:08:51.160 "flush": true, 00:08:51.160 "reset": true, 00:08:51.160 "nvme_admin": false, 00:08:51.160 "nvme_io": false, 00:08:51.160 "nvme_io_md": false, 00:08:51.160 "write_zeroes": true, 00:08:51.160 "zcopy": true, 00:08:51.160 "get_zone_info": false, 00:08:51.160 "zone_management": false, 00:08:51.160 "zone_append": false, 00:08:51.160 "compare": false, 00:08:51.160 "compare_and_write": false, 00:08:51.160 "abort": true, 00:08:51.160 "seek_hole": false, 00:08:51.160 "seek_data": false, 00:08:51.160 "copy": true, 00:08:51.160 "nvme_iov_md": false 00:08:51.160 }, 00:08:51.160 "memory_domains": [ 00:08:51.160 { 00:08:51.160 "dma_device_id": "system", 00:08:51.160 "dma_device_type": 1 00:08:51.160 }, 00:08:51.160 { 00:08:51.160 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:51.160 "dma_device_type": 2 00:08:51.160 } 00:08:51.160 ], 00:08:51.160 "driver_specific": {} 00:08:51.160 } 00:08:51.160 ] 00:08:51.160 23:42:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:51.160 23:42:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:08:51.160 23:42:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:08:51.160 23:42:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:08:51.160 23:42:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:08:51.160 23:42:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:51.160 23:42:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:51.160 BaseBdev3 00:08:51.160 23:42:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:51.160 23:42:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:08:51.160 23:42:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:08:51.160 23:42:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:51.160 23:42:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:08:51.160 23:42:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:51.160 23:42:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:51.160 23:42:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:51.160 23:42:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:51.160 23:42:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:51.160 23:42:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:51.160 23:42:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:08:51.160 23:42:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:51.160 23:42:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:51.160 [ 00:08:51.160 { 00:08:51.160 "name": "BaseBdev3", 00:08:51.160 "aliases": [ 00:08:51.160 "77ce42e3-9160-4528-b363-e54f9b8f8da6" 00:08:51.160 ], 00:08:51.160 "product_name": "Malloc disk", 00:08:51.160 "block_size": 512, 00:08:51.161 "num_blocks": 65536, 00:08:51.161 "uuid": "77ce42e3-9160-4528-b363-e54f9b8f8da6", 00:08:51.161 "assigned_rate_limits": { 00:08:51.161 "rw_ios_per_sec": 0, 00:08:51.161 "rw_mbytes_per_sec": 0, 00:08:51.161 "r_mbytes_per_sec": 0, 00:08:51.161 "w_mbytes_per_sec": 0 00:08:51.161 }, 00:08:51.161 "claimed": false, 00:08:51.161 "zoned": false, 00:08:51.161 "supported_io_types": { 00:08:51.161 "read": true, 00:08:51.161 "write": true, 00:08:51.161 "unmap": true, 00:08:51.161 "flush": true, 00:08:51.161 "reset": true, 00:08:51.161 "nvme_admin": false, 00:08:51.161 "nvme_io": false, 00:08:51.161 "nvme_io_md": false, 00:08:51.161 "write_zeroes": true, 00:08:51.161 "zcopy": true, 00:08:51.161 "get_zone_info": false, 00:08:51.161 "zone_management": false, 00:08:51.161 "zone_append": false, 00:08:51.161 "compare": false, 00:08:51.161 "compare_and_write": false, 00:08:51.161 "abort": true, 00:08:51.161 "seek_hole": false, 00:08:51.161 "seek_data": false, 00:08:51.161 "copy": true, 00:08:51.161 "nvme_iov_md": false 00:08:51.161 }, 00:08:51.161 "memory_domains": [ 00:08:51.161 { 00:08:51.161 "dma_device_id": "system", 00:08:51.161 "dma_device_type": 1 00:08:51.161 }, 00:08:51.161 { 00:08:51.161 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:51.161 "dma_device_type": 2 00:08:51.161 } 00:08:51.161 ], 00:08:51.161 "driver_specific": {} 00:08:51.161 } 00:08:51.161 ] 00:08:51.161 23:42:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:51.161 23:42:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:08:51.161 23:42:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:08:51.161 23:42:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:08:51.161 23:42:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:08:51.161 23:42:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:51.161 23:42:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:51.161 BaseBdev4 00:08:51.161 23:42:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:51.161 23:42:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev4 00:08:51.161 23:42:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:08:51.161 23:42:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:51.161 23:42:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:08:51.161 23:42:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:51.161 23:42:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:51.161 23:42:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:51.161 23:42:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:51.161 23:42:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:51.161 23:42:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:51.161 23:42:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:08:51.161 23:42:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:51.161 23:42:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:51.161 [ 00:08:51.161 { 00:08:51.161 "name": "BaseBdev4", 00:08:51.161 "aliases": [ 00:08:51.161 "604be6d1-6b0e-4805-9e72-d2730983fdaf" 00:08:51.161 ], 00:08:51.161 "product_name": "Malloc disk", 00:08:51.161 "block_size": 512, 00:08:51.161 "num_blocks": 65536, 00:08:51.161 "uuid": "604be6d1-6b0e-4805-9e72-d2730983fdaf", 00:08:51.161 "assigned_rate_limits": { 00:08:51.161 "rw_ios_per_sec": 0, 00:08:51.161 "rw_mbytes_per_sec": 0, 00:08:51.161 "r_mbytes_per_sec": 0, 00:08:51.161 "w_mbytes_per_sec": 0 00:08:51.161 }, 00:08:51.161 "claimed": false, 00:08:51.161 "zoned": false, 00:08:51.161 "supported_io_types": { 00:08:51.161 "read": true, 00:08:51.161 "write": true, 00:08:51.161 "unmap": true, 00:08:51.161 "flush": true, 00:08:51.161 "reset": true, 00:08:51.161 "nvme_admin": false, 00:08:51.161 "nvme_io": false, 00:08:51.161 "nvme_io_md": false, 00:08:51.161 "write_zeroes": true, 00:08:51.161 "zcopy": true, 00:08:51.161 "get_zone_info": false, 00:08:51.161 "zone_management": false, 00:08:51.161 "zone_append": false, 00:08:51.161 "compare": false, 00:08:51.161 "compare_and_write": false, 00:08:51.161 "abort": true, 00:08:51.161 "seek_hole": false, 00:08:51.161 "seek_data": false, 00:08:51.161 "copy": true, 00:08:51.161 "nvme_iov_md": false 00:08:51.161 }, 00:08:51.161 "memory_domains": [ 00:08:51.161 { 00:08:51.161 "dma_device_id": "system", 00:08:51.161 "dma_device_type": 1 00:08:51.161 }, 00:08:51.161 { 00:08:51.161 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:51.161 "dma_device_type": 2 00:08:51.161 } 00:08:51.161 ], 00:08:51.161 "driver_specific": {} 00:08:51.161 } 00:08:51.161 ] 00:08:51.161 23:42:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:51.161 23:42:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:08:51.161 23:42:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:08:51.161 23:42:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:08:51.161 23:42:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:08:51.161 23:42:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:51.161 23:42:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:51.161 [2024-11-26 23:42:39.223830] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:51.161 [2024-11-26 23:42:39.223934] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:51.161 [2024-11-26 23:42:39.223993] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:51.161 [2024-11-26 23:42:39.225772] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:51.161 [2024-11-26 23:42:39.225854] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:08:51.161 23:42:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:51.161 23:42:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:08:51.161 23:42:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:51.161 23:42:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:51.161 23:42:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:51.161 23:42:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:51.161 23:42:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:08:51.161 23:42:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:51.161 23:42:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:51.161 23:42:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:51.161 23:42:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:51.161 23:42:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:51.161 23:42:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:51.161 23:42:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:51.161 23:42:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:51.161 23:42:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:51.161 23:42:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:51.161 "name": "Existed_Raid", 00:08:51.161 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:51.161 "strip_size_kb": 64, 00:08:51.161 "state": "configuring", 00:08:51.161 "raid_level": "raid0", 00:08:51.161 "superblock": false, 00:08:51.161 "num_base_bdevs": 4, 00:08:51.161 "num_base_bdevs_discovered": 3, 00:08:51.161 "num_base_bdevs_operational": 4, 00:08:51.161 "base_bdevs_list": [ 00:08:51.161 { 00:08:51.161 "name": "BaseBdev1", 00:08:51.161 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:51.161 "is_configured": false, 00:08:51.161 "data_offset": 0, 00:08:51.161 "data_size": 0 00:08:51.161 }, 00:08:51.161 { 00:08:51.161 "name": "BaseBdev2", 00:08:51.161 "uuid": "f214a47e-6afc-47f2-b397-a265571a3fd2", 00:08:51.161 "is_configured": true, 00:08:51.161 "data_offset": 0, 00:08:51.161 "data_size": 65536 00:08:51.161 }, 00:08:51.161 { 00:08:51.161 "name": "BaseBdev3", 00:08:51.161 "uuid": "77ce42e3-9160-4528-b363-e54f9b8f8da6", 00:08:51.161 "is_configured": true, 00:08:51.161 "data_offset": 0, 00:08:51.161 "data_size": 65536 00:08:51.161 }, 00:08:51.161 { 00:08:51.161 "name": "BaseBdev4", 00:08:51.161 "uuid": "604be6d1-6b0e-4805-9e72-d2730983fdaf", 00:08:51.161 "is_configured": true, 00:08:51.161 "data_offset": 0, 00:08:51.161 "data_size": 65536 00:08:51.161 } 00:08:51.161 ] 00:08:51.161 }' 00:08:51.161 23:42:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:51.161 23:42:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:51.731 23:42:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:08:51.731 23:42:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:51.731 23:42:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:51.731 [2024-11-26 23:42:39.695036] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:08:51.731 23:42:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:51.731 23:42:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:08:51.731 23:42:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:51.731 23:42:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:51.731 23:42:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:51.731 23:42:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:51.731 23:42:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:08:51.731 23:42:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:51.731 23:42:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:51.731 23:42:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:51.731 23:42:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:51.731 23:42:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:51.731 23:42:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:51.731 23:42:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:51.731 23:42:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:51.731 23:42:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:51.731 23:42:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:51.731 "name": "Existed_Raid", 00:08:51.731 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:51.731 "strip_size_kb": 64, 00:08:51.731 "state": "configuring", 00:08:51.731 "raid_level": "raid0", 00:08:51.731 "superblock": false, 00:08:51.731 "num_base_bdevs": 4, 00:08:51.731 "num_base_bdevs_discovered": 2, 00:08:51.731 "num_base_bdevs_operational": 4, 00:08:51.731 "base_bdevs_list": [ 00:08:51.731 { 00:08:51.731 "name": "BaseBdev1", 00:08:51.731 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:51.731 "is_configured": false, 00:08:51.731 "data_offset": 0, 00:08:51.731 "data_size": 0 00:08:51.731 }, 00:08:51.731 { 00:08:51.731 "name": null, 00:08:51.731 "uuid": "f214a47e-6afc-47f2-b397-a265571a3fd2", 00:08:51.731 "is_configured": false, 00:08:51.731 "data_offset": 0, 00:08:51.731 "data_size": 65536 00:08:51.731 }, 00:08:51.731 { 00:08:51.731 "name": "BaseBdev3", 00:08:51.731 "uuid": "77ce42e3-9160-4528-b363-e54f9b8f8da6", 00:08:51.731 "is_configured": true, 00:08:51.731 "data_offset": 0, 00:08:51.731 "data_size": 65536 00:08:51.731 }, 00:08:51.731 { 00:08:51.731 "name": "BaseBdev4", 00:08:51.731 "uuid": "604be6d1-6b0e-4805-9e72-d2730983fdaf", 00:08:51.731 "is_configured": true, 00:08:51.731 "data_offset": 0, 00:08:51.731 "data_size": 65536 00:08:51.731 } 00:08:51.731 ] 00:08:51.731 }' 00:08:51.731 23:42:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:51.731 23:42:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:52.299 23:42:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:52.299 23:42:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:08:52.299 23:42:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:52.299 23:42:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:52.299 23:42:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:52.299 23:42:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:08:52.299 23:42:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:08:52.299 23:42:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:52.299 23:42:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:52.299 [2024-11-26 23:42:40.185284] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:52.299 BaseBdev1 00:08:52.299 23:42:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:52.299 23:42:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:08:52.299 23:42:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:08:52.300 23:42:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:52.300 23:42:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:08:52.300 23:42:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:52.300 23:42:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:52.300 23:42:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:52.300 23:42:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:52.300 23:42:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:52.300 23:42:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:52.300 23:42:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:08:52.300 23:42:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:52.300 23:42:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:52.300 [ 00:08:52.300 { 00:08:52.300 "name": "BaseBdev1", 00:08:52.300 "aliases": [ 00:08:52.300 "b9eea5fb-f21c-47bc-a8b2-79c965ee047c" 00:08:52.300 ], 00:08:52.300 "product_name": "Malloc disk", 00:08:52.300 "block_size": 512, 00:08:52.300 "num_blocks": 65536, 00:08:52.300 "uuid": "b9eea5fb-f21c-47bc-a8b2-79c965ee047c", 00:08:52.300 "assigned_rate_limits": { 00:08:52.300 "rw_ios_per_sec": 0, 00:08:52.300 "rw_mbytes_per_sec": 0, 00:08:52.300 "r_mbytes_per_sec": 0, 00:08:52.300 "w_mbytes_per_sec": 0 00:08:52.300 }, 00:08:52.300 "claimed": true, 00:08:52.300 "claim_type": "exclusive_write", 00:08:52.300 "zoned": false, 00:08:52.300 "supported_io_types": { 00:08:52.300 "read": true, 00:08:52.300 "write": true, 00:08:52.300 "unmap": true, 00:08:52.300 "flush": true, 00:08:52.300 "reset": true, 00:08:52.300 "nvme_admin": false, 00:08:52.300 "nvme_io": false, 00:08:52.300 "nvme_io_md": false, 00:08:52.300 "write_zeroes": true, 00:08:52.300 "zcopy": true, 00:08:52.300 "get_zone_info": false, 00:08:52.300 "zone_management": false, 00:08:52.300 "zone_append": false, 00:08:52.300 "compare": false, 00:08:52.300 "compare_and_write": false, 00:08:52.300 "abort": true, 00:08:52.300 "seek_hole": false, 00:08:52.300 "seek_data": false, 00:08:52.300 "copy": true, 00:08:52.300 "nvme_iov_md": false 00:08:52.300 }, 00:08:52.300 "memory_domains": [ 00:08:52.300 { 00:08:52.300 "dma_device_id": "system", 00:08:52.300 "dma_device_type": 1 00:08:52.300 }, 00:08:52.300 { 00:08:52.300 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:52.300 "dma_device_type": 2 00:08:52.300 } 00:08:52.300 ], 00:08:52.300 "driver_specific": {} 00:08:52.300 } 00:08:52.300 ] 00:08:52.300 23:42:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:52.300 23:42:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:08:52.300 23:42:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:08:52.300 23:42:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:52.300 23:42:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:52.300 23:42:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:52.300 23:42:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:52.300 23:42:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:08:52.300 23:42:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:52.300 23:42:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:52.300 23:42:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:52.300 23:42:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:52.300 23:42:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:52.300 23:42:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:52.300 23:42:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:52.300 23:42:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:52.300 23:42:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:52.300 23:42:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:52.300 "name": "Existed_Raid", 00:08:52.300 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:52.300 "strip_size_kb": 64, 00:08:52.300 "state": "configuring", 00:08:52.300 "raid_level": "raid0", 00:08:52.300 "superblock": false, 00:08:52.300 "num_base_bdevs": 4, 00:08:52.300 "num_base_bdevs_discovered": 3, 00:08:52.300 "num_base_bdevs_operational": 4, 00:08:52.300 "base_bdevs_list": [ 00:08:52.300 { 00:08:52.300 "name": "BaseBdev1", 00:08:52.300 "uuid": "b9eea5fb-f21c-47bc-a8b2-79c965ee047c", 00:08:52.300 "is_configured": true, 00:08:52.300 "data_offset": 0, 00:08:52.300 "data_size": 65536 00:08:52.300 }, 00:08:52.300 { 00:08:52.300 "name": null, 00:08:52.300 "uuid": "f214a47e-6afc-47f2-b397-a265571a3fd2", 00:08:52.300 "is_configured": false, 00:08:52.300 "data_offset": 0, 00:08:52.300 "data_size": 65536 00:08:52.300 }, 00:08:52.300 { 00:08:52.300 "name": "BaseBdev3", 00:08:52.300 "uuid": "77ce42e3-9160-4528-b363-e54f9b8f8da6", 00:08:52.300 "is_configured": true, 00:08:52.300 "data_offset": 0, 00:08:52.300 "data_size": 65536 00:08:52.300 }, 00:08:52.300 { 00:08:52.300 "name": "BaseBdev4", 00:08:52.300 "uuid": "604be6d1-6b0e-4805-9e72-d2730983fdaf", 00:08:52.300 "is_configured": true, 00:08:52.300 "data_offset": 0, 00:08:52.300 "data_size": 65536 00:08:52.300 } 00:08:52.300 ] 00:08:52.300 }' 00:08:52.300 23:42:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:52.300 23:42:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:52.560 23:42:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:08:52.560 23:42:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:52.560 23:42:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:52.560 23:42:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:52.830 23:42:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:52.830 23:42:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:08:52.830 23:42:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:08:52.830 23:42:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:52.830 23:42:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:52.830 [2024-11-26 23:42:40.724448] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:08:52.830 23:42:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:52.830 23:42:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:08:52.830 23:42:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:52.830 23:42:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:52.830 23:42:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:52.830 23:42:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:52.830 23:42:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:08:52.830 23:42:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:52.830 23:42:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:52.830 23:42:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:52.830 23:42:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:52.830 23:42:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:52.830 23:42:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:52.831 23:42:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:52.831 23:42:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:52.831 23:42:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:52.831 23:42:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:52.831 "name": "Existed_Raid", 00:08:52.831 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:52.831 "strip_size_kb": 64, 00:08:52.831 "state": "configuring", 00:08:52.831 "raid_level": "raid0", 00:08:52.831 "superblock": false, 00:08:52.831 "num_base_bdevs": 4, 00:08:52.831 "num_base_bdevs_discovered": 2, 00:08:52.831 "num_base_bdevs_operational": 4, 00:08:52.831 "base_bdevs_list": [ 00:08:52.831 { 00:08:52.831 "name": "BaseBdev1", 00:08:52.831 "uuid": "b9eea5fb-f21c-47bc-a8b2-79c965ee047c", 00:08:52.831 "is_configured": true, 00:08:52.831 "data_offset": 0, 00:08:52.831 "data_size": 65536 00:08:52.831 }, 00:08:52.831 { 00:08:52.831 "name": null, 00:08:52.831 "uuid": "f214a47e-6afc-47f2-b397-a265571a3fd2", 00:08:52.831 "is_configured": false, 00:08:52.831 "data_offset": 0, 00:08:52.831 "data_size": 65536 00:08:52.831 }, 00:08:52.831 { 00:08:52.831 "name": null, 00:08:52.831 "uuid": "77ce42e3-9160-4528-b363-e54f9b8f8da6", 00:08:52.831 "is_configured": false, 00:08:52.831 "data_offset": 0, 00:08:52.831 "data_size": 65536 00:08:52.831 }, 00:08:52.831 { 00:08:52.831 "name": "BaseBdev4", 00:08:52.831 "uuid": "604be6d1-6b0e-4805-9e72-d2730983fdaf", 00:08:52.831 "is_configured": true, 00:08:52.831 "data_offset": 0, 00:08:52.831 "data_size": 65536 00:08:52.831 } 00:08:52.831 ] 00:08:52.831 }' 00:08:52.831 23:42:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:52.831 23:42:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:53.091 23:42:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:53.091 23:42:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:08:53.091 23:42:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:53.091 23:42:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:53.351 23:42:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:53.351 23:42:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:08:53.351 23:42:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:08:53.351 23:42:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:53.351 23:42:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:53.351 [2024-11-26 23:42:41.263544] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:53.351 23:42:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:53.351 23:42:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:08:53.351 23:42:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:53.351 23:42:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:53.351 23:42:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:53.351 23:42:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:53.351 23:42:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:08:53.351 23:42:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:53.351 23:42:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:53.351 23:42:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:53.351 23:42:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:53.351 23:42:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:53.351 23:42:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:53.351 23:42:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:53.351 23:42:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:53.351 23:42:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:53.351 23:42:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:53.351 "name": "Existed_Raid", 00:08:53.351 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:53.351 "strip_size_kb": 64, 00:08:53.351 "state": "configuring", 00:08:53.351 "raid_level": "raid0", 00:08:53.351 "superblock": false, 00:08:53.351 "num_base_bdevs": 4, 00:08:53.351 "num_base_bdevs_discovered": 3, 00:08:53.351 "num_base_bdevs_operational": 4, 00:08:53.351 "base_bdevs_list": [ 00:08:53.351 { 00:08:53.351 "name": "BaseBdev1", 00:08:53.351 "uuid": "b9eea5fb-f21c-47bc-a8b2-79c965ee047c", 00:08:53.351 "is_configured": true, 00:08:53.351 "data_offset": 0, 00:08:53.351 "data_size": 65536 00:08:53.351 }, 00:08:53.351 { 00:08:53.351 "name": null, 00:08:53.351 "uuid": "f214a47e-6afc-47f2-b397-a265571a3fd2", 00:08:53.351 "is_configured": false, 00:08:53.351 "data_offset": 0, 00:08:53.352 "data_size": 65536 00:08:53.352 }, 00:08:53.352 { 00:08:53.352 "name": "BaseBdev3", 00:08:53.352 "uuid": "77ce42e3-9160-4528-b363-e54f9b8f8da6", 00:08:53.352 "is_configured": true, 00:08:53.352 "data_offset": 0, 00:08:53.352 "data_size": 65536 00:08:53.352 }, 00:08:53.352 { 00:08:53.352 "name": "BaseBdev4", 00:08:53.352 "uuid": "604be6d1-6b0e-4805-9e72-d2730983fdaf", 00:08:53.352 "is_configured": true, 00:08:53.352 "data_offset": 0, 00:08:53.352 "data_size": 65536 00:08:53.352 } 00:08:53.352 ] 00:08:53.352 }' 00:08:53.352 23:42:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:53.352 23:42:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:53.611 23:42:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:53.611 23:42:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:53.611 23:42:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:08:53.611 23:42:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:53.611 23:42:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:53.611 23:42:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:08:53.611 23:42:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:08:53.612 23:42:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:53.612 23:42:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:53.870 [2024-11-26 23:42:41.742742] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:08:53.870 23:42:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:53.870 23:42:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:08:53.870 23:42:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:53.870 23:42:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:53.870 23:42:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:53.870 23:42:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:53.870 23:42:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:08:53.870 23:42:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:53.870 23:42:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:53.870 23:42:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:53.870 23:42:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:53.870 23:42:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:53.870 23:42:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:53.870 23:42:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:53.870 23:42:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:53.870 23:42:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:53.870 23:42:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:53.870 "name": "Existed_Raid", 00:08:53.870 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:53.870 "strip_size_kb": 64, 00:08:53.871 "state": "configuring", 00:08:53.871 "raid_level": "raid0", 00:08:53.871 "superblock": false, 00:08:53.871 "num_base_bdevs": 4, 00:08:53.871 "num_base_bdevs_discovered": 2, 00:08:53.871 "num_base_bdevs_operational": 4, 00:08:53.871 "base_bdevs_list": [ 00:08:53.871 { 00:08:53.871 "name": null, 00:08:53.871 "uuid": "b9eea5fb-f21c-47bc-a8b2-79c965ee047c", 00:08:53.871 "is_configured": false, 00:08:53.871 "data_offset": 0, 00:08:53.871 "data_size": 65536 00:08:53.871 }, 00:08:53.871 { 00:08:53.871 "name": null, 00:08:53.871 "uuid": "f214a47e-6afc-47f2-b397-a265571a3fd2", 00:08:53.871 "is_configured": false, 00:08:53.871 "data_offset": 0, 00:08:53.871 "data_size": 65536 00:08:53.871 }, 00:08:53.871 { 00:08:53.871 "name": "BaseBdev3", 00:08:53.871 "uuid": "77ce42e3-9160-4528-b363-e54f9b8f8da6", 00:08:53.871 "is_configured": true, 00:08:53.871 "data_offset": 0, 00:08:53.871 "data_size": 65536 00:08:53.871 }, 00:08:53.871 { 00:08:53.871 "name": "BaseBdev4", 00:08:53.871 "uuid": "604be6d1-6b0e-4805-9e72-d2730983fdaf", 00:08:53.871 "is_configured": true, 00:08:53.871 "data_offset": 0, 00:08:53.871 "data_size": 65536 00:08:53.871 } 00:08:53.871 ] 00:08:53.871 }' 00:08:53.871 23:42:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:53.871 23:42:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:54.129 23:42:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:54.129 23:42:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:08:54.129 23:42:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:54.129 23:42:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:54.129 23:42:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:54.129 23:42:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:08:54.129 23:42:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:08:54.129 23:42:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:54.129 23:42:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:54.129 [2024-11-26 23:42:42.256377] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:54.389 23:42:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:54.389 23:42:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:08:54.389 23:42:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:54.389 23:42:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:54.389 23:42:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:54.389 23:42:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:54.389 23:42:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:08:54.389 23:42:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:54.389 23:42:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:54.389 23:42:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:54.389 23:42:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:54.389 23:42:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:54.389 23:42:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:54.389 23:42:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:54.389 23:42:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:54.389 23:42:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:54.389 23:42:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:54.389 "name": "Existed_Raid", 00:08:54.389 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:54.389 "strip_size_kb": 64, 00:08:54.389 "state": "configuring", 00:08:54.389 "raid_level": "raid0", 00:08:54.389 "superblock": false, 00:08:54.389 "num_base_bdevs": 4, 00:08:54.389 "num_base_bdevs_discovered": 3, 00:08:54.389 "num_base_bdevs_operational": 4, 00:08:54.389 "base_bdevs_list": [ 00:08:54.389 { 00:08:54.389 "name": null, 00:08:54.389 "uuid": "b9eea5fb-f21c-47bc-a8b2-79c965ee047c", 00:08:54.389 "is_configured": false, 00:08:54.389 "data_offset": 0, 00:08:54.389 "data_size": 65536 00:08:54.389 }, 00:08:54.389 { 00:08:54.389 "name": "BaseBdev2", 00:08:54.389 "uuid": "f214a47e-6afc-47f2-b397-a265571a3fd2", 00:08:54.389 "is_configured": true, 00:08:54.389 "data_offset": 0, 00:08:54.389 "data_size": 65536 00:08:54.389 }, 00:08:54.389 { 00:08:54.389 "name": "BaseBdev3", 00:08:54.389 "uuid": "77ce42e3-9160-4528-b363-e54f9b8f8da6", 00:08:54.389 "is_configured": true, 00:08:54.389 "data_offset": 0, 00:08:54.389 "data_size": 65536 00:08:54.389 }, 00:08:54.389 { 00:08:54.389 "name": "BaseBdev4", 00:08:54.389 "uuid": "604be6d1-6b0e-4805-9e72-d2730983fdaf", 00:08:54.389 "is_configured": true, 00:08:54.389 "data_offset": 0, 00:08:54.389 "data_size": 65536 00:08:54.389 } 00:08:54.389 ] 00:08:54.389 }' 00:08:54.389 23:42:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:54.389 23:42:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:54.648 23:42:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:08:54.648 23:42:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:54.648 23:42:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:54.648 23:42:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:54.648 23:42:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:54.648 23:42:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:08:54.648 23:42:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:54.648 23:42:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:54.648 23:42:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:54.648 23:42:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:08:54.648 23:42:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:54.648 23:42:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u b9eea5fb-f21c-47bc-a8b2-79c965ee047c 00:08:54.648 23:42:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:54.648 23:42:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:54.648 [2024-11-26 23:42:42.766572] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:08:54.648 [2024-11-26 23:42:42.766617] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:08:54.648 [2024-11-26 23:42:42.766624] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 262144, blocklen 512 00:08:54.648 [2024-11-26 23:42:42.766871] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002a10 00:08:54.648 [2024-11-26 23:42:42.766982] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:08:54.648 [2024-11-26 23:42:42.766992] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001c80 00:08:54.648 [2024-11-26 23:42:42.767166] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:54.648 NewBaseBdev 00:08:54.648 23:42:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:54.648 23:42:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:08:54.648 23:42:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:08:54.648 23:42:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:54.648 23:42:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:08:54.648 23:42:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:54.648 23:42:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:54.648 23:42:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:54.648 23:42:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:54.648 23:42:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:54.907 23:42:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:54.908 23:42:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:08:54.908 23:42:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:54.908 23:42:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:54.908 [ 00:08:54.908 { 00:08:54.908 "name": "NewBaseBdev", 00:08:54.908 "aliases": [ 00:08:54.908 "b9eea5fb-f21c-47bc-a8b2-79c965ee047c" 00:08:54.908 ], 00:08:54.908 "product_name": "Malloc disk", 00:08:54.908 "block_size": 512, 00:08:54.908 "num_blocks": 65536, 00:08:54.908 "uuid": "b9eea5fb-f21c-47bc-a8b2-79c965ee047c", 00:08:54.908 "assigned_rate_limits": { 00:08:54.908 "rw_ios_per_sec": 0, 00:08:54.908 "rw_mbytes_per_sec": 0, 00:08:54.908 "r_mbytes_per_sec": 0, 00:08:54.908 "w_mbytes_per_sec": 0 00:08:54.908 }, 00:08:54.908 "claimed": true, 00:08:54.908 "claim_type": "exclusive_write", 00:08:54.908 "zoned": false, 00:08:54.908 "supported_io_types": { 00:08:54.908 "read": true, 00:08:54.908 "write": true, 00:08:54.908 "unmap": true, 00:08:54.908 "flush": true, 00:08:54.908 "reset": true, 00:08:54.908 "nvme_admin": false, 00:08:54.908 "nvme_io": false, 00:08:54.908 "nvme_io_md": false, 00:08:54.908 "write_zeroes": true, 00:08:54.908 "zcopy": true, 00:08:54.908 "get_zone_info": false, 00:08:54.908 "zone_management": false, 00:08:54.908 "zone_append": false, 00:08:54.908 "compare": false, 00:08:54.908 "compare_and_write": false, 00:08:54.908 "abort": true, 00:08:54.908 "seek_hole": false, 00:08:54.908 "seek_data": false, 00:08:54.908 "copy": true, 00:08:54.908 "nvme_iov_md": false 00:08:54.908 }, 00:08:54.908 "memory_domains": [ 00:08:54.908 { 00:08:54.908 "dma_device_id": "system", 00:08:54.908 "dma_device_type": 1 00:08:54.908 }, 00:08:54.908 { 00:08:54.908 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:54.908 "dma_device_type": 2 00:08:54.908 } 00:08:54.908 ], 00:08:54.908 "driver_specific": {} 00:08:54.908 } 00:08:54.908 ] 00:08:54.908 23:42:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:54.908 23:42:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:08:54.908 23:42:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid0 64 4 00:08:54.908 23:42:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:54.908 23:42:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:54.908 23:42:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:54.908 23:42:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:54.908 23:42:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:08:54.908 23:42:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:54.908 23:42:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:54.908 23:42:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:54.908 23:42:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:54.908 23:42:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:54.908 23:42:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:54.908 23:42:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:54.908 23:42:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:54.908 23:42:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:54.908 23:42:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:54.908 "name": "Existed_Raid", 00:08:54.908 "uuid": "c70bc50f-3e0c-4894-8e66-5d15b2510ad3", 00:08:54.908 "strip_size_kb": 64, 00:08:54.908 "state": "online", 00:08:54.908 "raid_level": "raid0", 00:08:54.908 "superblock": false, 00:08:54.908 "num_base_bdevs": 4, 00:08:54.908 "num_base_bdevs_discovered": 4, 00:08:54.908 "num_base_bdevs_operational": 4, 00:08:54.908 "base_bdevs_list": [ 00:08:54.908 { 00:08:54.908 "name": "NewBaseBdev", 00:08:54.908 "uuid": "b9eea5fb-f21c-47bc-a8b2-79c965ee047c", 00:08:54.908 "is_configured": true, 00:08:54.908 "data_offset": 0, 00:08:54.908 "data_size": 65536 00:08:54.908 }, 00:08:54.908 { 00:08:54.908 "name": "BaseBdev2", 00:08:54.908 "uuid": "f214a47e-6afc-47f2-b397-a265571a3fd2", 00:08:54.908 "is_configured": true, 00:08:54.908 "data_offset": 0, 00:08:54.908 "data_size": 65536 00:08:54.908 }, 00:08:54.908 { 00:08:54.908 "name": "BaseBdev3", 00:08:54.908 "uuid": "77ce42e3-9160-4528-b363-e54f9b8f8da6", 00:08:54.908 "is_configured": true, 00:08:54.908 "data_offset": 0, 00:08:54.908 "data_size": 65536 00:08:54.908 }, 00:08:54.908 { 00:08:54.908 "name": "BaseBdev4", 00:08:54.908 "uuid": "604be6d1-6b0e-4805-9e72-d2730983fdaf", 00:08:54.908 "is_configured": true, 00:08:54.908 "data_offset": 0, 00:08:54.908 "data_size": 65536 00:08:54.908 } 00:08:54.908 ] 00:08:54.908 }' 00:08:54.908 23:42:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:54.908 23:42:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:55.168 23:42:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:08:55.168 23:42:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:08:55.168 23:42:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:55.168 23:42:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:55.168 23:42:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:08:55.168 23:42:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:55.168 23:42:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:55.168 23:42:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:08:55.168 23:42:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:55.168 23:42:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:55.168 [2024-11-26 23:42:43.266104] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:55.168 23:42:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:55.428 23:42:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:55.428 "name": "Existed_Raid", 00:08:55.428 "aliases": [ 00:08:55.429 "c70bc50f-3e0c-4894-8e66-5d15b2510ad3" 00:08:55.429 ], 00:08:55.429 "product_name": "Raid Volume", 00:08:55.429 "block_size": 512, 00:08:55.429 "num_blocks": 262144, 00:08:55.429 "uuid": "c70bc50f-3e0c-4894-8e66-5d15b2510ad3", 00:08:55.429 "assigned_rate_limits": { 00:08:55.429 "rw_ios_per_sec": 0, 00:08:55.429 "rw_mbytes_per_sec": 0, 00:08:55.429 "r_mbytes_per_sec": 0, 00:08:55.429 "w_mbytes_per_sec": 0 00:08:55.429 }, 00:08:55.429 "claimed": false, 00:08:55.429 "zoned": false, 00:08:55.429 "supported_io_types": { 00:08:55.429 "read": true, 00:08:55.429 "write": true, 00:08:55.429 "unmap": true, 00:08:55.429 "flush": true, 00:08:55.429 "reset": true, 00:08:55.429 "nvme_admin": false, 00:08:55.429 "nvme_io": false, 00:08:55.429 "nvme_io_md": false, 00:08:55.429 "write_zeroes": true, 00:08:55.429 "zcopy": false, 00:08:55.429 "get_zone_info": false, 00:08:55.429 "zone_management": false, 00:08:55.429 "zone_append": false, 00:08:55.429 "compare": false, 00:08:55.429 "compare_and_write": false, 00:08:55.429 "abort": false, 00:08:55.429 "seek_hole": false, 00:08:55.429 "seek_data": false, 00:08:55.429 "copy": false, 00:08:55.429 "nvme_iov_md": false 00:08:55.429 }, 00:08:55.429 "memory_domains": [ 00:08:55.429 { 00:08:55.429 "dma_device_id": "system", 00:08:55.429 "dma_device_type": 1 00:08:55.429 }, 00:08:55.429 { 00:08:55.429 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:55.429 "dma_device_type": 2 00:08:55.429 }, 00:08:55.429 { 00:08:55.429 "dma_device_id": "system", 00:08:55.429 "dma_device_type": 1 00:08:55.429 }, 00:08:55.429 { 00:08:55.429 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:55.429 "dma_device_type": 2 00:08:55.429 }, 00:08:55.429 { 00:08:55.429 "dma_device_id": "system", 00:08:55.429 "dma_device_type": 1 00:08:55.429 }, 00:08:55.429 { 00:08:55.429 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:55.429 "dma_device_type": 2 00:08:55.429 }, 00:08:55.429 { 00:08:55.429 "dma_device_id": "system", 00:08:55.429 "dma_device_type": 1 00:08:55.429 }, 00:08:55.429 { 00:08:55.429 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:55.429 "dma_device_type": 2 00:08:55.429 } 00:08:55.429 ], 00:08:55.429 "driver_specific": { 00:08:55.429 "raid": { 00:08:55.429 "uuid": "c70bc50f-3e0c-4894-8e66-5d15b2510ad3", 00:08:55.429 "strip_size_kb": 64, 00:08:55.429 "state": "online", 00:08:55.429 "raid_level": "raid0", 00:08:55.429 "superblock": false, 00:08:55.429 "num_base_bdevs": 4, 00:08:55.429 "num_base_bdevs_discovered": 4, 00:08:55.429 "num_base_bdevs_operational": 4, 00:08:55.429 "base_bdevs_list": [ 00:08:55.429 { 00:08:55.429 "name": "NewBaseBdev", 00:08:55.429 "uuid": "b9eea5fb-f21c-47bc-a8b2-79c965ee047c", 00:08:55.429 "is_configured": true, 00:08:55.429 "data_offset": 0, 00:08:55.429 "data_size": 65536 00:08:55.429 }, 00:08:55.429 { 00:08:55.429 "name": "BaseBdev2", 00:08:55.429 "uuid": "f214a47e-6afc-47f2-b397-a265571a3fd2", 00:08:55.429 "is_configured": true, 00:08:55.429 "data_offset": 0, 00:08:55.429 "data_size": 65536 00:08:55.429 }, 00:08:55.429 { 00:08:55.429 "name": "BaseBdev3", 00:08:55.429 "uuid": "77ce42e3-9160-4528-b363-e54f9b8f8da6", 00:08:55.429 "is_configured": true, 00:08:55.429 "data_offset": 0, 00:08:55.429 "data_size": 65536 00:08:55.429 }, 00:08:55.429 { 00:08:55.429 "name": "BaseBdev4", 00:08:55.429 "uuid": "604be6d1-6b0e-4805-9e72-d2730983fdaf", 00:08:55.429 "is_configured": true, 00:08:55.429 "data_offset": 0, 00:08:55.429 "data_size": 65536 00:08:55.429 } 00:08:55.429 ] 00:08:55.429 } 00:08:55.429 } 00:08:55.429 }' 00:08:55.429 23:42:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:55.429 23:42:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:08:55.429 BaseBdev2 00:08:55.429 BaseBdev3 00:08:55.429 BaseBdev4' 00:08:55.429 23:42:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:55.429 23:42:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:55.429 23:42:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:55.429 23:42:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:08:55.429 23:42:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:55.429 23:42:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:55.429 23:42:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:55.429 23:42:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:55.429 23:42:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:55.429 23:42:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:55.429 23:42:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:55.429 23:42:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:08:55.429 23:42:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:55.429 23:42:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:55.429 23:42:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:55.429 23:42:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:55.429 23:42:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:55.429 23:42:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:55.429 23:42:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:55.429 23:42:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:08:55.429 23:42:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:55.429 23:42:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:55.429 23:42:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:55.429 23:42:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:55.688 23:42:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:55.688 23:42:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:55.688 23:42:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:55.688 23:42:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:55.688 23:42:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:08:55.688 23:42:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:55.688 23:42:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:55.688 23:42:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:55.688 23:42:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:55.688 23:42:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:55.688 23:42:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:55.688 23:42:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:55.688 23:42:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:55.689 [2024-11-26 23:42:43.617227] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:55.689 [2024-11-26 23:42:43.617323] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:55.689 [2024-11-26 23:42:43.617426] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:55.689 [2024-11-26 23:42:43.617493] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:55.689 [2024-11-26 23:42:43.617512] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name Existed_Raid, state offline 00:08:55.689 23:42:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:55.689 23:42:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 80066 00:08:55.689 23:42:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # '[' -z 80066 ']' 00:08:55.689 23:42:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@958 -- # kill -0 80066 00:08:55.689 23:42:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # uname 00:08:55.689 23:42:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:08:55.689 23:42:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 80066 00:08:55.689 23:42:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:08:55.689 23:42:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:08:55.689 23:42:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 80066' 00:08:55.689 killing process with pid 80066 00:08:55.689 23:42:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@973 -- # kill 80066 00:08:55.689 [2024-11-26 23:42:43.666555] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:55.689 23:42:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@978 -- # wait 80066 00:08:55.689 [2024-11-26 23:42:43.708345] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:55.948 23:42:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:08:55.948 00:08:55.948 real 0m9.725s 00:08:55.948 user 0m16.802s 00:08:55.948 sys 0m1.939s 00:08:55.948 23:42:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:08:55.948 23:42:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:55.948 ************************************ 00:08:55.948 END TEST raid_state_function_test 00:08:55.948 ************************************ 00:08:55.948 23:42:43 bdev_raid -- bdev/bdev_raid.sh@969 -- # run_test raid_state_function_test_sb raid_state_function_test raid0 4 true 00:08:55.948 23:42:43 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:08:55.948 23:42:43 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:08:55.948 23:42:43 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:55.948 ************************************ 00:08:55.948 START TEST raid_state_function_test_sb 00:08:55.948 ************************************ 00:08:55.948 23:42:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1129 -- # raid_state_function_test raid0 4 true 00:08:55.948 23:42:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=raid0 00:08:55.948 23:42:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=4 00:08:55.948 23:42:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:08:55.948 23:42:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:08:55.948 23:42:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:08:55.948 23:42:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:55.948 23:42:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:08:55.948 23:42:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:55.948 23:42:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:55.948 23:42:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:08:55.948 23:42:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:55.948 23:42:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:55.948 23:42:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:08:55.948 23:42:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:55.948 23:42:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:55.948 23:42:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev4 00:08:55.948 23:42:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:55.948 23:42:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:55.948 23:42:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:08:55.948 23:42:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:08:55.948 23:42:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:08:55.948 23:42:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:08:55.948 23:42:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:08:55.948 23:42:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:08:55.948 23:42:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' raid0 '!=' raid1 ']' 00:08:55.948 23:42:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:08:55.948 23:42:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:08:55.948 23:42:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:08:55.948 23:42:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:08:55.948 Process raid pid: 80723 00:08:55.948 23:42:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=80723 00:08:55.948 23:42:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:08:55.948 23:42:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 80723' 00:08:55.948 23:42:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 80723 00:08:55.948 23:42:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@835 -- # '[' -z 80723 ']' 00:08:55.948 23:42:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:55.948 23:42:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@840 -- # local max_retries=100 00:08:55.948 23:42:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:55.948 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:55.948 23:42:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@844 -- # xtrace_disable 00:08:55.948 23:42:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:56.214 [2024-11-26 23:42:44.088438] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:08:56.214 [2024-11-26 23:42:44.088665] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:08:56.214 [2024-11-26 23:42:44.243686] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:56.214 [2024-11-26 23:42:44.272786] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:56.214 [2024-11-26 23:42:44.315993] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:56.214 [2024-11-26 23:42:44.316106] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:56.827 23:42:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:08:56.827 23:42:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@868 -- # return 0 00:08:56.827 23:42:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:08:56.827 23:42:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:56.827 23:42:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:56.827 [2024-11-26 23:42:44.935013] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:56.827 [2024-11-26 23:42:44.935131] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:56.827 [2024-11-26 23:42:44.935172] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:56.827 [2024-11-26 23:42:44.935220] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:56.827 [2024-11-26 23:42:44.935251] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:08:56.827 [2024-11-26 23:42:44.935282] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:08:56.827 [2024-11-26 23:42:44.935327] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:08:56.827 [2024-11-26 23:42:44.935375] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:08:56.827 23:42:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:56.827 23:42:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:08:56.827 23:42:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:56.827 23:42:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:56.827 23:42:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:56.827 23:42:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:56.827 23:42:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:08:56.827 23:42:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:56.827 23:42:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:56.827 23:42:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:56.827 23:42:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:56.827 23:42:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:56.827 23:42:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:56.827 23:42:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:56.827 23:42:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:57.086 23:42:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:57.086 23:42:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:57.086 "name": "Existed_Raid", 00:08:57.086 "uuid": "60b3ad98-14d2-464d-9faf-b234f20eaa79", 00:08:57.086 "strip_size_kb": 64, 00:08:57.086 "state": "configuring", 00:08:57.086 "raid_level": "raid0", 00:08:57.086 "superblock": true, 00:08:57.086 "num_base_bdevs": 4, 00:08:57.086 "num_base_bdevs_discovered": 0, 00:08:57.086 "num_base_bdevs_operational": 4, 00:08:57.086 "base_bdevs_list": [ 00:08:57.086 { 00:08:57.086 "name": "BaseBdev1", 00:08:57.086 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:57.086 "is_configured": false, 00:08:57.086 "data_offset": 0, 00:08:57.086 "data_size": 0 00:08:57.086 }, 00:08:57.086 { 00:08:57.086 "name": "BaseBdev2", 00:08:57.086 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:57.086 "is_configured": false, 00:08:57.086 "data_offset": 0, 00:08:57.086 "data_size": 0 00:08:57.086 }, 00:08:57.086 { 00:08:57.086 "name": "BaseBdev3", 00:08:57.086 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:57.086 "is_configured": false, 00:08:57.086 "data_offset": 0, 00:08:57.086 "data_size": 0 00:08:57.086 }, 00:08:57.086 { 00:08:57.086 "name": "BaseBdev4", 00:08:57.086 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:57.086 "is_configured": false, 00:08:57.086 "data_offset": 0, 00:08:57.086 "data_size": 0 00:08:57.086 } 00:08:57.086 ] 00:08:57.086 }' 00:08:57.086 23:42:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:57.086 23:42:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:57.346 23:42:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:57.346 23:42:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:57.346 23:42:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:57.346 [2024-11-26 23:42:45.394126] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:57.346 [2024-11-26 23:42:45.394230] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:08:57.346 23:42:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:57.346 23:42:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:08:57.346 23:42:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:57.346 23:42:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:57.346 [2024-11-26 23:42:45.406128] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:57.346 [2024-11-26 23:42:45.406176] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:57.346 [2024-11-26 23:42:45.406186] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:57.346 [2024-11-26 23:42:45.406195] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:57.346 [2024-11-26 23:42:45.406200] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:08:57.346 [2024-11-26 23:42:45.406209] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:08:57.346 [2024-11-26 23:42:45.406215] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:08:57.346 [2024-11-26 23:42:45.406223] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:08:57.346 23:42:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:57.346 23:42:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:08:57.346 23:42:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:57.346 23:42:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:57.346 [2024-11-26 23:42:45.427118] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:57.346 BaseBdev1 00:08:57.346 23:42:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:57.346 23:42:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:08:57.346 23:42:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:08:57.346 23:42:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:57.346 23:42:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:08:57.346 23:42:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:57.346 23:42:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:57.346 23:42:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:57.346 23:42:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:57.346 23:42:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:57.346 23:42:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:57.346 23:42:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:08:57.346 23:42:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:57.346 23:42:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:57.346 [ 00:08:57.346 { 00:08:57.346 "name": "BaseBdev1", 00:08:57.346 "aliases": [ 00:08:57.346 "417cfaa8-0648-4f72-886e-ef1167577da1" 00:08:57.346 ], 00:08:57.346 "product_name": "Malloc disk", 00:08:57.346 "block_size": 512, 00:08:57.346 "num_blocks": 65536, 00:08:57.346 "uuid": "417cfaa8-0648-4f72-886e-ef1167577da1", 00:08:57.346 "assigned_rate_limits": { 00:08:57.346 "rw_ios_per_sec": 0, 00:08:57.346 "rw_mbytes_per_sec": 0, 00:08:57.346 "r_mbytes_per_sec": 0, 00:08:57.346 "w_mbytes_per_sec": 0 00:08:57.346 }, 00:08:57.346 "claimed": true, 00:08:57.346 "claim_type": "exclusive_write", 00:08:57.346 "zoned": false, 00:08:57.346 "supported_io_types": { 00:08:57.346 "read": true, 00:08:57.346 "write": true, 00:08:57.346 "unmap": true, 00:08:57.346 "flush": true, 00:08:57.346 "reset": true, 00:08:57.346 "nvme_admin": false, 00:08:57.346 "nvme_io": false, 00:08:57.346 "nvme_io_md": false, 00:08:57.346 "write_zeroes": true, 00:08:57.346 "zcopy": true, 00:08:57.346 "get_zone_info": false, 00:08:57.346 "zone_management": false, 00:08:57.346 "zone_append": false, 00:08:57.346 "compare": false, 00:08:57.346 "compare_and_write": false, 00:08:57.346 "abort": true, 00:08:57.346 "seek_hole": false, 00:08:57.346 "seek_data": false, 00:08:57.346 "copy": true, 00:08:57.346 "nvme_iov_md": false 00:08:57.346 }, 00:08:57.346 "memory_domains": [ 00:08:57.346 { 00:08:57.346 "dma_device_id": "system", 00:08:57.346 "dma_device_type": 1 00:08:57.346 }, 00:08:57.346 { 00:08:57.346 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:57.346 "dma_device_type": 2 00:08:57.346 } 00:08:57.346 ], 00:08:57.346 "driver_specific": {} 00:08:57.346 } 00:08:57.346 ] 00:08:57.346 23:42:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:57.346 23:42:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:08:57.346 23:42:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:08:57.346 23:42:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:57.346 23:42:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:57.346 23:42:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:57.346 23:42:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:57.346 23:42:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:08:57.346 23:42:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:57.346 23:42:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:57.347 23:42:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:57.347 23:42:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:57.347 23:42:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:57.347 23:42:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:57.347 23:42:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:57.347 23:42:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:57.614 23:42:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:57.615 23:42:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:57.615 "name": "Existed_Raid", 00:08:57.615 "uuid": "12fbe0ae-ff25-4300-93f3-f294c9d4f4fd", 00:08:57.615 "strip_size_kb": 64, 00:08:57.615 "state": "configuring", 00:08:57.615 "raid_level": "raid0", 00:08:57.615 "superblock": true, 00:08:57.615 "num_base_bdevs": 4, 00:08:57.615 "num_base_bdevs_discovered": 1, 00:08:57.615 "num_base_bdevs_operational": 4, 00:08:57.615 "base_bdevs_list": [ 00:08:57.615 { 00:08:57.615 "name": "BaseBdev1", 00:08:57.615 "uuid": "417cfaa8-0648-4f72-886e-ef1167577da1", 00:08:57.615 "is_configured": true, 00:08:57.615 "data_offset": 2048, 00:08:57.615 "data_size": 63488 00:08:57.615 }, 00:08:57.615 { 00:08:57.615 "name": "BaseBdev2", 00:08:57.615 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:57.615 "is_configured": false, 00:08:57.615 "data_offset": 0, 00:08:57.615 "data_size": 0 00:08:57.615 }, 00:08:57.615 { 00:08:57.615 "name": "BaseBdev3", 00:08:57.615 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:57.615 "is_configured": false, 00:08:57.615 "data_offset": 0, 00:08:57.615 "data_size": 0 00:08:57.615 }, 00:08:57.615 { 00:08:57.615 "name": "BaseBdev4", 00:08:57.615 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:57.615 "is_configured": false, 00:08:57.615 "data_offset": 0, 00:08:57.615 "data_size": 0 00:08:57.615 } 00:08:57.615 ] 00:08:57.615 }' 00:08:57.615 23:42:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:57.615 23:42:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:57.876 23:42:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:57.876 23:42:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:57.876 23:42:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:57.876 [2024-11-26 23:42:45.910374] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:57.876 [2024-11-26 23:42:45.910477] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:08:57.876 23:42:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:57.876 23:42:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:08:57.876 23:42:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:57.876 23:42:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:57.876 [2024-11-26 23:42:45.922400] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:57.876 [2024-11-26 23:42:45.924341] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:57.876 [2024-11-26 23:42:45.924436] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:57.876 [2024-11-26 23:42:45.924471] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:08:57.876 [2024-11-26 23:42:45.924507] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:08:57.876 [2024-11-26 23:42:45.924533] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:08:57.876 [2024-11-26 23:42:45.924561] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:08:57.876 23:42:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:57.876 23:42:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:08:57.876 23:42:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:57.876 23:42:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:08:57.876 23:42:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:57.876 23:42:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:57.876 23:42:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:57.876 23:42:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:57.876 23:42:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:08:57.876 23:42:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:57.876 23:42:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:57.876 23:42:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:57.876 23:42:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:57.876 23:42:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:57.876 23:42:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:57.876 23:42:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:57.876 23:42:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:57.876 23:42:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:57.876 23:42:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:57.876 "name": "Existed_Raid", 00:08:57.876 "uuid": "b02b57d6-841a-424c-b8f2-13c3501360ed", 00:08:57.876 "strip_size_kb": 64, 00:08:57.876 "state": "configuring", 00:08:57.876 "raid_level": "raid0", 00:08:57.876 "superblock": true, 00:08:57.876 "num_base_bdevs": 4, 00:08:57.876 "num_base_bdevs_discovered": 1, 00:08:57.876 "num_base_bdevs_operational": 4, 00:08:57.876 "base_bdevs_list": [ 00:08:57.876 { 00:08:57.876 "name": "BaseBdev1", 00:08:57.876 "uuid": "417cfaa8-0648-4f72-886e-ef1167577da1", 00:08:57.876 "is_configured": true, 00:08:57.876 "data_offset": 2048, 00:08:57.876 "data_size": 63488 00:08:57.876 }, 00:08:57.876 { 00:08:57.876 "name": "BaseBdev2", 00:08:57.876 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:57.876 "is_configured": false, 00:08:57.876 "data_offset": 0, 00:08:57.876 "data_size": 0 00:08:57.876 }, 00:08:57.876 { 00:08:57.876 "name": "BaseBdev3", 00:08:57.876 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:57.876 "is_configured": false, 00:08:57.876 "data_offset": 0, 00:08:57.876 "data_size": 0 00:08:57.876 }, 00:08:57.876 { 00:08:57.876 "name": "BaseBdev4", 00:08:57.876 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:57.876 "is_configured": false, 00:08:57.876 "data_offset": 0, 00:08:57.876 "data_size": 0 00:08:57.876 } 00:08:57.876 ] 00:08:57.876 }' 00:08:57.876 23:42:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:57.876 23:42:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:58.446 23:42:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:08:58.446 23:42:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:58.446 23:42:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:58.446 [2024-11-26 23:42:46.408485] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:58.446 BaseBdev2 00:08:58.446 23:42:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:58.446 23:42:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:08:58.446 23:42:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:08:58.446 23:42:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:58.446 23:42:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:08:58.446 23:42:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:58.446 23:42:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:58.446 23:42:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:58.446 23:42:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:58.446 23:42:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:58.446 23:42:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:58.446 23:42:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:08:58.446 23:42:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:58.446 23:42:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:58.446 [ 00:08:58.446 { 00:08:58.446 "name": "BaseBdev2", 00:08:58.446 "aliases": [ 00:08:58.446 "af8bfd70-3e81-45fb-bdfb-f0beceea02d7" 00:08:58.446 ], 00:08:58.446 "product_name": "Malloc disk", 00:08:58.446 "block_size": 512, 00:08:58.446 "num_blocks": 65536, 00:08:58.446 "uuid": "af8bfd70-3e81-45fb-bdfb-f0beceea02d7", 00:08:58.446 "assigned_rate_limits": { 00:08:58.446 "rw_ios_per_sec": 0, 00:08:58.446 "rw_mbytes_per_sec": 0, 00:08:58.446 "r_mbytes_per_sec": 0, 00:08:58.446 "w_mbytes_per_sec": 0 00:08:58.446 }, 00:08:58.446 "claimed": true, 00:08:58.446 "claim_type": "exclusive_write", 00:08:58.446 "zoned": false, 00:08:58.446 "supported_io_types": { 00:08:58.446 "read": true, 00:08:58.446 "write": true, 00:08:58.446 "unmap": true, 00:08:58.446 "flush": true, 00:08:58.446 "reset": true, 00:08:58.446 "nvme_admin": false, 00:08:58.446 "nvme_io": false, 00:08:58.446 "nvme_io_md": false, 00:08:58.446 "write_zeroes": true, 00:08:58.446 "zcopy": true, 00:08:58.446 "get_zone_info": false, 00:08:58.446 "zone_management": false, 00:08:58.446 "zone_append": false, 00:08:58.446 "compare": false, 00:08:58.446 "compare_and_write": false, 00:08:58.446 "abort": true, 00:08:58.446 "seek_hole": false, 00:08:58.446 "seek_data": false, 00:08:58.446 "copy": true, 00:08:58.446 "nvme_iov_md": false 00:08:58.446 }, 00:08:58.446 "memory_domains": [ 00:08:58.446 { 00:08:58.446 "dma_device_id": "system", 00:08:58.446 "dma_device_type": 1 00:08:58.446 }, 00:08:58.446 { 00:08:58.446 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:58.446 "dma_device_type": 2 00:08:58.446 } 00:08:58.446 ], 00:08:58.446 "driver_specific": {} 00:08:58.446 } 00:08:58.446 ] 00:08:58.446 23:42:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:58.446 23:42:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:08:58.446 23:42:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:08:58.446 23:42:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:58.446 23:42:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:08:58.446 23:42:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:58.446 23:42:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:58.446 23:42:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:58.446 23:42:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:58.446 23:42:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:08:58.446 23:42:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:58.446 23:42:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:58.446 23:42:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:58.446 23:42:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:58.446 23:42:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:58.446 23:42:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:58.446 23:42:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:58.446 23:42:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:58.446 23:42:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:58.446 23:42:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:58.446 "name": "Existed_Raid", 00:08:58.446 "uuid": "b02b57d6-841a-424c-b8f2-13c3501360ed", 00:08:58.446 "strip_size_kb": 64, 00:08:58.446 "state": "configuring", 00:08:58.446 "raid_level": "raid0", 00:08:58.446 "superblock": true, 00:08:58.446 "num_base_bdevs": 4, 00:08:58.446 "num_base_bdevs_discovered": 2, 00:08:58.446 "num_base_bdevs_operational": 4, 00:08:58.446 "base_bdevs_list": [ 00:08:58.446 { 00:08:58.446 "name": "BaseBdev1", 00:08:58.446 "uuid": "417cfaa8-0648-4f72-886e-ef1167577da1", 00:08:58.446 "is_configured": true, 00:08:58.446 "data_offset": 2048, 00:08:58.446 "data_size": 63488 00:08:58.446 }, 00:08:58.446 { 00:08:58.446 "name": "BaseBdev2", 00:08:58.446 "uuid": "af8bfd70-3e81-45fb-bdfb-f0beceea02d7", 00:08:58.446 "is_configured": true, 00:08:58.446 "data_offset": 2048, 00:08:58.446 "data_size": 63488 00:08:58.447 }, 00:08:58.447 { 00:08:58.447 "name": "BaseBdev3", 00:08:58.447 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:58.447 "is_configured": false, 00:08:58.447 "data_offset": 0, 00:08:58.447 "data_size": 0 00:08:58.447 }, 00:08:58.447 { 00:08:58.447 "name": "BaseBdev4", 00:08:58.447 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:58.447 "is_configured": false, 00:08:58.447 "data_offset": 0, 00:08:58.447 "data_size": 0 00:08:58.447 } 00:08:58.447 ] 00:08:58.447 }' 00:08:58.447 23:42:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:58.447 23:42:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:59.016 23:42:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:08:59.016 23:42:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:59.016 23:42:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:59.016 [2024-11-26 23:42:46.938472] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:59.016 BaseBdev3 00:08:59.016 23:42:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:59.016 23:42:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:08:59.016 23:42:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:08:59.016 23:42:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:59.016 23:42:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:08:59.016 23:42:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:59.016 23:42:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:59.016 23:42:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:59.016 23:42:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:59.016 23:42:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:59.016 23:42:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:59.016 23:42:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:08:59.016 23:42:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:59.016 23:42:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:59.016 [ 00:08:59.016 { 00:08:59.016 "name": "BaseBdev3", 00:08:59.016 "aliases": [ 00:08:59.016 "03d8cb13-e06e-413a-ac4a-57dc407bb4c3" 00:08:59.016 ], 00:08:59.016 "product_name": "Malloc disk", 00:08:59.016 "block_size": 512, 00:08:59.016 "num_blocks": 65536, 00:08:59.016 "uuid": "03d8cb13-e06e-413a-ac4a-57dc407bb4c3", 00:08:59.016 "assigned_rate_limits": { 00:08:59.016 "rw_ios_per_sec": 0, 00:08:59.016 "rw_mbytes_per_sec": 0, 00:08:59.016 "r_mbytes_per_sec": 0, 00:08:59.016 "w_mbytes_per_sec": 0 00:08:59.016 }, 00:08:59.016 "claimed": true, 00:08:59.016 "claim_type": "exclusive_write", 00:08:59.016 "zoned": false, 00:08:59.016 "supported_io_types": { 00:08:59.016 "read": true, 00:08:59.016 "write": true, 00:08:59.016 "unmap": true, 00:08:59.016 "flush": true, 00:08:59.016 "reset": true, 00:08:59.016 "nvme_admin": false, 00:08:59.016 "nvme_io": false, 00:08:59.016 "nvme_io_md": false, 00:08:59.016 "write_zeroes": true, 00:08:59.016 "zcopy": true, 00:08:59.016 "get_zone_info": false, 00:08:59.016 "zone_management": false, 00:08:59.016 "zone_append": false, 00:08:59.016 "compare": false, 00:08:59.016 "compare_and_write": false, 00:08:59.016 "abort": true, 00:08:59.016 "seek_hole": false, 00:08:59.016 "seek_data": false, 00:08:59.016 "copy": true, 00:08:59.016 "nvme_iov_md": false 00:08:59.016 }, 00:08:59.016 "memory_domains": [ 00:08:59.016 { 00:08:59.016 "dma_device_id": "system", 00:08:59.016 "dma_device_type": 1 00:08:59.016 }, 00:08:59.016 { 00:08:59.016 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:59.016 "dma_device_type": 2 00:08:59.016 } 00:08:59.016 ], 00:08:59.016 "driver_specific": {} 00:08:59.016 } 00:08:59.016 ] 00:08:59.016 23:42:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:59.016 23:42:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:08:59.016 23:42:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:08:59.016 23:42:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:59.016 23:42:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:08:59.016 23:42:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:59.016 23:42:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:59.016 23:42:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:59.016 23:42:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:59.016 23:42:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:08:59.016 23:42:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:59.016 23:42:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:59.016 23:42:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:59.016 23:42:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:59.016 23:42:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:59.016 23:42:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:59.016 23:42:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:59.016 23:42:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:59.016 23:42:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:59.016 23:42:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:59.016 "name": "Existed_Raid", 00:08:59.016 "uuid": "b02b57d6-841a-424c-b8f2-13c3501360ed", 00:08:59.016 "strip_size_kb": 64, 00:08:59.016 "state": "configuring", 00:08:59.016 "raid_level": "raid0", 00:08:59.016 "superblock": true, 00:08:59.016 "num_base_bdevs": 4, 00:08:59.016 "num_base_bdevs_discovered": 3, 00:08:59.016 "num_base_bdevs_operational": 4, 00:08:59.016 "base_bdevs_list": [ 00:08:59.016 { 00:08:59.016 "name": "BaseBdev1", 00:08:59.016 "uuid": "417cfaa8-0648-4f72-886e-ef1167577da1", 00:08:59.016 "is_configured": true, 00:08:59.017 "data_offset": 2048, 00:08:59.017 "data_size": 63488 00:08:59.017 }, 00:08:59.017 { 00:08:59.017 "name": "BaseBdev2", 00:08:59.017 "uuid": "af8bfd70-3e81-45fb-bdfb-f0beceea02d7", 00:08:59.017 "is_configured": true, 00:08:59.017 "data_offset": 2048, 00:08:59.017 "data_size": 63488 00:08:59.017 }, 00:08:59.017 { 00:08:59.017 "name": "BaseBdev3", 00:08:59.017 "uuid": "03d8cb13-e06e-413a-ac4a-57dc407bb4c3", 00:08:59.017 "is_configured": true, 00:08:59.017 "data_offset": 2048, 00:08:59.017 "data_size": 63488 00:08:59.017 }, 00:08:59.017 { 00:08:59.017 "name": "BaseBdev4", 00:08:59.017 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:59.017 "is_configured": false, 00:08:59.017 "data_offset": 0, 00:08:59.017 "data_size": 0 00:08:59.017 } 00:08:59.017 ] 00:08:59.017 }' 00:08:59.017 23:42:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:59.017 23:42:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:59.585 23:42:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:08:59.585 23:42:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:59.585 23:42:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:59.585 [2024-11-26 23:42:47.444815] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:08:59.585 [2024-11-26 23:42:47.445034] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:08:59.585 [2024-11-26 23:42:47.445050] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:08:59.585 BaseBdev4 00:08:59.585 [2024-11-26 23:42:47.445317] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:08:59.585 [2024-11-26 23:42:47.445477] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:08:59.585 [2024-11-26 23:42:47.445490] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:08:59.585 [2024-11-26 23:42:47.445606] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:59.585 23:42:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:59.585 23:42:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev4 00:08:59.585 23:42:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:08:59.585 23:42:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:59.585 23:42:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:08:59.585 23:42:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:59.585 23:42:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:59.585 23:42:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:59.585 23:42:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:59.585 23:42:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:59.585 23:42:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:59.585 23:42:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:08:59.585 23:42:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:59.585 23:42:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:59.585 [ 00:08:59.585 { 00:08:59.585 "name": "BaseBdev4", 00:08:59.585 "aliases": [ 00:08:59.585 "09ac4a94-acc2-4c39-b144-348427daf010" 00:08:59.585 ], 00:08:59.585 "product_name": "Malloc disk", 00:08:59.585 "block_size": 512, 00:08:59.585 "num_blocks": 65536, 00:08:59.585 "uuid": "09ac4a94-acc2-4c39-b144-348427daf010", 00:08:59.585 "assigned_rate_limits": { 00:08:59.585 "rw_ios_per_sec": 0, 00:08:59.585 "rw_mbytes_per_sec": 0, 00:08:59.585 "r_mbytes_per_sec": 0, 00:08:59.585 "w_mbytes_per_sec": 0 00:08:59.585 }, 00:08:59.585 "claimed": true, 00:08:59.585 "claim_type": "exclusive_write", 00:08:59.585 "zoned": false, 00:08:59.585 "supported_io_types": { 00:08:59.585 "read": true, 00:08:59.585 "write": true, 00:08:59.585 "unmap": true, 00:08:59.585 "flush": true, 00:08:59.585 "reset": true, 00:08:59.585 "nvme_admin": false, 00:08:59.585 "nvme_io": false, 00:08:59.585 "nvme_io_md": false, 00:08:59.585 "write_zeroes": true, 00:08:59.585 "zcopy": true, 00:08:59.585 "get_zone_info": false, 00:08:59.585 "zone_management": false, 00:08:59.585 "zone_append": false, 00:08:59.585 "compare": false, 00:08:59.585 "compare_and_write": false, 00:08:59.585 "abort": true, 00:08:59.585 "seek_hole": false, 00:08:59.586 "seek_data": false, 00:08:59.586 "copy": true, 00:08:59.586 "nvme_iov_md": false 00:08:59.586 }, 00:08:59.586 "memory_domains": [ 00:08:59.586 { 00:08:59.586 "dma_device_id": "system", 00:08:59.586 "dma_device_type": 1 00:08:59.586 }, 00:08:59.586 { 00:08:59.586 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:59.586 "dma_device_type": 2 00:08:59.586 } 00:08:59.586 ], 00:08:59.586 "driver_specific": {} 00:08:59.586 } 00:08:59.586 ] 00:08:59.586 23:42:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:59.586 23:42:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:08:59.586 23:42:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:08:59.586 23:42:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:59.586 23:42:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid0 64 4 00:08:59.586 23:42:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:59.586 23:42:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:59.586 23:42:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:59.586 23:42:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:59.586 23:42:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:08:59.586 23:42:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:59.586 23:42:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:59.586 23:42:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:59.586 23:42:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:59.586 23:42:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:59.586 23:42:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:59.586 23:42:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:59.586 23:42:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:59.586 23:42:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:59.586 23:42:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:59.586 "name": "Existed_Raid", 00:08:59.586 "uuid": "b02b57d6-841a-424c-b8f2-13c3501360ed", 00:08:59.586 "strip_size_kb": 64, 00:08:59.586 "state": "online", 00:08:59.586 "raid_level": "raid0", 00:08:59.586 "superblock": true, 00:08:59.586 "num_base_bdevs": 4, 00:08:59.586 "num_base_bdevs_discovered": 4, 00:08:59.586 "num_base_bdevs_operational": 4, 00:08:59.586 "base_bdevs_list": [ 00:08:59.586 { 00:08:59.586 "name": "BaseBdev1", 00:08:59.586 "uuid": "417cfaa8-0648-4f72-886e-ef1167577da1", 00:08:59.586 "is_configured": true, 00:08:59.586 "data_offset": 2048, 00:08:59.586 "data_size": 63488 00:08:59.586 }, 00:08:59.586 { 00:08:59.586 "name": "BaseBdev2", 00:08:59.586 "uuid": "af8bfd70-3e81-45fb-bdfb-f0beceea02d7", 00:08:59.586 "is_configured": true, 00:08:59.586 "data_offset": 2048, 00:08:59.586 "data_size": 63488 00:08:59.586 }, 00:08:59.586 { 00:08:59.586 "name": "BaseBdev3", 00:08:59.586 "uuid": "03d8cb13-e06e-413a-ac4a-57dc407bb4c3", 00:08:59.586 "is_configured": true, 00:08:59.586 "data_offset": 2048, 00:08:59.586 "data_size": 63488 00:08:59.586 }, 00:08:59.586 { 00:08:59.586 "name": "BaseBdev4", 00:08:59.586 "uuid": "09ac4a94-acc2-4c39-b144-348427daf010", 00:08:59.586 "is_configured": true, 00:08:59.586 "data_offset": 2048, 00:08:59.586 "data_size": 63488 00:08:59.586 } 00:08:59.586 ] 00:08:59.586 }' 00:08:59.586 23:42:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:59.586 23:42:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:59.846 23:42:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:08:59.846 23:42:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:08:59.846 23:42:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:59.846 23:42:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:59.846 23:42:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:08:59.846 23:42:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:59.846 23:42:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:59.846 23:42:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:08:59.846 23:42:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:59.846 23:42:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:59.846 [2024-11-26 23:42:47.944385] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:59.846 23:42:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:59.846 23:42:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:59.846 "name": "Existed_Raid", 00:08:59.846 "aliases": [ 00:08:59.846 "b02b57d6-841a-424c-b8f2-13c3501360ed" 00:08:59.846 ], 00:08:59.846 "product_name": "Raid Volume", 00:08:59.846 "block_size": 512, 00:08:59.846 "num_blocks": 253952, 00:08:59.846 "uuid": "b02b57d6-841a-424c-b8f2-13c3501360ed", 00:08:59.846 "assigned_rate_limits": { 00:08:59.846 "rw_ios_per_sec": 0, 00:08:59.846 "rw_mbytes_per_sec": 0, 00:08:59.846 "r_mbytes_per_sec": 0, 00:08:59.846 "w_mbytes_per_sec": 0 00:08:59.846 }, 00:08:59.846 "claimed": false, 00:08:59.846 "zoned": false, 00:08:59.846 "supported_io_types": { 00:08:59.846 "read": true, 00:08:59.846 "write": true, 00:08:59.846 "unmap": true, 00:08:59.846 "flush": true, 00:08:59.846 "reset": true, 00:08:59.846 "nvme_admin": false, 00:08:59.846 "nvme_io": false, 00:08:59.846 "nvme_io_md": false, 00:08:59.846 "write_zeroes": true, 00:08:59.846 "zcopy": false, 00:08:59.846 "get_zone_info": false, 00:08:59.846 "zone_management": false, 00:08:59.846 "zone_append": false, 00:08:59.846 "compare": false, 00:08:59.846 "compare_and_write": false, 00:08:59.846 "abort": false, 00:08:59.846 "seek_hole": false, 00:08:59.846 "seek_data": false, 00:08:59.846 "copy": false, 00:08:59.846 "nvme_iov_md": false 00:08:59.846 }, 00:08:59.846 "memory_domains": [ 00:08:59.846 { 00:08:59.846 "dma_device_id": "system", 00:08:59.846 "dma_device_type": 1 00:08:59.846 }, 00:08:59.846 { 00:08:59.846 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:59.846 "dma_device_type": 2 00:08:59.846 }, 00:08:59.846 { 00:08:59.846 "dma_device_id": "system", 00:08:59.846 "dma_device_type": 1 00:08:59.846 }, 00:08:59.846 { 00:08:59.846 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:59.846 "dma_device_type": 2 00:08:59.846 }, 00:08:59.846 { 00:08:59.846 "dma_device_id": "system", 00:08:59.846 "dma_device_type": 1 00:08:59.846 }, 00:08:59.846 { 00:08:59.846 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:59.846 "dma_device_type": 2 00:08:59.846 }, 00:08:59.846 { 00:08:59.846 "dma_device_id": "system", 00:08:59.846 "dma_device_type": 1 00:08:59.846 }, 00:08:59.846 { 00:08:59.846 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:59.846 "dma_device_type": 2 00:08:59.846 } 00:08:59.846 ], 00:08:59.846 "driver_specific": { 00:08:59.846 "raid": { 00:08:59.846 "uuid": "b02b57d6-841a-424c-b8f2-13c3501360ed", 00:08:59.846 "strip_size_kb": 64, 00:08:59.846 "state": "online", 00:08:59.846 "raid_level": "raid0", 00:08:59.846 "superblock": true, 00:08:59.846 "num_base_bdevs": 4, 00:08:59.846 "num_base_bdevs_discovered": 4, 00:08:59.846 "num_base_bdevs_operational": 4, 00:08:59.846 "base_bdevs_list": [ 00:08:59.846 { 00:08:59.846 "name": "BaseBdev1", 00:08:59.846 "uuid": "417cfaa8-0648-4f72-886e-ef1167577da1", 00:08:59.846 "is_configured": true, 00:08:59.846 "data_offset": 2048, 00:08:59.846 "data_size": 63488 00:08:59.846 }, 00:08:59.846 { 00:08:59.846 "name": "BaseBdev2", 00:08:59.846 "uuid": "af8bfd70-3e81-45fb-bdfb-f0beceea02d7", 00:08:59.846 "is_configured": true, 00:08:59.846 "data_offset": 2048, 00:08:59.846 "data_size": 63488 00:08:59.846 }, 00:08:59.846 { 00:08:59.846 "name": "BaseBdev3", 00:08:59.846 "uuid": "03d8cb13-e06e-413a-ac4a-57dc407bb4c3", 00:08:59.846 "is_configured": true, 00:08:59.846 "data_offset": 2048, 00:08:59.846 "data_size": 63488 00:08:59.846 }, 00:08:59.846 { 00:08:59.846 "name": "BaseBdev4", 00:08:59.846 "uuid": "09ac4a94-acc2-4c39-b144-348427daf010", 00:08:59.846 "is_configured": true, 00:08:59.846 "data_offset": 2048, 00:08:59.846 "data_size": 63488 00:08:59.846 } 00:08:59.846 ] 00:08:59.846 } 00:08:59.846 } 00:08:59.846 }' 00:09:00.106 23:42:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:00.106 23:42:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:09:00.106 BaseBdev2 00:09:00.106 BaseBdev3 00:09:00.106 BaseBdev4' 00:09:00.106 23:42:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:00.106 23:42:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:00.106 23:42:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:00.106 23:42:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:09:00.106 23:42:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:00.106 23:42:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:00.106 23:42:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:00.106 23:42:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:00.106 23:42:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:00.106 23:42:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:00.106 23:42:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:00.106 23:42:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:09:00.106 23:42:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:00.106 23:42:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:00.106 23:42:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:00.106 23:42:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:00.106 23:42:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:00.106 23:42:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:00.106 23:42:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:00.106 23:42:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:09:00.106 23:42:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:00.106 23:42:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:00.106 23:42:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:00.106 23:42:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:00.106 23:42:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:00.106 23:42:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:00.106 23:42:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:00.106 23:42:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:09:00.106 23:42:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:00.106 23:42:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:00.106 23:42:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:00.106 23:42:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:00.367 23:42:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:00.367 23:42:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:00.367 23:42:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:09:00.367 23:42:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:00.367 23:42:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:00.367 [2024-11-26 23:42:48.259561] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:09:00.367 [2024-11-26 23:42:48.259650] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:00.367 [2024-11-26 23:42:48.259724] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:00.367 23:42:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:00.367 23:42:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:09:00.367 23:42:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy raid0 00:09:00.367 23:42:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:09:00.367 23:42:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # return 1 00:09:00.367 23:42:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:09:00.367 23:42:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline raid0 64 3 00:09:00.367 23:42:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:00.367 23:42:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:09:00.367 23:42:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:00.367 23:42:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:00.367 23:42:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:00.367 23:42:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:00.367 23:42:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:00.367 23:42:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:00.367 23:42:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:00.367 23:42:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:00.367 23:42:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:00.367 23:42:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:00.367 23:42:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:00.367 23:42:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:00.367 23:42:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:00.367 "name": "Existed_Raid", 00:09:00.367 "uuid": "b02b57d6-841a-424c-b8f2-13c3501360ed", 00:09:00.367 "strip_size_kb": 64, 00:09:00.367 "state": "offline", 00:09:00.367 "raid_level": "raid0", 00:09:00.367 "superblock": true, 00:09:00.367 "num_base_bdevs": 4, 00:09:00.367 "num_base_bdevs_discovered": 3, 00:09:00.367 "num_base_bdevs_operational": 3, 00:09:00.367 "base_bdevs_list": [ 00:09:00.367 { 00:09:00.367 "name": null, 00:09:00.367 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:00.367 "is_configured": false, 00:09:00.367 "data_offset": 0, 00:09:00.367 "data_size": 63488 00:09:00.367 }, 00:09:00.367 { 00:09:00.367 "name": "BaseBdev2", 00:09:00.367 "uuid": "af8bfd70-3e81-45fb-bdfb-f0beceea02d7", 00:09:00.367 "is_configured": true, 00:09:00.367 "data_offset": 2048, 00:09:00.367 "data_size": 63488 00:09:00.367 }, 00:09:00.367 { 00:09:00.367 "name": "BaseBdev3", 00:09:00.367 "uuid": "03d8cb13-e06e-413a-ac4a-57dc407bb4c3", 00:09:00.367 "is_configured": true, 00:09:00.367 "data_offset": 2048, 00:09:00.367 "data_size": 63488 00:09:00.367 }, 00:09:00.367 { 00:09:00.367 "name": "BaseBdev4", 00:09:00.367 "uuid": "09ac4a94-acc2-4c39-b144-348427daf010", 00:09:00.367 "is_configured": true, 00:09:00.367 "data_offset": 2048, 00:09:00.367 "data_size": 63488 00:09:00.367 } 00:09:00.367 ] 00:09:00.367 }' 00:09:00.367 23:42:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:00.367 23:42:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:00.627 23:42:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:09:00.627 23:42:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:00.627 23:42:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:09:00.627 23:42:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:00.627 23:42:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:00.627 23:42:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:00.627 23:42:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:00.627 23:42:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:09:00.627 23:42:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:09:00.627 23:42:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:09:00.627 23:42:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:00.627 23:42:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:00.627 [2024-11-26 23:42:48.742014] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:09:00.627 23:42:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:00.627 23:42:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:09:00.627 23:42:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:00.886 23:42:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:00.886 23:42:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:00.886 23:42:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:09:00.886 23:42:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:00.886 23:42:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:00.886 23:42:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:09:00.886 23:42:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:09:00.886 23:42:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:09:00.886 23:42:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:00.886 23:42:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:00.886 [2024-11-26 23:42:48.813229] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:09:00.886 23:42:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:00.886 23:42:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:09:00.886 23:42:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:00.886 23:42:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:00.886 23:42:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:09:00.886 23:42:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:00.886 23:42:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:00.886 23:42:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:00.886 23:42:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:09:00.886 23:42:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:09:00.886 23:42:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev4 00:09:00.886 23:42:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:00.886 23:42:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:00.886 [2024-11-26 23:42:48.884453] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev4 00:09:00.886 [2024-11-26 23:42:48.884496] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:09:00.887 23:42:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:00.887 23:42:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:09:00.887 23:42:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:00.887 23:42:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:00.887 23:42:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:09:00.887 23:42:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:00.887 23:42:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:00.887 23:42:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:00.887 23:42:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:09:00.887 23:42:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:09:00.887 23:42:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 4 -gt 2 ']' 00:09:00.887 23:42:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:09:00.887 23:42:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:00.887 23:42:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:09:00.887 23:42:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:00.887 23:42:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:00.887 BaseBdev2 00:09:00.887 23:42:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:00.887 23:42:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:09:00.887 23:42:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:09:00.887 23:42:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:00.887 23:42:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:09:00.887 23:42:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:00.887 23:42:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:00.887 23:42:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:00.887 23:42:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:00.887 23:42:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:00.887 23:42:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:00.887 23:42:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:09:00.887 23:42:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:00.887 23:42:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:00.887 [ 00:09:00.887 { 00:09:00.887 "name": "BaseBdev2", 00:09:00.887 "aliases": [ 00:09:00.887 "54fae0e9-e3aa-4910-90e9-0d240981390f" 00:09:00.887 ], 00:09:00.887 "product_name": "Malloc disk", 00:09:00.887 "block_size": 512, 00:09:00.887 "num_blocks": 65536, 00:09:00.887 "uuid": "54fae0e9-e3aa-4910-90e9-0d240981390f", 00:09:00.887 "assigned_rate_limits": { 00:09:00.887 "rw_ios_per_sec": 0, 00:09:00.887 "rw_mbytes_per_sec": 0, 00:09:00.887 "r_mbytes_per_sec": 0, 00:09:00.887 "w_mbytes_per_sec": 0 00:09:00.887 }, 00:09:00.887 "claimed": false, 00:09:00.887 "zoned": false, 00:09:00.887 "supported_io_types": { 00:09:00.887 "read": true, 00:09:00.887 "write": true, 00:09:00.887 "unmap": true, 00:09:00.887 "flush": true, 00:09:00.887 "reset": true, 00:09:00.887 "nvme_admin": false, 00:09:00.887 "nvme_io": false, 00:09:00.887 "nvme_io_md": false, 00:09:00.887 "write_zeroes": true, 00:09:00.887 "zcopy": true, 00:09:00.887 "get_zone_info": false, 00:09:00.887 "zone_management": false, 00:09:00.887 "zone_append": false, 00:09:00.887 "compare": false, 00:09:00.887 "compare_and_write": false, 00:09:00.887 "abort": true, 00:09:00.887 "seek_hole": false, 00:09:00.887 "seek_data": false, 00:09:00.887 "copy": true, 00:09:00.887 "nvme_iov_md": false 00:09:00.887 }, 00:09:00.887 "memory_domains": [ 00:09:00.887 { 00:09:00.887 "dma_device_id": "system", 00:09:00.887 "dma_device_type": 1 00:09:00.887 }, 00:09:00.887 { 00:09:00.887 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:00.887 "dma_device_type": 2 00:09:00.887 } 00:09:00.887 ], 00:09:00.887 "driver_specific": {} 00:09:00.887 } 00:09:00.887 ] 00:09:00.887 23:42:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:00.887 23:42:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:09:00.887 23:42:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:09:00.887 23:42:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:00.887 23:42:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:09:00.887 23:42:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:00.887 23:42:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:01.147 BaseBdev3 00:09:01.147 23:42:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:01.147 23:42:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:09:01.147 23:42:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:09:01.147 23:42:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:01.147 23:42:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:09:01.147 23:42:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:01.147 23:42:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:01.147 23:42:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:01.147 23:42:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:01.147 23:42:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:01.147 23:42:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:01.147 23:42:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:09:01.147 23:42:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:01.147 23:42:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:01.147 [ 00:09:01.147 { 00:09:01.147 "name": "BaseBdev3", 00:09:01.147 "aliases": [ 00:09:01.147 "84e6a8e7-7c16-479b-9900-221eed699b29" 00:09:01.147 ], 00:09:01.147 "product_name": "Malloc disk", 00:09:01.147 "block_size": 512, 00:09:01.147 "num_blocks": 65536, 00:09:01.147 "uuid": "84e6a8e7-7c16-479b-9900-221eed699b29", 00:09:01.147 "assigned_rate_limits": { 00:09:01.147 "rw_ios_per_sec": 0, 00:09:01.147 "rw_mbytes_per_sec": 0, 00:09:01.147 "r_mbytes_per_sec": 0, 00:09:01.147 "w_mbytes_per_sec": 0 00:09:01.147 }, 00:09:01.147 "claimed": false, 00:09:01.148 "zoned": false, 00:09:01.148 "supported_io_types": { 00:09:01.148 "read": true, 00:09:01.148 "write": true, 00:09:01.148 "unmap": true, 00:09:01.148 "flush": true, 00:09:01.148 "reset": true, 00:09:01.148 "nvme_admin": false, 00:09:01.148 "nvme_io": false, 00:09:01.148 "nvme_io_md": false, 00:09:01.148 "write_zeroes": true, 00:09:01.148 "zcopy": true, 00:09:01.148 "get_zone_info": false, 00:09:01.148 "zone_management": false, 00:09:01.148 "zone_append": false, 00:09:01.148 "compare": false, 00:09:01.148 "compare_and_write": false, 00:09:01.148 "abort": true, 00:09:01.148 "seek_hole": false, 00:09:01.148 "seek_data": false, 00:09:01.148 "copy": true, 00:09:01.148 "nvme_iov_md": false 00:09:01.148 }, 00:09:01.148 "memory_domains": [ 00:09:01.148 { 00:09:01.148 "dma_device_id": "system", 00:09:01.148 "dma_device_type": 1 00:09:01.148 }, 00:09:01.148 { 00:09:01.148 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:01.148 "dma_device_type": 2 00:09:01.148 } 00:09:01.148 ], 00:09:01.148 "driver_specific": {} 00:09:01.148 } 00:09:01.148 ] 00:09:01.148 23:42:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:01.148 23:42:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:09:01.148 23:42:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:09:01.148 23:42:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:01.148 23:42:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:09:01.148 23:42:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:01.148 23:42:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:01.148 BaseBdev4 00:09:01.148 23:42:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:01.148 23:42:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev4 00:09:01.148 23:42:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:09:01.148 23:42:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:01.148 23:42:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:09:01.148 23:42:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:01.148 23:42:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:01.148 23:42:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:01.148 23:42:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:01.148 23:42:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:01.148 23:42:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:01.148 23:42:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:09:01.148 23:42:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:01.148 23:42:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:01.148 [ 00:09:01.148 { 00:09:01.148 "name": "BaseBdev4", 00:09:01.148 "aliases": [ 00:09:01.148 "366edb40-a913-47b8-9642-56b06f51f309" 00:09:01.148 ], 00:09:01.148 "product_name": "Malloc disk", 00:09:01.148 "block_size": 512, 00:09:01.148 "num_blocks": 65536, 00:09:01.148 "uuid": "366edb40-a913-47b8-9642-56b06f51f309", 00:09:01.148 "assigned_rate_limits": { 00:09:01.148 "rw_ios_per_sec": 0, 00:09:01.148 "rw_mbytes_per_sec": 0, 00:09:01.148 "r_mbytes_per_sec": 0, 00:09:01.148 "w_mbytes_per_sec": 0 00:09:01.148 }, 00:09:01.148 "claimed": false, 00:09:01.148 "zoned": false, 00:09:01.148 "supported_io_types": { 00:09:01.148 "read": true, 00:09:01.148 "write": true, 00:09:01.148 "unmap": true, 00:09:01.148 "flush": true, 00:09:01.148 "reset": true, 00:09:01.148 "nvme_admin": false, 00:09:01.148 "nvme_io": false, 00:09:01.148 "nvme_io_md": false, 00:09:01.148 "write_zeroes": true, 00:09:01.148 "zcopy": true, 00:09:01.148 "get_zone_info": false, 00:09:01.148 "zone_management": false, 00:09:01.148 "zone_append": false, 00:09:01.148 "compare": false, 00:09:01.148 "compare_and_write": false, 00:09:01.148 "abort": true, 00:09:01.148 "seek_hole": false, 00:09:01.148 "seek_data": false, 00:09:01.148 "copy": true, 00:09:01.148 "nvme_iov_md": false 00:09:01.148 }, 00:09:01.148 "memory_domains": [ 00:09:01.148 { 00:09:01.148 "dma_device_id": "system", 00:09:01.148 "dma_device_type": 1 00:09:01.148 }, 00:09:01.148 { 00:09:01.148 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:01.148 "dma_device_type": 2 00:09:01.148 } 00:09:01.148 ], 00:09:01.148 "driver_specific": {} 00:09:01.148 } 00:09:01.148 ] 00:09:01.148 23:42:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:01.148 23:42:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:09:01.148 23:42:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:09:01.148 23:42:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:01.148 23:42:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:09:01.148 23:42:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:01.148 23:42:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:01.148 [2024-11-26 23:42:49.114001] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:01.148 [2024-11-26 23:42:49.114089] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:01.148 [2024-11-26 23:42:49.114162] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:01.148 [2024-11-26 23:42:49.118423] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:01.148 [2024-11-26 23:42:49.118518] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:09:01.148 23:42:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:01.148 23:42:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:01.148 23:42:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:01.148 23:42:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:01.148 23:42:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:01.148 23:42:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:01.148 23:42:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:01.148 23:42:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:01.148 23:42:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:01.148 23:42:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:01.148 23:42:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:01.148 23:42:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:01.148 23:42:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:01.148 23:42:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:01.148 23:42:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:01.148 23:42:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:01.148 23:42:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:01.148 "name": "Existed_Raid", 00:09:01.148 "uuid": "056c3fba-be90-45ba-8133-24e12ca385aa", 00:09:01.148 "strip_size_kb": 64, 00:09:01.148 "state": "configuring", 00:09:01.148 "raid_level": "raid0", 00:09:01.148 "superblock": true, 00:09:01.148 "num_base_bdevs": 4, 00:09:01.148 "num_base_bdevs_discovered": 3, 00:09:01.148 "num_base_bdevs_operational": 4, 00:09:01.148 "base_bdevs_list": [ 00:09:01.148 { 00:09:01.148 "name": "BaseBdev1", 00:09:01.148 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:01.148 "is_configured": false, 00:09:01.148 "data_offset": 0, 00:09:01.148 "data_size": 0 00:09:01.148 }, 00:09:01.148 { 00:09:01.148 "name": "BaseBdev2", 00:09:01.148 "uuid": "54fae0e9-e3aa-4910-90e9-0d240981390f", 00:09:01.148 "is_configured": true, 00:09:01.148 "data_offset": 2048, 00:09:01.148 "data_size": 63488 00:09:01.148 }, 00:09:01.148 { 00:09:01.148 "name": "BaseBdev3", 00:09:01.148 "uuid": "84e6a8e7-7c16-479b-9900-221eed699b29", 00:09:01.148 "is_configured": true, 00:09:01.148 "data_offset": 2048, 00:09:01.148 "data_size": 63488 00:09:01.148 }, 00:09:01.148 { 00:09:01.148 "name": "BaseBdev4", 00:09:01.148 "uuid": "366edb40-a913-47b8-9642-56b06f51f309", 00:09:01.148 "is_configured": true, 00:09:01.148 "data_offset": 2048, 00:09:01.148 "data_size": 63488 00:09:01.148 } 00:09:01.148 ] 00:09:01.148 }' 00:09:01.148 23:42:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:01.148 23:42:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:01.717 23:42:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:09:01.717 23:42:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:01.717 23:42:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:01.717 [2024-11-26 23:42:49.545995] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:09:01.717 23:42:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:01.717 23:42:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:01.717 23:42:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:01.717 23:42:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:01.717 23:42:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:01.717 23:42:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:01.717 23:42:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:01.717 23:42:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:01.717 23:42:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:01.717 23:42:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:01.717 23:42:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:01.717 23:42:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:01.717 23:42:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:01.717 23:42:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:01.717 23:42:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:01.717 23:42:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:01.717 23:42:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:01.717 "name": "Existed_Raid", 00:09:01.717 "uuid": "056c3fba-be90-45ba-8133-24e12ca385aa", 00:09:01.717 "strip_size_kb": 64, 00:09:01.717 "state": "configuring", 00:09:01.717 "raid_level": "raid0", 00:09:01.717 "superblock": true, 00:09:01.717 "num_base_bdevs": 4, 00:09:01.717 "num_base_bdevs_discovered": 2, 00:09:01.717 "num_base_bdevs_operational": 4, 00:09:01.717 "base_bdevs_list": [ 00:09:01.717 { 00:09:01.717 "name": "BaseBdev1", 00:09:01.717 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:01.717 "is_configured": false, 00:09:01.717 "data_offset": 0, 00:09:01.717 "data_size": 0 00:09:01.717 }, 00:09:01.717 { 00:09:01.717 "name": null, 00:09:01.717 "uuid": "54fae0e9-e3aa-4910-90e9-0d240981390f", 00:09:01.717 "is_configured": false, 00:09:01.717 "data_offset": 0, 00:09:01.717 "data_size": 63488 00:09:01.717 }, 00:09:01.717 { 00:09:01.717 "name": "BaseBdev3", 00:09:01.717 "uuid": "84e6a8e7-7c16-479b-9900-221eed699b29", 00:09:01.717 "is_configured": true, 00:09:01.718 "data_offset": 2048, 00:09:01.718 "data_size": 63488 00:09:01.718 }, 00:09:01.718 { 00:09:01.718 "name": "BaseBdev4", 00:09:01.718 "uuid": "366edb40-a913-47b8-9642-56b06f51f309", 00:09:01.718 "is_configured": true, 00:09:01.718 "data_offset": 2048, 00:09:01.718 "data_size": 63488 00:09:01.718 } 00:09:01.718 ] 00:09:01.718 }' 00:09:01.718 23:42:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:01.718 23:42:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:01.977 23:42:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:01.977 23:42:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:01.977 23:42:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:01.977 23:42:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:09:01.977 23:42:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:01.977 23:42:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:09:01.977 23:42:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:09:01.977 23:42:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:01.977 23:42:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:01.977 [2024-11-26 23:42:50.048075] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:01.977 BaseBdev1 00:09:01.977 23:42:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:01.977 23:42:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:09:01.977 23:42:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:09:01.977 23:42:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:01.977 23:42:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:09:01.977 23:42:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:01.977 23:42:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:01.977 23:42:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:01.978 23:42:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:01.978 23:42:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:01.978 23:42:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:01.978 23:42:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:09:01.978 23:42:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:01.978 23:42:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:01.978 [ 00:09:01.978 { 00:09:01.978 "name": "BaseBdev1", 00:09:01.978 "aliases": [ 00:09:01.978 "16227b2c-b735-4ab7-bf87-f6ffa73fca82" 00:09:01.978 ], 00:09:01.978 "product_name": "Malloc disk", 00:09:01.978 "block_size": 512, 00:09:01.978 "num_blocks": 65536, 00:09:01.978 "uuid": "16227b2c-b735-4ab7-bf87-f6ffa73fca82", 00:09:01.978 "assigned_rate_limits": { 00:09:01.978 "rw_ios_per_sec": 0, 00:09:01.978 "rw_mbytes_per_sec": 0, 00:09:01.978 "r_mbytes_per_sec": 0, 00:09:01.978 "w_mbytes_per_sec": 0 00:09:01.978 }, 00:09:01.978 "claimed": true, 00:09:01.978 "claim_type": "exclusive_write", 00:09:01.978 "zoned": false, 00:09:01.978 "supported_io_types": { 00:09:01.978 "read": true, 00:09:01.978 "write": true, 00:09:01.978 "unmap": true, 00:09:01.978 "flush": true, 00:09:01.978 "reset": true, 00:09:01.978 "nvme_admin": false, 00:09:01.978 "nvme_io": false, 00:09:01.978 "nvme_io_md": false, 00:09:01.978 "write_zeroes": true, 00:09:01.978 "zcopy": true, 00:09:01.978 "get_zone_info": false, 00:09:01.978 "zone_management": false, 00:09:01.978 "zone_append": false, 00:09:01.978 "compare": false, 00:09:01.978 "compare_and_write": false, 00:09:01.978 "abort": true, 00:09:01.978 "seek_hole": false, 00:09:01.978 "seek_data": false, 00:09:01.978 "copy": true, 00:09:01.978 "nvme_iov_md": false 00:09:01.978 }, 00:09:01.978 "memory_domains": [ 00:09:01.978 { 00:09:01.978 "dma_device_id": "system", 00:09:01.978 "dma_device_type": 1 00:09:01.978 }, 00:09:01.978 { 00:09:01.978 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:01.978 "dma_device_type": 2 00:09:01.978 } 00:09:01.978 ], 00:09:01.978 "driver_specific": {} 00:09:01.978 } 00:09:01.978 ] 00:09:01.978 23:42:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:01.978 23:42:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:09:01.978 23:42:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:01.978 23:42:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:01.978 23:42:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:01.978 23:42:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:01.978 23:42:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:01.978 23:42:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:01.978 23:42:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:01.978 23:42:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:01.978 23:42:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:01.978 23:42:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:01.978 23:42:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:01.978 23:42:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:01.978 23:42:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:01.978 23:42:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:02.238 23:42:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:02.238 23:42:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:02.238 "name": "Existed_Raid", 00:09:02.238 "uuid": "056c3fba-be90-45ba-8133-24e12ca385aa", 00:09:02.238 "strip_size_kb": 64, 00:09:02.238 "state": "configuring", 00:09:02.238 "raid_level": "raid0", 00:09:02.238 "superblock": true, 00:09:02.238 "num_base_bdevs": 4, 00:09:02.238 "num_base_bdevs_discovered": 3, 00:09:02.238 "num_base_bdevs_operational": 4, 00:09:02.238 "base_bdevs_list": [ 00:09:02.238 { 00:09:02.238 "name": "BaseBdev1", 00:09:02.238 "uuid": "16227b2c-b735-4ab7-bf87-f6ffa73fca82", 00:09:02.238 "is_configured": true, 00:09:02.238 "data_offset": 2048, 00:09:02.238 "data_size": 63488 00:09:02.238 }, 00:09:02.238 { 00:09:02.238 "name": null, 00:09:02.238 "uuid": "54fae0e9-e3aa-4910-90e9-0d240981390f", 00:09:02.238 "is_configured": false, 00:09:02.238 "data_offset": 0, 00:09:02.238 "data_size": 63488 00:09:02.238 }, 00:09:02.238 { 00:09:02.238 "name": "BaseBdev3", 00:09:02.238 "uuid": "84e6a8e7-7c16-479b-9900-221eed699b29", 00:09:02.238 "is_configured": true, 00:09:02.238 "data_offset": 2048, 00:09:02.238 "data_size": 63488 00:09:02.238 }, 00:09:02.238 { 00:09:02.238 "name": "BaseBdev4", 00:09:02.238 "uuid": "366edb40-a913-47b8-9642-56b06f51f309", 00:09:02.238 "is_configured": true, 00:09:02.238 "data_offset": 2048, 00:09:02.238 "data_size": 63488 00:09:02.238 } 00:09:02.238 ] 00:09:02.238 }' 00:09:02.238 23:42:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:02.238 23:42:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:02.498 23:42:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:09:02.498 23:42:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:02.498 23:42:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:02.498 23:42:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:02.498 23:42:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:02.498 23:42:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:09:02.498 23:42:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:09:02.498 23:42:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:02.498 23:42:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:02.498 [2024-11-26 23:42:50.571284] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:09:02.498 23:42:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:02.498 23:42:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:02.498 23:42:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:02.498 23:42:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:02.498 23:42:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:02.498 23:42:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:02.498 23:42:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:02.498 23:42:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:02.498 23:42:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:02.498 23:42:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:02.498 23:42:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:02.498 23:42:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:02.498 23:42:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:02.498 23:42:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:02.498 23:42:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:02.498 23:42:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:02.758 23:42:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:02.758 "name": "Existed_Raid", 00:09:02.758 "uuid": "056c3fba-be90-45ba-8133-24e12ca385aa", 00:09:02.758 "strip_size_kb": 64, 00:09:02.758 "state": "configuring", 00:09:02.758 "raid_level": "raid0", 00:09:02.758 "superblock": true, 00:09:02.758 "num_base_bdevs": 4, 00:09:02.758 "num_base_bdevs_discovered": 2, 00:09:02.758 "num_base_bdevs_operational": 4, 00:09:02.758 "base_bdevs_list": [ 00:09:02.758 { 00:09:02.758 "name": "BaseBdev1", 00:09:02.758 "uuid": "16227b2c-b735-4ab7-bf87-f6ffa73fca82", 00:09:02.758 "is_configured": true, 00:09:02.758 "data_offset": 2048, 00:09:02.758 "data_size": 63488 00:09:02.758 }, 00:09:02.758 { 00:09:02.758 "name": null, 00:09:02.758 "uuid": "54fae0e9-e3aa-4910-90e9-0d240981390f", 00:09:02.758 "is_configured": false, 00:09:02.758 "data_offset": 0, 00:09:02.758 "data_size": 63488 00:09:02.758 }, 00:09:02.758 { 00:09:02.758 "name": null, 00:09:02.758 "uuid": "84e6a8e7-7c16-479b-9900-221eed699b29", 00:09:02.758 "is_configured": false, 00:09:02.758 "data_offset": 0, 00:09:02.758 "data_size": 63488 00:09:02.758 }, 00:09:02.758 { 00:09:02.758 "name": "BaseBdev4", 00:09:02.758 "uuid": "366edb40-a913-47b8-9642-56b06f51f309", 00:09:02.758 "is_configured": true, 00:09:02.758 "data_offset": 2048, 00:09:02.758 "data_size": 63488 00:09:02.758 } 00:09:02.758 ] 00:09:02.758 }' 00:09:02.758 23:42:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:02.758 23:42:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:03.019 23:42:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:09:03.019 23:42:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:03.019 23:42:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:03.019 23:42:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:03.019 23:42:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:03.019 23:42:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:09:03.019 23:42:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:09:03.019 23:42:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:03.019 23:42:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:03.019 [2024-11-26 23:42:51.054473] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:03.019 23:42:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:03.019 23:42:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:03.019 23:42:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:03.019 23:42:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:03.019 23:42:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:03.019 23:42:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:03.019 23:42:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:03.019 23:42:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:03.019 23:42:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:03.019 23:42:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:03.019 23:42:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:03.019 23:42:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:03.019 23:42:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:03.019 23:42:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:03.019 23:42:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:03.019 23:42:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:03.019 23:42:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:03.019 "name": "Existed_Raid", 00:09:03.019 "uuid": "056c3fba-be90-45ba-8133-24e12ca385aa", 00:09:03.019 "strip_size_kb": 64, 00:09:03.019 "state": "configuring", 00:09:03.019 "raid_level": "raid0", 00:09:03.019 "superblock": true, 00:09:03.019 "num_base_bdevs": 4, 00:09:03.019 "num_base_bdevs_discovered": 3, 00:09:03.019 "num_base_bdevs_operational": 4, 00:09:03.019 "base_bdevs_list": [ 00:09:03.019 { 00:09:03.019 "name": "BaseBdev1", 00:09:03.019 "uuid": "16227b2c-b735-4ab7-bf87-f6ffa73fca82", 00:09:03.019 "is_configured": true, 00:09:03.019 "data_offset": 2048, 00:09:03.019 "data_size": 63488 00:09:03.019 }, 00:09:03.019 { 00:09:03.019 "name": null, 00:09:03.019 "uuid": "54fae0e9-e3aa-4910-90e9-0d240981390f", 00:09:03.019 "is_configured": false, 00:09:03.019 "data_offset": 0, 00:09:03.019 "data_size": 63488 00:09:03.019 }, 00:09:03.019 { 00:09:03.019 "name": "BaseBdev3", 00:09:03.019 "uuid": "84e6a8e7-7c16-479b-9900-221eed699b29", 00:09:03.019 "is_configured": true, 00:09:03.019 "data_offset": 2048, 00:09:03.019 "data_size": 63488 00:09:03.019 }, 00:09:03.019 { 00:09:03.019 "name": "BaseBdev4", 00:09:03.019 "uuid": "366edb40-a913-47b8-9642-56b06f51f309", 00:09:03.019 "is_configured": true, 00:09:03.019 "data_offset": 2048, 00:09:03.019 "data_size": 63488 00:09:03.019 } 00:09:03.019 ] 00:09:03.019 }' 00:09:03.019 23:42:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:03.019 23:42:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:03.589 23:42:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:09:03.589 23:42:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:03.589 23:42:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:03.589 23:42:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:03.589 23:42:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:03.589 23:42:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:09:03.589 23:42:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:09:03.589 23:42:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:03.590 23:42:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:03.590 [2024-11-26 23:42:51.577652] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:09:03.590 23:42:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:03.590 23:42:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:03.590 23:42:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:03.590 23:42:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:03.590 23:42:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:03.590 23:42:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:03.590 23:42:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:03.590 23:42:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:03.590 23:42:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:03.590 23:42:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:03.590 23:42:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:03.590 23:42:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:03.590 23:42:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:03.590 23:42:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:03.590 23:42:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:03.590 23:42:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:03.590 23:42:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:03.590 "name": "Existed_Raid", 00:09:03.590 "uuid": "056c3fba-be90-45ba-8133-24e12ca385aa", 00:09:03.590 "strip_size_kb": 64, 00:09:03.590 "state": "configuring", 00:09:03.590 "raid_level": "raid0", 00:09:03.590 "superblock": true, 00:09:03.590 "num_base_bdevs": 4, 00:09:03.590 "num_base_bdevs_discovered": 2, 00:09:03.590 "num_base_bdevs_operational": 4, 00:09:03.590 "base_bdevs_list": [ 00:09:03.590 { 00:09:03.590 "name": null, 00:09:03.590 "uuid": "16227b2c-b735-4ab7-bf87-f6ffa73fca82", 00:09:03.590 "is_configured": false, 00:09:03.590 "data_offset": 0, 00:09:03.590 "data_size": 63488 00:09:03.590 }, 00:09:03.590 { 00:09:03.590 "name": null, 00:09:03.590 "uuid": "54fae0e9-e3aa-4910-90e9-0d240981390f", 00:09:03.590 "is_configured": false, 00:09:03.590 "data_offset": 0, 00:09:03.590 "data_size": 63488 00:09:03.590 }, 00:09:03.590 { 00:09:03.590 "name": "BaseBdev3", 00:09:03.590 "uuid": "84e6a8e7-7c16-479b-9900-221eed699b29", 00:09:03.590 "is_configured": true, 00:09:03.590 "data_offset": 2048, 00:09:03.590 "data_size": 63488 00:09:03.590 }, 00:09:03.590 { 00:09:03.590 "name": "BaseBdev4", 00:09:03.590 "uuid": "366edb40-a913-47b8-9642-56b06f51f309", 00:09:03.590 "is_configured": true, 00:09:03.590 "data_offset": 2048, 00:09:03.590 "data_size": 63488 00:09:03.590 } 00:09:03.590 ] 00:09:03.590 }' 00:09:03.590 23:42:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:03.590 23:42:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:04.158 23:42:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:04.158 23:42:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:09:04.158 23:42:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:04.158 23:42:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:04.158 23:42:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:04.158 23:42:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:09:04.158 23:42:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:09:04.158 23:42:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:04.158 23:42:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:04.158 [2024-11-26 23:42:52.051320] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:04.158 23:42:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:04.158 23:42:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:04.158 23:42:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:04.158 23:42:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:04.158 23:42:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:04.158 23:42:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:04.158 23:42:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:04.158 23:42:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:04.158 23:42:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:04.158 23:42:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:04.158 23:42:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:04.158 23:42:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:04.158 23:42:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:04.158 23:42:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:04.158 23:42:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:04.158 23:42:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:04.158 23:42:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:04.158 "name": "Existed_Raid", 00:09:04.158 "uuid": "056c3fba-be90-45ba-8133-24e12ca385aa", 00:09:04.158 "strip_size_kb": 64, 00:09:04.158 "state": "configuring", 00:09:04.158 "raid_level": "raid0", 00:09:04.158 "superblock": true, 00:09:04.158 "num_base_bdevs": 4, 00:09:04.158 "num_base_bdevs_discovered": 3, 00:09:04.158 "num_base_bdevs_operational": 4, 00:09:04.158 "base_bdevs_list": [ 00:09:04.158 { 00:09:04.158 "name": null, 00:09:04.158 "uuid": "16227b2c-b735-4ab7-bf87-f6ffa73fca82", 00:09:04.158 "is_configured": false, 00:09:04.158 "data_offset": 0, 00:09:04.158 "data_size": 63488 00:09:04.158 }, 00:09:04.158 { 00:09:04.158 "name": "BaseBdev2", 00:09:04.158 "uuid": "54fae0e9-e3aa-4910-90e9-0d240981390f", 00:09:04.158 "is_configured": true, 00:09:04.158 "data_offset": 2048, 00:09:04.158 "data_size": 63488 00:09:04.158 }, 00:09:04.158 { 00:09:04.158 "name": "BaseBdev3", 00:09:04.158 "uuid": "84e6a8e7-7c16-479b-9900-221eed699b29", 00:09:04.159 "is_configured": true, 00:09:04.159 "data_offset": 2048, 00:09:04.159 "data_size": 63488 00:09:04.159 }, 00:09:04.159 { 00:09:04.159 "name": "BaseBdev4", 00:09:04.159 "uuid": "366edb40-a913-47b8-9642-56b06f51f309", 00:09:04.159 "is_configured": true, 00:09:04.159 "data_offset": 2048, 00:09:04.159 "data_size": 63488 00:09:04.159 } 00:09:04.159 ] 00:09:04.159 }' 00:09:04.159 23:42:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:04.159 23:42:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:04.417 23:42:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:09:04.417 23:42:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:04.417 23:42:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:04.417 23:42:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:04.417 23:42:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:04.417 23:42:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:09:04.417 23:42:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:04.417 23:42:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:04.417 23:42:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:04.417 23:42:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:09:04.677 23:42:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:04.677 23:42:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 16227b2c-b735-4ab7-bf87-f6ffa73fca82 00:09:04.677 23:42:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:04.677 23:42:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:04.677 [2024-11-26 23:42:52.597508] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:09:04.677 [2024-11-26 23:42:52.597697] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:09:04.677 [2024-11-26 23:42:52.597710] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:09:04.677 [2024-11-26 23:42:52.597951] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002a10 00:09:04.677 NewBaseBdev 00:09:04.677 [2024-11-26 23:42:52.598056] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:09:04.678 [2024-11-26 23:42:52.598066] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001c80 00:09:04.678 [2024-11-26 23:42:52.598182] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:04.678 23:42:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:04.678 23:42:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:09:04.678 23:42:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:09:04.678 23:42:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:04.678 23:42:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:09:04.678 23:42:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:04.678 23:42:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:04.678 23:42:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:04.678 23:42:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:04.678 23:42:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:04.678 23:42:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:04.678 23:42:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:09:04.678 23:42:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:04.678 23:42:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:04.678 [ 00:09:04.678 { 00:09:04.678 "name": "NewBaseBdev", 00:09:04.678 "aliases": [ 00:09:04.678 "16227b2c-b735-4ab7-bf87-f6ffa73fca82" 00:09:04.678 ], 00:09:04.678 "product_name": "Malloc disk", 00:09:04.678 "block_size": 512, 00:09:04.678 "num_blocks": 65536, 00:09:04.678 "uuid": "16227b2c-b735-4ab7-bf87-f6ffa73fca82", 00:09:04.678 "assigned_rate_limits": { 00:09:04.678 "rw_ios_per_sec": 0, 00:09:04.678 "rw_mbytes_per_sec": 0, 00:09:04.678 "r_mbytes_per_sec": 0, 00:09:04.678 "w_mbytes_per_sec": 0 00:09:04.678 }, 00:09:04.678 "claimed": true, 00:09:04.678 "claim_type": "exclusive_write", 00:09:04.678 "zoned": false, 00:09:04.678 "supported_io_types": { 00:09:04.678 "read": true, 00:09:04.678 "write": true, 00:09:04.678 "unmap": true, 00:09:04.678 "flush": true, 00:09:04.678 "reset": true, 00:09:04.678 "nvme_admin": false, 00:09:04.678 "nvme_io": false, 00:09:04.678 "nvme_io_md": false, 00:09:04.678 "write_zeroes": true, 00:09:04.678 "zcopy": true, 00:09:04.678 "get_zone_info": false, 00:09:04.678 "zone_management": false, 00:09:04.678 "zone_append": false, 00:09:04.678 "compare": false, 00:09:04.678 "compare_and_write": false, 00:09:04.678 "abort": true, 00:09:04.678 "seek_hole": false, 00:09:04.678 "seek_data": false, 00:09:04.678 "copy": true, 00:09:04.678 "nvme_iov_md": false 00:09:04.678 }, 00:09:04.678 "memory_domains": [ 00:09:04.678 { 00:09:04.678 "dma_device_id": "system", 00:09:04.678 "dma_device_type": 1 00:09:04.678 }, 00:09:04.678 { 00:09:04.678 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:04.678 "dma_device_type": 2 00:09:04.678 } 00:09:04.678 ], 00:09:04.678 "driver_specific": {} 00:09:04.678 } 00:09:04.678 ] 00:09:04.678 23:42:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:04.678 23:42:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:09:04.678 23:42:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid0 64 4 00:09:04.678 23:42:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:04.678 23:42:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:04.678 23:42:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:04.678 23:42:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:04.678 23:42:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:04.678 23:42:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:04.678 23:42:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:04.678 23:42:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:04.678 23:42:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:04.678 23:42:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:04.678 23:42:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:04.678 23:42:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:04.678 23:42:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:04.678 23:42:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:04.678 23:42:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:04.678 "name": "Existed_Raid", 00:09:04.678 "uuid": "056c3fba-be90-45ba-8133-24e12ca385aa", 00:09:04.678 "strip_size_kb": 64, 00:09:04.678 "state": "online", 00:09:04.678 "raid_level": "raid0", 00:09:04.678 "superblock": true, 00:09:04.678 "num_base_bdevs": 4, 00:09:04.678 "num_base_bdevs_discovered": 4, 00:09:04.678 "num_base_bdevs_operational": 4, 00:09:04.678 "base_bdevs_list": [ 00:09:04.678 { 00:09:04.678 "name": "NewBaseBdev", 00:09:04.678 "uuid": "16227b2c-b735-4ab7-bf87-f6ffa73fca82", 00:09:04.678 "is_configured": true, 00:09:04.678 "data_offset": 2048, 00:09:04.678 "data_size": 63488 00:09:04.678 }, 00:09:04.678 { 00:09:04.678 "name": "BaseBdev2", 00:09:04.678 "uuid": "54fae0e9-e3aa-4910-90e9-0d240981390f", 00:09:04.678 "is_configured": true, 00:09:04.678 "data_offset": 2048, 00:09:04.678 "data_size": 63488 00:09:04.678 }, 00:09:04.678 { 00:09:04.678 "name": "BaseBdev3", 00:09:04.678 "uuid": "84e6a8e7-7c16-479b-9900-221eed699b29", 00:09:04.678 "is_configured": true, 00:09:04.678 "data_offset": 2048, 00:09:04.678 "data_size": 63488 00:09:04.678 }, 00:09:04.678 { 00:09:04.678 "name": "BaseBdev4", 00:09:04.678 "uuid": "366edb40-a913-47b8-9642-56b06f51f309", 00:09:04.678 "is_configured": true, 00:09:04.678 "data_offset": 2048, 00:09:04.678 "data_size": 63488 00:09:04.678 } 00:09:04.678 ] 00:09:04.678 }' 00:09:04.678 23:42:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:04.678 23:42:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:05.248 23:42:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:09:05.248 23:42:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:09:05.248 23:42:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:05.248 23:42:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:05.248 23:42:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:09:05.248 23:42:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:05.248 23:42:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:09:05.248 23:42:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:05.248 23:42:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:05.248 23:42:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:05.248 [2024-11-26 23:42:53.101081] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:05.248 23:42:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:05.248 23:42:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:05.248 "name": "Existed_Raid", 00:09:05.248 "aliases": [ 00:09:05.248 "056c3fba-be90-45ba-8133-24e12ca385aa" 00:09:05.248 ], 00:09:05.248 "product_name": "Raid Volume", 00:09:05.248 "block_size": 512, 00:09:05.248 "num_blocks": 253952, 00:09:05.248 "uuid": "056c3fba-be90-45ba-8133-24e12ca385aa", 00:09:05.248 "assigned_rate_limits": { 00:09:05.248 "rw_ios_per_sec": 0, 00:09:05.248 "rw_mbytes_per_sec": 0, 00:09:05.248 "r_mbytes_per_sec": 0, 00:09:05.248 "w_mbytes_per_sec": 0 00:09:05.248 }, 00:09:05.248 "claimed": false, 00:09:05.248 "zoned": false, 00:09:05.248 "supported_io_types": { 00:09:05.248 "read": true, 00:09:05.248 "write": true, 00:09:05.248 "unmap": true, 00:09:05.248 "flush": true, 00:09:05.248 "reset": true, 00:09:05.248 "nvme_admin": false, 00:09:05.248 "nvme_io": false, 00:09:05.248 "nvme_io_md": false, 00:09:05.248 "write_zeroes": true, 00:09:05.248 "zcopy": false, 00:09:05.248 "get_zone_info": false, 00:09:05.248 "zone_management": false, 00:09:05.248 "zone_append": false, 00:09:05.248 "compare": false, 00:09:05.248 "compare_and_write": false, 00:09:05.248 "abort": false, 00:09:05.248 "seek_hole": false, 00:09:05.248 "seek_data": false, 00:09:05.248 "copy": false, 00:09:05.248 "nvme_iov_md": false 00:09:05.248 }, 00:09:05.248 "memory_domains": [ 00:09:05.248 { 00:09:05.248 "dma_device_id": "system", 00:09:05.248 "dma_device_type": 1 00:09:05.248 }, 00:09:05.248 { 00:09:05.248 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:05.248 "dma_device_type": 2 00:09:05.248 }, 00:09:05.248 { 00:09:05.248 "dma_device_id": "system", 00:09:05.248 "dma_device_type": 1 00:09:05.248 }, 00:09:05.248 { 00:09:05.248 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:05.248 "dma_device_type": 2 00:09:05.248 }, 00:09:05.248 { 00:09:05.248 "dma_device_id": "system", 00:09:05.248 "dma_device_type": 1 00:09:05.248 }, 00:09:05.248 { 00:09:05.248 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:05.248 "dma_device_type": 2 00:09:05.248 }, 00:09:05.248 { 00:09:05.248 "dma_device_id": "system", 00:09:05.248 "dma_device_type": 1 00:09:05.248 }, 00:09:05.248 { 00:09:05.248 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:05.248 "dma_device_type": 2 00:09:05.248 } 00:09:05.248 ], 00:09:05.248 "driver_specific": { 00:09:05.248 "raid": { 00:09:05.248 "uuid": "056c3fba-be90-45ba-8133-24e12ca385aa", 00:09:05.248 "strip_size_kb": 64, 00:09:05.248 "state": "online", 00:09:05.249 "raid_level": "raid0", 00:09:05.249 "superblock": true, 00:09:05.249 "num_base_bdevs": 4, 00:09:05.249 "num_base_bdevs_discovered": 4, 00:09:05.249 "num_base_bdevs_operational": 4, 00:09:05.249 "base_bdevs_list": [ 00:09:05.249 { 00:09:05.249 "name": "NewBaseBdev", 00:09:05.249 "uuid": "16227b2c-b735-4ab7-bf87-f6ffa73fca82", 00:09:05.249 "is_configured": true, 00:09:05.249 "data_offset": 2048, 00:09:05.249 "data_size": 63488 00:09:05.249 }, 00:09:05.249 { 00:09:05.249 "name": "BaseBdev2", 00:09:05.249 "uuid": "54fae0e9-e3aa-4910-90e9-0d240981390f", 00:09:05.249 "is_configured": true, 00:09:05.249 "data_offset": 2048, 00:09:05.249 "data_size": 63488 00:09:05.249 }, 00:09:05.249 { 00:09:05.249 "name": "BaseBdev3", 00:09:05.249 "uuid": "84e6a8e7-7c16-479b-9900-221eed699b29", 00:09:05.249 "is_configured": true, 00:09:05.249 "data_offset": 2048, 00:09:05.249 "data_size": 63488 00:09:05.249 }, 00:09:05.249 { 00:09:05.249 "name": "BaseBdev4", 00:09:05.249 "uuid": "366edb40-a913-47b8-9642-56b06f51f309", 00:09:05.249 "is_configured": true, 00:09:05.249 "data_offset": 2048, 00:09:05.249 "data_size": 63488 00:09:05.249 } 00:09:05.249 ] 00:09:05.249 } 00:09:05.249 } 00:09:05.249 }' 00:09:05.249 23:42:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:05.249 23:42:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:09:05.249 BaseBdev2 00:09:05.249 BaseBdev3 00:09:05.249 BaseBdev4' 00:09:05.249 23:42:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:05.249 23:42:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:05.249 23:42:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:05.249 23:42:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:09:05.249 23:42:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:05.249 23:42:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:05.249 23:42:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:05.249 23:42:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:05.249 23:42:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:05.249 23:42:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:05.249 23:42:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:05.249 23:42:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:09:05.249 23:42:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:05.249 23:42:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:05.249 23:42:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:05.249 23:42:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:05.249 23:42:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:05.249 23:42:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:05.249 23:42:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:05.249 23:42:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:09:05.249 23:42:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:05.249 23:42:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:05.249 23:42:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:05.249 23:42:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:05.509 23:42:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:05.509 23:42:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:05.509 23:42:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:05.509 23:42:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:05.509 23:42:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:09:05.509 23:42:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:05.509 23:42:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:05.509 23:42:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:05.509 23:42:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:05.509 23:42:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:05.509 23:42:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:05.509 23:42:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:05.509 23:42:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:05.509 [2024-11-26 23:42:53.452177] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:05.509 [2024-11-26 23:42:53.452209] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:05.509 [2024-11-26 23:42:53.452298] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:05.509 [2024-11-26 23:42:53.452378] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:05.509 [2024-11-26 23:42:53.452388] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name Existed_Raid, state offline 00:09:05.509 23:42:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:05.509 23:42:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 80723 00:09:05.509 23:42:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # '[' -z 80723 ']' 00:09:05.509 23:42:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@958 -- # kill -0 80723 00:09:05.509 23:42:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # uname 00:09:05.509 23:42:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:09:05.509 23:42:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 80723 00:09:05.509 23:42:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:09:05.509 23:42:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:09:05.509 23:42:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@972 -- # echo 'killing process with pid 80723' 00:09:05.509 killing process with pid 80723 00:09:05.509 23:42:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@973 -- # kill 80723 00:09:05.509 [2024-11-26 23:42:53.501529] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:09:05.509 23:42:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@978 -- # wait 80723 00:09:05.509 [2024-11-26 23:42:53.542166] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:09:05.785 23:42:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:09:05.785 00:09:05.785 real 0m9.751s 00:09:05.785 user 0m16.776s 00:09:05.785 sys 0m2.000s 00:09:05.785 23:42:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1130 -- # xtrace_disable 00:09:05.785 23:42:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:05.785 ************************************ 00:09:05.785 END TEST raid_state_function_test_sb 00:09:05.785 ************************************ 00:09:05.785 23:42:53 bdev_raid -- bdev/bdev_raid.sh@970 -- # run_test raid_superblock_test raid_superblock_test raid0 4 00:09:05.785 23:42:53 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:09:05.785 23:42:53 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:09:05.785 23:42:53 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:09:05.785 ************************************ 00:09:05.785 START TEST raid_superblock_test 00:09:05.785 ************************************ 00:09:05.785 23:42:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1129 -- # raid_superblock_test raid0 4 00:09:05.785 23:42:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=raid0 00:09:05.785 23:42:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=4 00:09:05.785 23:42:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:09:05.785 23:42:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:09:05.785 23:42:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:09:05.785 23:42:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:09:05.785 23:42:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:09:05.785 23:42:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:09:05.785 23:42:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:09:05.785 23:42:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:09:05.785 23:42:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:09:05.785 23:42:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:09:05.785 23:42:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:09:05.785 23:42:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' raid0 '!=' raid1 ']' 00:09:05.785 23:42:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@405 -- # strip_size=64 00:09:05.785 23:42:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@406 -- # strip_size_create_arg='-z 64' 00:09:05.785 23:42:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=81371 00:09:05.785 23:42:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:09:05.785 23:42:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 81371 00:09:05.785 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:05.785 23:42:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@835 -- # '[' -z 81371 ']' 00:09:05.785 23:42:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:05.785 23:42:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:09:05.785 23:42:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:05.785 23:42:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:09:05.785 23:42:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:05.785 [2024-11-26 23:42:53.904990] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:09:05.785 [2024-11-26 23:42:53.905203] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid81371 ] 00:09:06.045 [2024-11-26 23:42:54.056257] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:06.045 [2024-11-26 23:42:54.083791] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:09:06.045 [2024-11-26 23:42:54.127401] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:06.045 [2024-11-26 23:42:54.127524] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:06.984 23:42:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:09:06.984 23:42:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@868 -- # return 0 00:09:06.984 23:42:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:09:06.984 23:42:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:09:06.984 23:42:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:09:06.984 23:42:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:09:06.984 23:42:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:09:06.984 23:42:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:09:06.984 23:42:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:09:06.984 23:42:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:09:06.984 23:42:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:09:06.984 23:42:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:06.984 23:42:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:06.984 malloc1 00:09:06.984 23:42:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:06.984 23:42:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:09:06.984 23:42:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:06.984 23:42:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:06.984 [2024-11-26 23:42:54.799543] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:09:06.984 [2024-11-26 23:42:54.799608] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:06.984 [2024-11-26 23:42:54.799643] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:09:06.984 [2024-11-26 23:42:54.799666] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:06.984 [2024-11-26 23:42:54.801760] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:06.984 [2024-11-26 23:42:54.801801] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:09:06.984 pt1 00:09:06.984 23:42:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:06.984 23:42:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:09:06.984 23:42:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:09:06.984 23:42:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:09:06.984 23:42:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:09:06.984 23:42:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:09:06.984 23:42:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:09:06.984 23:42:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:09:06.984 23:42:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:09:06.984 23:42:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:09:06.984 23:42:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:06.984 23:42:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:06.984 malloc2 00:09:06.984 23:42:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:06.984 23:42:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:09:06.984 23:42:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:06.984 23:42:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:06.984 [2024-11-26 23:42:54.828063] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:09:06.984 [2024-11-26 23:42:54.828161] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:06.984 [2024-11-26 23:42:54.828212] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:09:06.984 [2024-11-26 23:42:54.828241] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:06.984 [2024-11-26 23:42:54.830388] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:06.984 [2024-11-26 23:42:54.830459] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:09:06.984 pt2 00:09:06.984 23:42:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:06.984 23:42:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:09:06.984 23:42:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:09:06.984 23:42:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc3 00:09:06.984 23:42:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt3 00:09:06.984 23:42:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000003 00:09:06.984 23:42:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:09:06.984 23:42:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:09:06.984 23:42:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:09:06.984 23:42:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc3 00:09:06.985 23:42:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:06.985 23:42:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:06.985 malloc3 00:09:06.985 23:42:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:06.985 23:42:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:09:06.985 23:42:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:06.985 23:42:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:06.985 [2024-11-26 23:42:54.860652] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:09:06.985 [2024-11-26 23:42:54.860747] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:06.985 [2024-11-26 23:42:54.860780] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:09:06.985 [2024-11-26 23:42:54.860809] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:06.985 [2024-11-26 23:42:54.862946] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:06.985 [2024-11-26 23:42:54.863037] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:09:06.985 pt3 00:09:06.985 23:42:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:06.985 23:42:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:09:06.985 23:42:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:09:06.985 23:42:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc4 00:09:06.985 23:42:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt4 00:09:06.985 23:42:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000004 00:09:06.985 23:42:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:09:06.985 23:42:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:09:06.985 23:42:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:09:06.985 23:42:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc4 00:09:06.985 23:42:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:06.985 23:42:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:06.985 malloc4 00:09:06.985 23:42:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:06.985 23:42:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:09:06.985 23:42:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:06.985 23:42:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:06.985 [2024-11-26 23:42:54.903841] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:09:06.985 [2024-11-26 23:42:54.903894] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:06.985 [2024-11-26 23:42:54.903909] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:09:06.985 [2024-11-26 23:42:54.903921] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:06.985 [2024-11-26 23:42:54.905995] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:06.985 [2024-11-26 23:42:54.906032] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:09:06.985 pt4 00:09:06.985 23:42:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:06.985 23:42:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:09:06.985 23:42:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:09:06.985 23:42:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''pt1 pt2 pt3 pt4'\''' -n raid_bdev1 -s 00:09:06.985 23:42:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:06.985 23:42:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:06.985 [2024-11-26 23:42:54.915865] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:09:06.985 [2024-11-26 23:42:54.917813] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:09:06.985 [2024-11-26 23:42:54.917879] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:09:06.985 [2024-11-26 23:42:54.917946] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:09:06.985 [2024-11-26 23:42:54.918098] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:09:06.985 [2024-11-26 23:42:54.918112] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:09:06.985 [2024-11-26 23:42:54.918387] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:09:06.985 [2024-11-26 23:42:54.918524] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:09:06.985 [2024-11-26 23:42:54.918533] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:09:06.985 [2024-11-26 23:42:54.918656] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:06.985 23:42:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:06.985 23:42:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 4 00:09:06.985 23:42:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:06.985 23:42:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:06.985 23:42:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:06.985 23:42:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:06.985 23:42:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:06.985 23:42:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:06.985 23:42:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:06.985 23:42:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:06.985 23:42:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:06.985 23:42:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:06.985 23:42:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:06.985 23:42:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:06.985 23:42:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:06.985 23:42:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:06.985 23:42:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:06.985 "name": "raid_bdev1", 00:09:06.985 "uuid": "d151ea8c-de3f-4bab-8dbf-26742b8f64ab", 00:09:06.985 "strip_size_kb": 64, 00:09:06.985 "state": "online", 00:09:06.985 "raid_level": "raid0", 00:09:06.985 "superblock": true, 00:09:06.985 "num_base_bdevs": 4, 00:09:06.985 "num_base_bdevs_discovered": 4, 00:09:06.985 "num_base_bdevs_operational": 4, 00:09:06.985 "base_bdevs_list": [ 00:09:06.985 { 00:09:06.985 "name": "pt1", 00:09:06.985 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:06.985 "is_configured": true, 00:09:06.985 "data_offset": 2048, 00:09:06.985 "data_size": 63488 00:09:06.985 }, 00:09:06.985 { 00:09:06.985 "name": "pt2", 00:09:06.985 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:06.985 "is_configured": true, 00:09:06.985 "data_offset": 2048, 00:09:06.985 "data_size": 63488 00:09:06.985 }, 00:09:06.985 { 00:09:06.985 "name": "pt3", 00:09:06.985 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:06.985 "is_configured": true, 00:09:06.985 "data_offset": 2048, 00:09:06.985 "data_size": 63488 00:09:06.985 }, 00:09:06.985 { 00:09:06.985 "name": "pt4", 00:09:06.985 "uuid": "00000000-0000-0000-0000-000000000004", 00:09:06.985 "is_configured": true, 00:09:06.985 "data_offset": 2048, 00:09:06.985 "data_size": 63488 00:09:06.985 } 00:09:06.985 ] 00:09:06.985 }' 00:09:06.985 23:42:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:06.985 23:42:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:07.558 23:42:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:09:07.558 23:42:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:09:07.558 23:42:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:07.558 23:42:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:07.558 23:42:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:09:07.558 23:42:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:07.558 23:42:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:09:07.558 23:42:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:07.558 23:42:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:07.558 23:42:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:07.558 [2024-11-26 23:42:55.415331] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:07.558 23:42:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:07.558 23:42:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:07.558 "name": "raid_bdev1", 00:09:07.558 "aliases": [ 00:09:07.558 "d151ea8c-de3f-4bab-8dbf-26742b8f64ab" 00:09:07.558 ], 00:09:07.558 "product_name": "Raid Volume", 00:09:07.558 "block_size": 512, 00:09:07.558 "num_blocks": 253952, 00:09:07.558 "uuid": "d151ea8c-de3f-4bab-8dbf-26742b8f64ab", 00:09:07.558 "assigned_rate_limits": { 00:09:07.558 "rw_ios_per_sec": 0, 00:09:07.558 "rw_mbytes_per_sec": 0, 00:09:07.558 "r_mbytes_per_sec": 0, 00:09:07.558 "w_mbytes_per_sec": 0 00:09:07.558 }, 00:09:07.558 "claimed": false, 00:09:07.558 "zoned": false, 00:09:07.558 "supported_io_types": { 00:09:07.558 "read": true, 00:09:07.558 "write": true, 00:09:07.558 "unmap": true, 00:09:07.558 "flush": true, 00:09:07.558 "reset": true, 00:09:07.558 "nvme_admin": false, 00:09:07.558 "nvme_io": false, 00:09:07.558 "nvme_io_md": false, 00:09:07.558 "write_zeroes": true, 00:09:07.558 "zcopy": false, 00:09:07.558 "get_zone_info": false, 00:09:07.558 "zone_management": false, 00:09:07.558 "zone_append": false, 00:09:07.558 "compare": false, 00:09:07.558 "compare_and_write": false, 00:09:07.558 "abort": false, 00:09:07.558 "seek_hole": false, 00:09:07.558 "seek_data": false, 00:09:07.558 "copy": false, 00:09:07.558 "nvme_iov_md": false 00:09:07.558 }, 00:09:07.558 "memory_domains": [ 00:09:07.558 { 00:09:07.558 "dma_device_id": "system", 00:09:07.558 "dma_device_type": 1 00:09:07.558 }, 00:09:07.558 { 00:09:07.558 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:07.558 "dma_device_type": 2 00:09:07.558 }, 00:09:07.558 { 00:09:07.558 "dma_device_id": "system", 00:09:07.558 "dma_device_type": 1 00:09:07.558 }, 00:09:07.558 { 00:09:07.558 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:07.558 "dma_device_type": 2 00:09:07.558 }, 00:09:07.558 { 00:09:07.558 "dma_device_id": "system", 00:09:07.559 "dma_device_type": 1 00:09:07.559 }, 00:09:07.559 { 00:09:07.559 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:07.559 "dma_device_type": 2 00:09:07.559 }, 00:09:07.559 { 00:09:07.559 "dma_device_id": "system", 00:09:07.559 "dma_device_type": 1 00:09:07.559 }, 00:09:07.559 { 00:09:07.559 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:07.559 "dma_device_type": 2 00:09:07.559 } 00:09:07.559 ], 00:09:07.559 "driver_specific": { 00:09:07.559 "raid": { 00:09:07.559 "uuid": "d151ea8c-de3f-4bab-8dbf-26742b8f64ab", 00:09:07.559 "strip_size_kb": 64, 00:09:07.559 "state": "online", 00:09:07.559 "raid_level": "raid0", 00:09:07.559 "superblock": true, 00:09:07.559 "num_base_bdevs": 4, 00:09:07.559 "num_base_bdevs_discovered": 4, 00:09:07.559 "num_base_bdevs_operational": 4, 00:09:07.559 "base_bdevs_list": [ 00:09:07.559 { 00:09:07.559 "name": "pt1", 00:09:07.559 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:07.559 "is_configured": true, 00:09:07.559 "data_offset": 2048, 00:09:07.559 "data_size": 63488 00:09:07.559 }, 00:09:07.559 { 00:09:07.559 "name": "pt2", 00:09:07.559 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:07.559 "is_configured": true, 00:09:07.559 "data_offset": 2048, 00:09:07.559 "data_size": 63488 00:09:07.559 }, 00:09:07.559 { 00:09:07.559 "name": "pt3", 00:09:07.559 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:07.559 "is_configured": true, 00:09:07.559 "data_offset": 2048, 00:09:07.559 "data_size": 63488 00:09:07.559 }, 00:09:07.559 { 00:09:07.559 "name": "pt4", 00:09:07.559 "uuid": "00000000-0000-0000-0000-000000000004", 00:09:07.559 "is_configured": true, 00:09:07.559 "data_offset": 2048, 00:09:07.559 "data_size": 63488 00:09:07.559 } 00:09:07.559 ] 00:09:07.559 } 00:09:07.559 } 00:09:07.559 }' 00:09:07.559 23:42:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:07.559 23:42:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:09:07.559 pt2 00:09:07.559 pt3 00:09:07.559 pt4' 00:09:07.559 23:42:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:07.559 23:42:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:07.559 23:42:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:07.559 23:42:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:09:07.559 23:42:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:07.559 23:42:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:07.559 23:42:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:07.559 23:42:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:07.559 23:42:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:07.559 23:42:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:07.559 23:42:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:07.559 23:42:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:09:07.559 23:42:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:07.559 23:42:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:07.559 23:42:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:07.559 23:42:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:07.559 23:42:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:07.559 23:42:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:07.559 23:42:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:07.559 23:42:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:09:07.559 23:42:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:07.559 23:42:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:07.559 23:42:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:07.559 23:42:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:07.559 23:42:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:07.559 23:42:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:07.559 23:42:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:07.559 23:42:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:07.559 23:42:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt4 00:09:07.559 23:42:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:07.559 23:42:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:07.825 23:42:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:07.826 23:42:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:07.826 23:42:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:07.826 23:42:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:09:07.826 23:42:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:09:07.826 23:42:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:07.826 23:42:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:07.826 [2024-11-26 23:42:55.722727] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:07.826 23:42:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:07.826 23:42:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=d151ea8c-de3f-4bab-8dbf-26742b8f64ab 00:09:07.826 23:42:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z d151ea8c-de3f-4bab-8dbf-26742b8f64ab ']' 00:09:07.826 23:42:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:09:07.826 23:42:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:07.826 23:42:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:07.826 [2024-11-26 23:42:55.766384] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:07.826 [2024-11-26 23:42:55.766452] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:07.826 [2024-11-26 23:42:55.766572] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:07.826 [2024-11-26 23:42:55.766665] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:07.826 [2024-11-26 23:42:55.766734] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:09:07.826 23:42:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:07.826 23:42:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:07.826 23:42:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:07.826 23:42:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:07.826 23:42:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:09:07.826 23:42:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:07.826 23:42:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:09:07.826 23:42:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:09:07.826 23:42:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:09:07.826 23:42:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:09:07.826 23:42:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:07.826 23:42:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:07.826 23:42:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:07.826 23:42:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:09:07.826 23:42:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:09:07.826 23:42:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:07.826 23:42:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:07.826 23:42:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:07.826 23:42:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:09:07.826 23:42:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt3 00:09:07.826 23:42:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:07.826 23:42:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:07.826 23:42:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:07.826 23:42:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:09:07.826 23:42:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt4 00:09:07.826 23:42:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:07.826 23:42:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:07.826 23:42:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:07.826 23:42:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:09:07.826 23:42:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:07.826 23:42:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:07.826 23:42:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:09:07.826 23:42:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:07.826 23:42:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:09:07.826 23:42:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:09:07.826 23:42:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@652 -- # local es=0 00:09:07.826 23:42:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:09:07.826 23:42:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:09:07.826 23:42:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:09:07.826 23:42:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:09:07.826 23:42:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:09:07.826 23:42:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:09:07.826 23:42:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:07.826 23:42:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:07.826 [2024-11-26 23:42:55.914159] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:09:07.826 [2024-11-26 23:42:55.916069] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:09:07.826 [2024-11-26 23:42:55.916117] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc3 is claimed 00:09:07.826 [2024-11-26 23:42:55.916144] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc4 is claimed 00:09:07.826 [2024-11-26 23:42:55.916190] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:09:07.826 [2024-11-26 23:42:55.916233] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:09:07.826 [2024-11-26 23:42:55.916252] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc3 00:09:07.826 [2024-11-26 23:42:55.916267] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc4 00:09:07.826 [2024-11-26 23:42:55.916280] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:07.826 [2024-11-26 23:42:55.916290] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state configuring 00:09:07.826 request: 00:09:07.826 { 00:09:07.826 "name": "raid_bdev1", 00:09:07.826 "raid_level": "raid0", 00:09:07.826 "base_bdevs": [ 00:09:07.826 "malloc1", 00:09:07.826 "malloc2", 00:09:07.826 "malloc3", 00:09:07.826 "malloc4" 00:09:07.826 ], 00:09:07.826 "strip_size_kb": 64, 00:09:07.826 "superblock": false, 00:09:07.826 "method": "bdev_raid_create", 00:09:07.826 "req_id": 1 00:09:07.826 } 00:09:07.826 Got JSON-RPC error response 00:09:07.826 response: 00:09:07.826 { 00:09:07.826 "code": -17, 00:09:07.826 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:09:07.826 } 00:09:07.826 23:42:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:09:07.826 23:42:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # es=1 00:09:07.826 23:42:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:09:07.826 23:42:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:09:07.826 23:42:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:09:07.826 23:42:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:07.826 23:42:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:09:07.826 23:42:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:07.826 23:42:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:07.826 23:42:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:08.085 23:42:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:09:08.085 23:42:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:09:08.085 23:42:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:09:08.085 23:42:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:08.085 23:42:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:08.085 [2024-11-26 23:42:55.982003] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:09:08.085 [2024-11-26 23:42:55.982101] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:08.085 [2024-11-26 23:42:55.982165] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:09:08.085 [2024-11-26 23:42:55.982194] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:08.085 [2024-11-26 23:42:55.984329] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:08.085 [2024-11-26 23:42:55.984407] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:09:08.085 [2024-11-26 23:42:55.984515] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:09:08.085 [2024-11-26 23:42:55.984573] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:09:08.085 pt1 00:09:08.085 23:42:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:08.085 23:42:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid0 64 4 00:09:08.085 23:42:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:08.085 23:42:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:08.085 23:42:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:08.085 23:42:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:08.085 23:42:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:08.085 23:42:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:08.085 23:42:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:08.085 23:42:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:08.085 23:42:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:08.085 23:42:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:08.085 23:42:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:08.085 23:42:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:08.085 23:42:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:08.085 23:42:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:08.085 23:42:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:08.085 "name": "raid_bdev1", 00:09:08.085 "uuid": "d151ea8c-de3f-4bab-8dbf-26742b8f64ab", 00:09:08.085 "strip_size_kb": 64, 00:09:08.085 "state": "configuring", 00:09:08.085 "raid_level": "raid0", 00:09:08.085 "superblock": true, 00:09:08.085 "num_base_bdevs": 4, 00:09:08.085 "num_base_bdevs_discovered": 1, 00:09:08.085 "num_base_bdevs_operational": 4, 00:09:08.085 "base_bdevs_list": [ 00:09:08.085 { 00:09:08.085 "name": "pt1", 00:09:08.085 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:08.085 "is_configured": true, 00:09:08.085 "data_offset": 2048, 00:09:08.085 "data_size": 63488 00:09:08.085 }, 00:09:08.085 { 00:09:08.085 "name": null, 00:09:08.085 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:08.085 "is_configured": false, 00:09:08.085 "data_offset": 2048, 00:09:08.085 "data_size": 63488 00:09:08.085 }, 00:09:08.085 { 00:09:08.085 "name": null, 00:09:08.085 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:08.085 "is_configured": false, 00:09:08.085 "data_offset": 2048, 00:09:08.085 "data_size": 63488 00:09:08.085 }, 00:09:08.085 { 00:09:08.085 "name": null, 00:09:08.085 "uuid": "00000000-0000-0000-0000-000000000004", 00:09:08.085 "is_configured": false, 00:09:08.085 "data_offset": 2048, 00:09:08.085 "data_size": 63488 00:09:08.085 } 00:09:08.085 ] 00:09:08.085 }' 00:09:08.085 23:42:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:08.085 23:42:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:08.343 23:42:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 4 -gt 2 ']' 00:09:08.343 23:42:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@472 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:09:08.343 23:42:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:08.343 23:42:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:08.343 [2024-11-26 23:42:56.425285] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:09:08.343 [2024-11-26 23:42:56.425414] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:08.343 [2024-11-26 23:42:56.425441] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009c80 00:09:08.343 [2024-11-26 23:42:56.425450] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:08.343 [2024-11-26 23:42:56.425872] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:08.343 [2024-11-26 23:42:56.425899] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:09:08.344 [2024-11-26 23:42:56.425980] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:09:08.344 [2024-11-26 23:42:56.426000] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:09:08.344 pt2 00:09:08.344 23:42:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:08.344 23:42:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@473 -- # rpc_cmd bdev_passthru_delete pt2 00:09:08.344 23:42:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:08.344 23:42:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:08.344 [2024-11-26 23:42:56.437270] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt2 00:09:08.344 23:42:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:08.344 23:42:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@474 -- # verify_raid_bdev_state raid_bdev1 configuring raid0 64 4 00:09:08.344 23:42:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:08.344 23:42:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:08.344 23:42:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:08.344 23:42:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:08.344 23:42:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:08.344 23:42:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:08.344 23:42:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:08.344 23:42:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:08.344 23:42:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:08.344 23:42:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:08.344 23:42:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:08.344 23:42:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:08.344 23:42:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:08.344 23:42:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:08.602 23:42:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:08.602 "name": "raid_bdev1", 00:09:08.602 "uuid": "d151ea8c-de3f-4bab-8dbf-26742b8f64ab", 00:09:08.602 "strip_size_kb": 64, 00:09:08.602 "state": "configuring", 00:09:08.602 "raid_level": "raid0", 00:09:08.602 "superblock": true, 00:09:08.602 "num_base_bdevs": 4, 00:09:08.602 "num_base_bdevs_discovered": 1, 00:09:08.602 "num_base_bdevs_operational": 4, 00:09:08.602 "base_bdevs_list": [ 00:09:08.602 { 00:09:08.602 "name": "pt1", 00:09:08.602 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:08.602 "is_configured": true, 00:09:08.602 "data_offset": 2048, 00:09:08.602 "data_size": 63488 00:09:08.602 }, 00:09:08.602 { 00:09:08.602 "name": null, 00:09:08.602 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:08.602 "is_configured": false, 00:09:08.602 "data_offset": 0, 00:09:08.602 "data_size": 63488 00:09:08.602 }, 00:09:08.602 { 00:09:08.602 "name": null, 00:09:08.602 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:08.602 "is_configured": false, 00:09:08.602 "data_offset": 2048, 00:09:08.602 "data_size": 63488 00:09:08.602 }, 00:09:08.602 { 00:09:08.602 "name": null, 00:09:08.602 "uuid": "00000000-0000-0000-0000-000000000004", 00:09:08.602 "is_configured": false, 00:09:08.602 "data_offset": 2048, 00:09:08.602 "data_size": 63488 00:09:08.602 } 00:09:08.602 ] 00:09:08.602 }' 00:09:08.602 23:42:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:08.602 23:42:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:08.862 23:42:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:09:08.862 23:42:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:09:08.862 23:42:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:09:08.862 23:42:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:08.862 23:42:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:08.862 [2024-11-26 23:42:56.888495] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:09:08.862 [2024-11-26 23:42:56.888623] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:08.862 [2024-11-26 23:42:56.888657] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009f80 00:09:08.862 [2024-11-26 23:42:56.888686] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:08.862 [2024-11-26 23:42:56.889091] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:08.862 [2024-11-26 23:42:56.889155] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:09:08.863 [2024-11-26 23:42:56.889254] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:09:08.863 [2024-11-26 23:42:56.889311] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:09:08.863 pt2 00:09:08.863 23:42:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:08.863 23:42:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:09:08.863 23:42:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:09:08.863 23:42:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:09:08.863 23:42:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:08.863 23:42:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:08.863 [2024-11-26 23:42:56.900459] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:09:08.863 [2024-11-26 23:42:56.900543] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:08.863 [2024-11-26 23:42:56.900574] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:09:08.863 [2024-11-26 23:42:56.900601] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:08.863 [2024-11-26 23:42:56.900948] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:08.864 [2024-11-26 23:42:56.901009] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:09:08.864 [2024-11-26 23:42:56.901088] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:09:08.864 [2024-11-26 23:42:56.901134] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:09:08.864 pt3 00:09:08.864 23:42:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:08.864 23:42:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:09:08.864 23:42:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:09:08.864 23:42:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:09:08.864 23:42:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:08.864 23:42:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:08.864 [2024-11-26 23:42:56.912437] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:09:08.864 [2024-11-26 23:42:56.912485] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:08.865 [2024-11-26 23:42:56.912499] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a580 00:09:08.865 [2024-11-26 23:42:56.912508] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:08.865 [2024-11-26 23:42:56.912787] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:08.865 [2024-11-26 23:42:56.912805] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:09:08.865 [2024-11-26 23:42:56.912853] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt4 00:09:08.865 [2024-11-26 23:42:56.912871] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:09:08.865 [2024-11-26 23:42:56.912963] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:09:08.865 [2024-11-26 23:42:56.912974] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:09:08.865 [2024-11-26 23:42:56.913188] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:09:08.865 [2024-11-26 23:42:56.913300] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:09:08.865 [2024-11-26 23:42:56.913307] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:09:08.865 [2024-11-26 23:42:56.913417] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:08.865 pt4 00:09:08.865 23:42:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:08.865 23:42:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:09:08.865 23:42:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:09:08.865 23:42:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 4 00:09:08.866 23:42:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:08.866 23:42:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:08.866 23:42:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:08.866 23:42:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:08.866 23:42:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:08.866 23:42:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:08.866 23:42:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:08.866 23:42:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:08.866 23:42:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:08.866 23:42:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:08.866 23:42:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:08.866 23:42:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:08.866 23:42:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:08.866 23:42:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:08.866 23:42:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:08.866 "name": "raid_bdev1", 00:09:08.866 "uuid": "d151ea8c-de3f-4bab-8dbf-26742b8f64ab", 00:09:08.867 "strip_size_kb": 64, 00:09:08.867 "state": "online", 00:09:08.867 "raid_level": "raid0", 00:09:08.867 "superblock": true, 00:09:08.867 "num_base_bdevs": 4, 00:09:08.867 "num_base_bdevs_discovered": 4, 00:09:08.867 "num_base_bdevs_operational": 4, 00:09:08.867 "base_bdevs_list": [ 00:09:08.867 { 00:09:08.867 "name": "pt1", 00:09:08.867 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:08.867 "is_configured": true, 00:09:08.867 "data_offset": 2048, 00:09:08.867 "data_size": 63488 00:09:08.867 }, 00:09:08.867 { 00:09:08.867 "name": "pt2", 00:09:08.867 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:08.868 "is_configured": true, 00:09:08.868 "data_offset": 2048, 00:09:08.868 "data_size": 63488 00:09:08.868 }, 00:09:08.868 { 00:09:08.868 "name": "pt3", 00:09:08.868 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:08.868 "is_configured": true, 00:09:08.868 "data_offset": 2048, 00:09:08.868 "data_size": 63488 00:09:08.868 }, 00:09:08.868 { 00:09:08.868 "name": "pt4", 00:09:08.868 "uuid": "00000000-0000-0000-0000-000000000004", 00:09:08.868 "is_configured": true, 00:09:08.868 "data_offset": 2048, 00:09:08.868 "data_size": 63488 00:09:08.868 } 00:09:08.868 ] 00:09:08.868 }' 00:09:08.868 23:42:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:08.868 23:42:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:09.445 23:42:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:09:09.445 23:42:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:09:09.445 23:42:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:09.445 23:42:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:09.445 23:42:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:09:09.445 23:42:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:09.445 23:42:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:09:09.445 23:42:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:09.445 23:42:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:09.445 23:42:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:09.445 [2024-11-26 23:42:57.340005] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:09.445 23:42:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:09.445 23:42:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:09.445 "name": "raid_bdev1", 00:09:09.445 "aliases": [ 00:09:09.445 "d151ea8c-de3f-4bab-8dbf-26742b8f64ab" 00:09:09.445 ], 00:09:09.445 "product_name": "Raid Volume", 00:09:09.445 "block_size": 512, 00:09:09.445 "num_blocks": 253952, 00:09:09.445 "uuid": "d151ea8c-de3f-4bab-8dbf-26742b8f64ab", 00:09:09.445 "assigned_rate_limits": { 00:09:09.445 "rw_ios_per_sec": 0, 00:09:09.445 "rw_mbytes_per_sec": 0, 00:09:09.445 "r_mbytes_per_sec": 0, 00:09:09.445 "w_mbytes_per_sec": 0 00:09:09.445 }, 00:09:09.445 "claimed": false, 00:09:09.445 "zoned": false, 00:09:09.445 "supported_io_types": { 00:09:09.445 "read": true, 00:09:09.445 "write": true, 00:09:09.445 "unmap": true, 00:09:09.445 "flush": true, 00:09:09.445 "reset": true, 00:09:09.445 "nvme_admin": false, 00:09:09.445 "nvme_io": false, 00:09:09.445 "nvme_io_md": false, 00:09:09.445 "write_zeroes": true, 00:09:09.445 "zcopy": false, 00:09:09.445 "get_zone_info": false, 00:09:09.445 "zone_management": false, 00:09:09.445 "zone_append": false, 00:09:09.445 "compare": false, 00:09:09.445 "compare_and_write": false, 00:09:09.445 "abort": false, 00:09:09.445 "seek_hole": false, 00:09:09.445 "seek_data": false, 00:09:09.445 "copy": false, 00:09:09.445 "nvme_iov_md": false 00:09:09.445 }, 00:09:09.445 "memory_domains": [ 00:09:09.445 { 00:09:09.445 "dma_device_id": "system", 00:09:09.445 "dma_device_type": 1 00:09:09.445 }, 00:09:09.445 { 00:09:09.445 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:09.445 "dma_device_type": 2 00:09:09.445 }, 00:09:09.445 { 00:09:09.445 "dma_device_id": "system", 00:09:09.445 "dma_device_type": 1 00:09:09.445 }, 00:09:09.445 { 00:09:09.445 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:09.445 "dma_device_type": 2 00:09:09.445 }, 00:09:09.445 { 00:09:09.445 "dma_device_id": "system", 00:09:09.445 "dma_device_type": 1 00:09:09.445 }, 00:09:09.445 { 00:09:09.445 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:09.445 "dma_device_type": 2 00:09:09.445 }, 00:09:09.445 { 00:09:09.445 "dma_device_id": "system", 00:09:09.445 "dma_device_type": 1 00:09:09.445 }, 00:09:09.445 { 00:09:09.445 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:09.445 "dma_device_type": 2 00:09:09.445 } 00:09:09.445 ], 00:09:09.445 "driver_specific": { 00:09:09.445 "raid": { 00:09:09.445 "uuid": "d151ea8c-de3f-4bab-8dbf-26742b8f64ab", 00:09:09.445 "strip_size_kb": 64, 00:09:09.445 "state": "online", 00:09:09.445 "raid_level": "raid0", 00:09:09.445 "superblock": true, 00:09:09.445 "num_base_bdevs": 4, 00:09:09.445 "num_base_bdevs_discovered": 4, 00:09:09.445 "num_base_bdevs_operational": 4, 00:09:09.446 "base_bdevs_list": [ 00:09:09.446 { 00:09:09.446 "name": "pt1", 00:09:09.446 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:09.446 "is_configured": true, 00:09:09.446 "data_offset": 2048, 00:09:09.446 "data_size": 63488 00:09:09.446 }, 00:09:09.446 { 00:09:09.446 "name": "pt2", 00:09:09.446 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:09.446 "is_configured": true, 00:09:09.446 "data_offset": 2048, 00:09:09.446 "data_size": 63488 00:09:09.446 }, 00:09:09.446 { 00:09:09.446 "name": "pt3", 00:09:09.446 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:09.446 "is_configured": true, 00:09:09.446 "data_offset": 2048, 00:09:09.446 "data_size": 63488 00:09:09.446 }, 00:09:09.446 { 00:09:09.446 "name": "pt4", 00:09:09.446 "uuid": "00000000-0000-0000-0000-000000000004", 00:09:09.446 "is_configured": true, 00:09:09.446 "data_offset": 2048, 00:09:09.446 "data_size": 63488 00:09:09.446 } 00:09:09.446 ] 00:09:09.446 } 00:09:09.446 } 00:09:09.446 }' 00:09:09.446 23:42:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:09.446 23:42:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:09:09.446 pt2 00:09:09.446 pt3 00:09:09.446 pt4' 00:09:09.446 23:42:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:09.446 23:42:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:09.446 23:42:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:09.446 23:42:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:09.446 23:42:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:09:09.446 23:42:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:09.446 23:42:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:09.446 23:42:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:09.446 23:42:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:09.446 23:42:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:09.446 23:42:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:09.446 23:42:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:09:09.446 23:42:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:09.446 23:42:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:09.446 23:42:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:09.446 23:42:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:09.446 23:42:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:09.446 23:42:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:09.446 23:42:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:09.446 23:42:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:09:09.446 23:42:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:09.446 23:42:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:09.446 23:42:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:09.705 23:42:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:09.705 23:42:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:09.705 23:42:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:09.705 23:42:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:09.705 23:42:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:09.705 23:42:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt4 00:09:09.705 23:42:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:09.705 23:42:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:09.705 23:42:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:09.705 23:42:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:09.705 23:42:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:09.705 23:42:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:09:09.705 23:42:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:09:09.705 23:42:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:09.705 23:42:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:09.705 [2024-11-26 23:42:57.651504] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:09.705 23:42:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:09.705 23:42:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' d151ea8c-de3f-4bab-8dbf-26742b8f64ab '!=' d151ea8c-de3f-4bab-8dbf-26742b8f64ab ']' 00:09:09.705 23:42:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy raid0 00:09:09.705 23:42:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:09:09.705 23:42:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # return 1 00:09:09.705 23:42:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 81371 00:09:09.705 23:42:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # '[' -z 81371 ']' 00:09:09.705 23:42:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@958 -- # kill -0 81371 00:09:09.705 23:42:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # uname 00:09:09.705 23:42:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:09:09.705 23:42:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 81371 00:09:09.705 23:42:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:09:09.705 23:42:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:09:09.705 23:42:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 81371' 00:09:09.705 killing process with pid 81371 00:09:09.705 23:42:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@973 -- # kill 81371 00:09:09.705 [2024-11-26 23:42:57.750073] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:09:09.705 [2024-11-26 23:42:57.750190] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:09.705 [2024-11-26 23:42:57.750268] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:09.705 [2024-11-26 23:42:57.750282] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:09:09.705 23:42:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@978 -- # wait 81371 00:09:09.705 [2024-11-26 23:42:57.793978] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:09:09.963 23:42:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:09:09.963 00:09:09.963 real 0m4.175s 00:09:09.963 user 0m6.644s 00:09:09.963 sys 0m0.883s 00:09:09.963 23:42:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:09:09.963 ************************************ 00:09:09.963 END TEST raid_superblock_test 00:09:09.963 ************************************ 00:09:09.963 23:42:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:09.963 23:42:58 bdev_raid -- bdev/bdev_raid.sh@971 -- # run_test raid_read_error_test raid_io_error_test raid0 4 read 00:09:09.963 23:42:58 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:09:09.963 23:42:58 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:09:09.963 23:42:58 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:09:09.963 ************************************ 00:09:09.963 START TEST raid_read_error_test 00:09:09.963 ************************************ 00:09:09.963 23:42:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test raid0 4 read 00:09:09.963 23:42:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid0 00:09:09.963 23:42:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=4 00:09:09.963 23:42:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=read 00:09:09.963 23:42:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:09:09.963 23:42:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:09.963 23:42:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:09:09.963 23:42:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:09.963 23:42:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:09.963 23:42:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:09:09.963 23:42:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:09.963 23:42:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:09.963 23:42:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:09:09.963 23:42:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:09.963 23:42:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:09.963 23:42:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev4 00:09:09.963 23:42:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:09.963 23:42:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:09.963 23:42:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:09:09.963 23:42:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:09:09.963 23:42:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:09:09.963 23:42:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:09:09.963 23:42:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:09:09.963 23:42:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:09:09.963 23:42:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:09:09.963 23:42:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid0 '!=' raid1 ']' 00:09:09.963 23:42:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:09:09.963 23:42:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:09:09.963 23:42:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:09:09.963 23:42:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.cVvkE6uKqz 00:09:09.963 23:42:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=81619 00:09:09.964 23:42:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:09:09.964 23:42:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 81619 00:09:09.964 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:09.964 23:42:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@835 -- # '[' -z 81619 ']' 00:09:09.964 23:42:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:09.964 23:42:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:09:09.964 23:42:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:09.964 23:42:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:09:09.964 23:42:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:10.222 [2024-11-26 23:42:58.162278] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:09:10.222 [2024-11-26 23:42:58.162433] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid81619 ] 00:09:10.222 [2024-11-26 23:42:58.314080] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:10.222 [2024-11-26 23:42:58.339909] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:09:10.482 [2024-11-26 23:42:58.383240] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:10.482 [2024-11-26 23:42:58.383273] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:11.050 23:42:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:09:11.050 23:42:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@868 -- # return 0 00:09:11.050 23:42:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:11.050 23:42:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:09:11.050 23:42:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:11.050 23:42:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:11.050 BaseBdev1_malloc 00:09:11.050 23:42:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:11.050 23:42:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:09:11.050 23:42:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:11.050 23:42:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:11.050 true 00:09:11.050 23:42:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:11.050 23:42:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:09:11.050 23:42:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:11.050 23:42:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:11.050 [2024-11-26 23:42:59.014879] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:09:11.050 [2024-11-26 23:42:59.014972] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:11.050 [2024-11-26 23:42:59.015001] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006980 00:09:11.050 [2024-11-26 23:42:59.015012] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:11.050 [2024-11-26 23:42:59.017137] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:11.050 [2024-11-26 23:42:59.017172] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:09:11.050 BaseBdev1 00:09:11.051 23:42:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:11.051 23:42:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:11.051 23:42:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:09:11.051 23:42:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:11.051 23:42:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:11.051 BaseBdev2_malloc 00:09:11.051 23:42:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:11.051 23:42:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:09:11.051 23:42:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:11.051 23:42:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:11.051 true 00:09:11.051 23:42:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:11.051 23:42:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:09:11.051 23:42:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:11.051 23:42:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:11.051 [2024-11-26 23:42:59.055583] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:09:11.051 [2024-11-26 23:42:59.055704] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:11.051 [2024-11-26 23:42:59.055740] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:09:11.051 [2024-11-26 23:42:59.055780] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:11.051 [2024-11-26 23:42:59.057970] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:11.051 [2024-11-26 23:42:59.058043] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:09:11.051 BaseBdev2 00:09:11.051 23:42:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:11.051 23:42:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:11.051 23:42:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:09:11.051 23:42:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:11.051 23:42:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:11.051 BaseBdev3_malloc 00:09:11.051 23:42:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:11.051 23:42:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:09:11.051 23:42:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:11.051 23:42:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:11.051 true 00:09:11.051 23:42:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:11.051 23:42:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:09:11.051 23:42:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:11.051 23:42:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:11.051 [2024-11-26 23:42:59.096203] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:09:11.051 [2024-11-26 23:42:59.096304] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:11.051 [2024-11-26 23:42:59.096340] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008780 00:09:11.051 [2024-11-26 23:42:59.096381] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:11.051 [2024-11-26 23:42:59.098456] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:11.051 [2024-11-26 23:42:59.098526] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:09:11.051 BaseBdev3 00:09:11.051 23:42:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:11.051 23:42:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:11.051 23:42:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:09:11.051 23:42:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:11.051 23:42:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:11.051 BaseBdev4_malloc 00:09:11.051 23:42:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:11.051 23:42:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev4_malloc 00:09:11.051 23:42:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:11.051 23:42:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:11.051 true 00:09:11.051 23:42:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:11.051 23:42:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev4_malloc -p BaseBdev4 00:09:11.051 23:42:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:11.051 23:42:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:11.051 [2024-11-26 23:42:59.148079] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev4_malloc 00:09:11.051 [2024-11-26 23:42:59.148125] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:11.051 [2024-11-26 23:42:59.148146] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:09:11.051 [2024-11-26 23:42:59.148154] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:11.051 [2024-11-26 23:42:59.150224] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:11.051 [2024-11-26 23:42:59.150258] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:09:11.051 BaseBdev4 00:09:11.051 23:42:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:11.051 23:42:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 -s 00:09:11.051 23:42:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:11.051 23:42:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:11.051 [2024-11-26 23:42:59.160114] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:11.051 [2024-11-26 23:42:59.161945] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:11.051 [2024-11-26 23:42:59.162023] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:11.051 [2024-11-26 23:42:59.162074] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:09:11.051 [2024-11-26 23:42:59.162276] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000002000 00:09:11.051 [2024-11-26 23:42:59.162288] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:09:11.051 [2024-11-26 23:42:59.162553] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002ef0 00:09:11.051 [2024-11-26 23:42:59.162690] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000002000 00:09:11.051 [2024-11-26 23:42:59.162703] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000002000 00:09:11.051 [2024-11-26 23:42:59.162830] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:11.051 23:42:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:11.051 23:42:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 4 00:09:11.051 23:42:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:11.051 23:42:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:11.051 23:42:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:11.051 23:42:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:11.051 23:42:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:11.051 23:42:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:11.051 23:42:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:11.051 23:42:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:11.051 23:42:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:11.051 23:42:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:11.051 23:42:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:11.051 23:42:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:11.051 23:42:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:11.309 23:42:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:11.309 23:42:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:11.309 "name": "raid_bdev1", 00:09:11.309 "uuid": "5b352be3-634d-4c61-9503-587dcf048852", 00:09:11.309 "strip_size_kb": 64, 00:09:11.309 "state": "online", 00:09:11.309 "raid_level": "raid0", 00:09:11.309 "superblock": true, 00:09:11.309 "num_base_bdevs": 4, 00:09:11.309 "num_base_bdevs_discovered": 4, 00:09:11.309 "num_base_bdevs_operational": 4, 00:09:11.309 "base_bdevs_list": [ 00:09:11.309 { 00:09:11.309 "name": "BaseBdev1", 00:09:11.309 "uuid": "4e3123ba-063d-5600-b569-2f20bbea73c2", 00:09:11.309 "is_configured": true, 00:09:11.309 "data_offset": 2048, 00:09:11.309 "data_size": 63488 00:09:11.309 }, 00:09:11.309 { 00:09:11.309 "name": "BaseBdev2", 00:09:11.309 "uuid": "7413f8d8-b319-5d7f-aff8-b47960c47b30", 00:09:11.309 "is_configured": true, 00:09:11.309 "data_offset": 2048, 00:09:11.309 "data_size": 63488 00:09:11.309 }, 00:09:11.309 { 00:09:11.309 "name": "BaseBdev3", 00:09:11.309 "uuid": "e722638f-4843-59f1-b0cc-944d87c0d5b9", 00:09:11.309 "is_configured": true, 00:09:11.309 "data_offset": 2048, 00:09:11.309 "data_size": 63488 00:09:11.309 }, 00:09:11.309 { 00:09:11.309 "name": "BaseBdev4", 00:09:11.309 "uuid": "6c185b26-c3a0-507f-95c1-a0fa62109e88", 00:09:11.309 "is_configured": true, 00:09:11.309 "data_offset": 2048, 00:09:11.309 "data_size": 63488 00:09:11.309 } 00:09:11.309 ] 00:09:11.309 }' 00:09:11.309 23:42:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:11.309 23:42:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:11.567 23:42:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:09:11.567 23:42:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:09:11.567 [2024-11-26 23:42:59.651686] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000003090 00:09:12.512 23:43:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:09:12.513 23:43:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:12.513 23:43:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:12.513 23:43:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:12.513 23:43:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:09:12.513 23:43:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid0 = \r\a\i\d\1 ]] 00:09:12.513 23:43:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=4 00:09:12.513 23:43:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 4 00:09:12.513 23:43:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:12.513 23:43:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:12.513 23:43:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:12.513 23:43:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:12.513 23:43:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:12.513 23:43:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:12.513 23:43:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:12.513 23:43:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:12.513 23:43:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:12.513 23:43:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:12.513 23:43:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:12.513 23:43:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:12.513 23:43:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:12.513 23:43:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:12.513 23:43:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:12.513 "name": "raid_bdev1", 00:09:12.513 "uuid": "5b352be3-634d-4c61-9503-587dcf048852", 00:09:12.513 "strip_size_kb": 64, 00:09:12.513 "state": "online", 00:09:12.514 "raid_level": "raid0", 00:09:12.514 "superblock": true, 00:09:12.514 "num_base_bdevs": 4, 00:09:12.514 "num_base_bdevs_discovered": 4, 00:09:12.514 "num_base_bdevs_operational": 4, 00:09:12.514 "base_bdevs_list": [ 00:09:12.514 { 00:09:12.514 "name": "BaseBdev1", 00:09:12.514 "uuid": "4e3123ba-063d-5600-b569-2f20bbea73c2", 00:09:12.514 "is_configured": true, 00:09:12.514 "data_offset": 2048, 00:09:12.514 "data_size": 63488 00:09:12.514 }, 00:09:12.514 { 00:09:12.514 "name": "BaseBdev2", 00:09:12.514 "uuid": "7413f8d8-b319-5d7f-aff8-b47960c47b30", 00:09:12.514 "is_configured": true, 00:09:12.514 "data_offset": 2048, 00:09:12.514 "data_size": 63488 00:09:12.514 }, 00:09:12.514 { 00:09:12.514 "name": "BaseBdev3", 00:09:12.514 "uuid": "e722638f-4843-59f1-b0cc-944d87c0d5b9", 00:09:12.514 "is_configured": true, 00:09:12.514 "data_offset": 2048, 00:09:12.514 "data_size": 63488 00:09:12.514 }, 00:09:12.514 { 00:09:12.514 "name": "BaseBdev4", 00:09:12.514 "uuid": "6c185b26-c3a0-507f-95c1-a0fa62109e88", 00:09:12.514 "is_configured": true, 00:09:12.514 "data_offset": 2048, 00:09:12.514 "data_size": 63488 00:09:12.514 } 00:09:12.514 ] 00:09:12.514 }' 00:09:12.514 23:43:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:12.514 23:43:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:13.089 23:43:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:09:13.089 23:43:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:13.089 23:43:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:13.089 [2024-11-26 23:43:01.048121] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:13.089 [2024-11-26 23:43:01.048156] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:13.089 [2024-11-26 23:43:01.050883] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:13.089 [2024-11-26 23:43:01.051047] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:13.089 [2024-11-26 23:43:01.051105] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:13.089 [2024-11-26 23:43:01.051126] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002000 name raid_bdev1, state offline 00:09:13.089 { 00:09:13.089 "results": [ 00:09:13.089 { 00:09:13.089 "job": "raid_bdev1", 00:09:13.089 "core_mask": "0x1", 00:09:13.089 "workload": "randrw", 00:09:13.089 "percentage": 50, 00:09:13.089 "status": "finished", 00:09:13.089 "queue_depth": 1, 00:09:13.089 "io_size": 131072, 00:09:13.089 "runtime": 1.397311, 00:09:13.089 "iops": 16415.100145923134, 00:09:13.089 "mibps": 2051.8875182403917, 00:09:13.089 "io_failed": 1, 00:09:13.089 "io_timeout": 0, 00:09:13.089 "avg_latency_us": 84.11596675450551, 00:09:13.089 "min_latency_us": 24.929257641921396, 00:09:13.089 "max_latency_us": 1366.5257641921398 00:09:13.089 } 00:09:13.089 ], 00:09:13.089 "core_count": 1 00:09:13.089 } 00:09:13.089 23:43:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:13.089 23:43:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 81619 00:09:13.089 23:43:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # '[' -z 81619 ']' 00:09:13.089 23:43:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@958 -- # kill -0 81619 00:09:13.089 23:43:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # uname 00:09:13.089 23:43:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:09:13.089 23:43:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 81619 00:09:13.089 23:43:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:09:13.089 23:43:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:09:13.089 killing process with pid 81619 00:09:13.089 23:43:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 81619' 00:09:13.089 23:43:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@973 -- # kill 81619 00:09:13.089 [2024-11-26 23:43:01.097152] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:09:13.089 23:43:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@978 -- # wait 81619 00:09:13.089 [2024-11-26 23:43:01.132932] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:09:13.348 23:43:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.cVvkE6uKqz 00:09:13.348 23:43:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:09:13.348 23:43:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:09:13.348 23:43:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.72 00:09:13.348 23:43:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid0 00:09:13.348 23:43:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:09:13.348 23:43:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:09:13.349 23:43:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.72 != \0\.\0\0 ]] 00:09:13.349 00:09:13.349 real 0m3.279s 00:09:13.349 user 0m4.166s 00:09:13.349 sys 0m0.500s 00:09:13.349 23:43:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:09:13.349 ************************************ 00:09:13.349 END TEST raid_read_error_test 00:09:13.349 ************************************ 00:09:13.349 23:43:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:13.349 23:43:01 bdev_raid -- bdev/bdev_raid.sh@972 -- # run_test raid_write_error_test raid_io_error_test raid0 4 write 00:09:13.349 23:43:01 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:09:13.349 23:43:01 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:09:13.349 23:43:01 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:09:13.349 ************************************ 00:09:13.349 START TEST raid_write_error_test 00:09:13.349 ************************************ 00:09:13.349 23:43:01 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test raid0 4 write 00:09:13.349 23:43:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid0 00:09:13.349 23:43:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=4 00:09:13.349 23:43:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=write 00:09:13.349 23:43:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:09:13.349 23:43:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:13.349 23:43:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:09:13.349 23:43:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:13.349 23:43:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:13.349 23:43:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:09:13.349 23:43:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:13.349 23:43:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:13.349 23:43:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:09:13.349 23:43:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:13.349 23:43:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:13.349 23:43:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev4 00:09:13.349 23:43:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:13.349 23:43:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:13.349 23:43:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:09:13.349 23:43:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:09:13.349 23:43:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:09:13.349 23:43:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:09:13.349 23:43:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:09:13.349 23:43:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:09:13.349 23:43:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:09:13.349 23:43:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid0 '!=' raid1 ']' 00:09:13.349 23:43:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:09:13.349 23:43:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:09:13.349 23:43:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:09:13.349 23:43:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.sce2Kxff5l 00:09:13.349 23:43:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=81755 00:09:13.349 23:43:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:09:13.349 23:43:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 81755 00:09:13.349 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:13.349 23:43:01 bdev_raid.raid_write_error_test -- common/autotest_common.sh@835 -- # '[' -z 81755 ']' 00:09:13.349 23:43:01 bdev_raid.raid_write_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:13.349 23:43:01 bdev_raid.raid_write_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:09:13.349 23:43:01 bdev_raid.raid_write_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:13.349 23:43:01 bdev_raid.raid_write_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:09:13.349 23:43:01 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:13.608 [2024-11-26 23:43:01.518121] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:09:13.608 [2024-11-26 23:43:01.518269] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid81755 ] 00:09:13.608 [2024-11-26 23:43:01.671803] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:13.608 [2024-11-26 23:43:01.697280] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:09:13.867 [2024-11-26 23:43:01.739805] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:13.867 [2024-11-26 23:43:01.739839] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:14.434 23:43:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:09:14.434 23:43:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@868 -- # return 0 00:09:14.434 23:43:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:14.434 23:43:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:09:14.434 23:43:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:14.434 23:43:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:14.434 BaseBdev1_malloc 00:09:14.434 23:43:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:14.434 23:43:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:09:14.434 23:43:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:14.434 23:43:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:14.434 true 00:09:14.434 23:43:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:14.434 23:43:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:09:14.434 23:43:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:14.434 23:43:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:14.434 [2024-11-26 23:43:02.363217] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:09:14.434 [2024-11-26 23:43:02.363367] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:14.434 [2024-11-26 23:43:02.363392] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006980 00:09:14.434 [2024-11-26 23:43:02.363401] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:14.434 [2024-11-26 23:43:02.365474] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:14.434 [2024-11-26 23:43:02.365510] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:09:14.434 BaseBdev1 00:09:14.435 23:43:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:14.435 23:43:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:14.435 23:43:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:09:14.435 23:43:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:14.435 23:43:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:14.435 BaseBdev2_malloc 00:09:14.435 23:43:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:14.435 23:43:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:09:14.435 23:43:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:14.435 23:43:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:14.435 true 00:09:14.435 23:43:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:14.435 23:43:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:09:14.435 23:43:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:14.435 23:43:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:14.435 [2024-11-26 23:43:02.399723] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:09:14.435 [2024-11-26 23:43:02.399771] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:14.435 [2024-11-26 23:43:02.399789] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:09:14.435 [2024-11-26 23:43:02.399804] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:14.435 [2024-11-26 23:43:02.401902] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:14.435 [2024-11-26 23:43:02.401951] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:09:14.435 BaseBdev2 00:09:14.435 23:43:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:14.435 23:43:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:14.435 23:43:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:09:14.435 23:43:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:14.435 23:43:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:14.435 BaseBdev3_malloc 00:09:14.435 23:43:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:14.435 23:43:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:09:14.435 23:43:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:14.435 23:43:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:14.435 true 00:09:14.435 23:43:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:14.435 23:43:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:09:14.435 23:43:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:14.435 23:43:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:14.435 [2024-11-26 23:43:02.440299] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:09:14.435 [2024-11-26 23:43:02.440366] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:14.435 [2024-11-26 23:43:02.440401] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008780 00:09:14.435 [2024-11-26 23:43:02.440409] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:14.435 [2024-11-26 23:43:02.442423] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:14.435 [2024-11-26 23:43:02.442456] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:09:14.435 BaseBdev3 00:09:14.435 23:43:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:14.435 23:43:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:14.435 23:43:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:09:14.435 23:43:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:14.435 23:43:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:14.435 BaseBdev4_malloc 00:09:14.435 23:43:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:14.435 23:43:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev4_malloc 00:09:14.435 23:43:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:14.435 23:43:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:14.435 true 00:09:14.435 23:43:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:14.435 23:43:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev4_malloc -p BaseBdev4 00:09:14.435 23:43:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:14.435 23:43:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:14.435 [2024-11-26 23:43:02.489943] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev4_malloc 00:09:14.435 [2024-11-26 23:43:02.489991] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:14.435 [2024-11-26 23:43:02.490011] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:09:14.435 [2024-11-26 23:43:02.490019] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:14.435 [2024-11-26 23:43:02.492015] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:14.435 [2024-11-26 23:43:02.492100] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:09:14.435 BaseBdev4 00:09:14.435 23:43:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:14.435 23:43:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 -s 00:09:14.435 23:43:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:14.435 23:43:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:14.435 [2024-11-26 23:43:02.501973] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:14.435 [2024-11-26 23:43:02.503775] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:14.435 [2024-11-26 23:43:02.503850] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:14.435 [2024-11-26 23:43:02.503900] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:09:14.435 [2024-11-26 23:43:02.504085] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000002000 00:09:14.435 [2024-11-26 23:43:02.504096] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:09:14.435 [2024-11-26 23:43:02.504347] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002ef0 00:09:14.435 [2024-11-26 23:43:02.504498] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000002000 00:09:14.435 [2024-11-26 23:43:02.504511] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000002000 00:09:14.435 [2024-11-26 23:43:02.504628] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:14.435 23:43:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:14.435 23:43:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 4 00:09:14.435 23:43:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:14.435 23:43:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:14.435 23:43:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:14.435 23:43:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:14.435 23:43:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:14.435 23:43:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:14.435 23:43:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:14.435 23:43:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:14.435 23:43:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:14.435 23:43:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:14.435 23:43:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:14.435 23:43:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:14.435 23:43:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:14.435 23:43:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:14.435 23:43:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:14.435 "name": "raid_bdev1", 00:09:14.435 "uuid": "daef0d53-0b80-4b61-9835-a2bfec123185", 00:09:14.435 "strip_size_kb": 64, 00:09:14.435 "state": "online", 00:09:14.435 "raid_level": "raid0", 00:09:14.435 "superblock": true, 00:09:14.435 "num_base_bdevs": 4, 00:09:14.435 "num_base_bdevs_discovered": 4, 00:09:14.435 "num_base_bdevs_operational": 4, 00:09:14.435 "base_bdevs_list": [ 00:09:14.435 { 00:09:14.435 "name": "BaseBdev1", 00:09:14.435 "uuid": "58ad296e-85e7-5702-800e-c61a28529c66", 00:09:14.435 "is_configured": true, 00:09:14.435 "data_offset": 2048, 00:09:14.435 "data_size": 63488 00:09:14.435 }, 00:09:14.435 { 00:09:14.435 "name": "BaseBdev2", 00:09:14.435 "uuid": "3bdd5ee0-f4a0-5ab5-94f1-25ee4ef044dd", 00:09:14.435 "is_configured": true, 00:09:14.435 "data_offset": 2048, 00:09:14.435 "data_size": 63488 00:09:14.435 }, 00:09:14.435 { 00:09:14.435 "name": "BaseBdev3", 00:09:14.435 "uuid": "1eed4097-e557-5b94-b74a-b246c49603c5", 00:09:14.435 "is_configured": true, 00:09:14.435 "data_offset": 2048, 00:09:14.435 "data_size": 63488 00:09:14.435 }, 00:09:14.435 { 00:09:14.435 "name": "BaseBdev4", 00:09:14.435 "uuid": "b1cfd161-0736-52a9-a8d6-88df7839b245", 00:09:14.435 "is_configured": true, 00:09:14.435 "data_offset": 2048, 00:09:14.435 "data_size": 63488 00:09:14.435 } 00:09:14.435 ] 00:09:14.435 }' 00:09:14.435 23:43:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:14.435 23:43:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:15.001 23:43:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:09:15.001 23:43:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:09:15.001 [2024-11-26 23:43:03.045421] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000003090 00:09:15.936 23:43:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:09:15.936 23:43:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:15.936 23:43:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:15.936 23:43:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:15.937 23:43:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:09:15.937 23:43:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid0 = \r\a\i\d\1 ]] 00:09:15.937 23:43:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=4 00:09:15.937 23:43:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 4 00:09:15.937 23:43:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:15.937 23:43:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:15.937 23:43:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:15.937 23:43:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:15.937 23:43:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:15.937 23:43:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:15.937 23:43:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:15.937 23:43:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:15.937 23:43:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:15.937 23:43:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:15.937 23:43:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:15.937 23:43:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:15.937 23:43:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:15.937 23:43:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:15.937 23:43:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:15.937 "name": "raid_bdev1", 00:09:15.937 "uuid": "daef0d53-0b80-4b61-9835-a2bfec123185", 00:09:15.937 "strip_size_kb": 64, 00:09:15.937 "state": "online", 00:09:15.937 "raid_level": "raid0", 00:09:15.937 "superblock": true, 00:09:15.937 "num_base_bdevs": 4, 00:09:15.937 "num_base_bdevs_discovered": 4, 00:09:15.937 "num_base_bdevs_operational": 4, 00:09:15.937 "base_bdevs_list": [ 00:09:15.937 { 00:09:15.937 "name": "BaseBdev1", 00:09:15.937 "uuid": "58ad296e-85e7-5702-800e-c61a28529c66", 00:09:15.937 "is_configured": true, 00:09:15.937 "data_offset": 2048, 00:09:15.937 "data_size": 63488 00:09:15.937 }, 00:09:15.937 { 00:09:15.937 "name": "BaseBdev2", 00:09:15.937 "uuid": "3bdd5ee0-f4a0-5ab5-94f1-25ee4ef044dd", 00:09:15.937 "is_configured": true, 00:09:15.937 "data_offset": 2048, 00:09:15.937 "data_size": 63488 00:09:15.937 }, 00:09:15.937 { 00:09:15.937 "name": "BaseBdev3", 00:09:15.937 "uuid": "1eed4097-e557-5b94-b74a-b246c49603c5", 00:09:15.937 "is_configured": true, 00:09:15.937 "data_offset": 2048, 00:09:15.937 "data_size": 63488 00:09:15.937 }, 00:09:15.937 { 00:09:15.937 "name": "BaseBdev4", 00:09:15.937 "uuid": "b1cfd161-0736-52a9-a8d6-88df7839b245", 00:09:15.937 "is_configured": true, 00:09:15.937 "data_offset": 2048, 00:09:15.937 "data_size": 63488 00:09:15.937 } 00:09:15.937 ] 00:09:15.937 }' 00:09:15.937 23:43:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:15.937 23:43:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:16.505 23:43:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:09:16.505 23:43:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:16.505 23:43:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:16.505 [2024-11-26 23:43:04.478055] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:16.505 [2024-11-26 23:43:04.478136] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:16.505 [2024-11-26 23:43:04.480792] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:16.505 [2024-11-26 23:43:04.480900] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:16.505 [2024-11-26 23:43:04.480968] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:16.505 [2024-11-26 23:43:04.481011] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002000 name raid_bdev1, state offline 00:09:16.505 { 00:09:16.505 "results": [ 00:09:16.505 { 00:09:16.505 "job": "raid_bdev1", 00:09:16.505 "core_mask": "0x1", 00:09:16.505 "workload": "randrw", 00:09:16.505 "percentage": 50, 00:09:16.505 "status": "finished", 00:09:16.505 "queue_depth": 1, 00:09:16.505 "io_size": 131072, 00:09:16.505 "runtime": 1.433602, 00:09:16.505 "iops": 16059.54790799678, 00:09:16.505 "mibps": 2007.4434884995976, 00:09:16.505 "io_failed": 1, 00:09:16.505 "io_timeout": 0, 00:09:16.505 "avg_latency_us": 85.949352261244, 00:09:16.505 "min_latency_us": 26.494323144104804, 00:09:16.505 "max_latency_us": 1359.3711790393013 00:09:16.505 } 00:09:16.505 ], 00:09:16.505 "core_count": 1 00:09:16.505 } 00:09:16.505 23:43:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:16.505 23:43:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 81755 00:09:16.505 23:43:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # '[' -z 81755 ']' 00:09:16.505 23:43:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@958 -- # kill -0 81755 00:09:16.505 23:43:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # uname 00:09:16.505 23:43:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:09:16.505 23:43:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 81755 00:09:16.505 23:43:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:09:16.505 23:43:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:09:16.505 23:43:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 81755' 00:09:16.505 killing process with pid 81755 00:09:16.505 23:43:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@973 -- # kill 81755 00:09:16.505 [2024-11-26 23:43:04.525053] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:09:16.505 23:43:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@978 -- # wait 81755 00:09:16.505 [2024-11-26 23:43:04.559694] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:09:16.765 23:43:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:09:16.765 23:43:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.sce2Kxff5l 00:09:16.765 23:43:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:09:16.765 23:43:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.70 00:09:16.765 23:43:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid0 00:09:16.765 ************************************ 00:09:16.765 END TEST raid_write_error_test 00:09:16.765 ************************************ 00:09:16.765 23:43:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:09:16.765 23:43:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:09:16.765 23:43:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.70 != \0\.\0\0 ]] 00:09:16.765 00:09:16.765 real 0m3.357s 00:09:16.765 user 0m4.285s 00:09:16.765 sys 0m0.530s 00:09:16.765 23:43:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:09:16.765 23:43:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:16.765 23:43:04 bdev_raid -- bdev/bdev_raid.sh@967 -- # for level in raid0 concat raid1 00:09:16.765 23:43:04 bdev_raid -- bdev/bdev_raid.sh@968 -- # run_test raid_state_function_test raid_state_function_test concat 4 false 00:09:16.765 23:43:04 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:09:16.765 23:43:04 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:09:16.765 23:43:04 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:09:16.765 ************************************ 00:09:16.765 START TEST raid_state_function_test 00:09:16.765 ************************************ 00:09:16.765 23:43:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1129 -- # raid_state_function_test concat 4 false 00:09:16.765 23:43:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=concat 00:09:16.765 23:43:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=4 00:09:16.765 23:43:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:09:16.765 23:43:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:09:16.765 23:43:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:09:16.765 23:43:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:16.765 23:43:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:09:16.765 23:43:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:16.765 23:43:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:16.765 23:43:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:09:16.765 23:43:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:16.765 23:43:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:16.765 23:43:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:09:16.765 23:43:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:16.765 23:43:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:16.765 23:43:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev4 00:09:16.765 23:43:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:16.765 23:43:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:16.765 23:43:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:09:16.765 23:43:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:09:16.765 23:43:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:09:16.765 23:43:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:09:16.765 23:43:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:09:16.765 23:43:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:09:16.765 23:43:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' concat '!=' raid1 ']' 00:09:16.765 23:43:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:09:16.765 23:43:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:09:16.765 23:43:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:09:16.765 23:43:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:09:16.765 23:43:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=81882 00:09:16.765 23:43:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:09:16.765 23:43:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 81882' 00:09:16.765 Process raid pid: 81882 00:09:16.766 23:43:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 81882 00:09:16.766 23:43:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@835 -- # '[' -z 81882 ']' 00:09:16.766 23:43:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:16.766 23:43:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:09:16.766 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:16.766 23:43:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:16.766 23:43:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:09:16.766 23:43:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:17.030 [2024-11-26 23:43:04.934207] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:09:17.030 [2024-11-26 23:43:04.934325] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:09:17.030 [2024-11-26 23:43:05.089951] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:17.030 [2024-11-26 23:43:05.114716] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:09:17.297 [2024-11-26 23:43:05.157804] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:17.297 [2024-11-26 23:43:05.157839] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:17.863 23:43:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:09:17.863 23:43:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@868 -- # return 0 00:09:17.863 23:43:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:09:17.863 23:43:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:17.863 23:43:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:17.863 [2024-11-26 23:43:05.744646] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:17.863 [2024-11-26 23:43:05.744704] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:17.863 [2024-11-26 23:43:05.744714] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:17.863 [2024-11-26 23:43:05.744723] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:17.863 [2024-11-26 23:43:05.744729] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:17.863 [2024-11-26 23:43:05.744739] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:17.863 [2024-11-26 23:43:05.744745] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:09:17.863 [2024-11-26 23:43:05.744753] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:09:17.863 23:43:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:17.863 23:43:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:09:17.863 23:43:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:17.863 23:43:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:17.863 23:43:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:17.863 23:43:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:17.863 23:43:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:17.863 23:43:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:17.863 23:43:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:17.863 23:43:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:17.863 23:43:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:17.863 23:43:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:17.863 23:43:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:17.863 23:43:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:17.863 23:43:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:17.863 23:43:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:17.863 23:43:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:17.863 "name": "Existed_Raid", 00:09:17.863 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:17.863 "strip_size_kb": 64, 00:09:17.863 "state": "configuring", 00:09:17.863 "raid_level": "concat", 00:09:17.863 "superblock": false, 00:09:17.863 "num_base_bdevs": 4, 00:09:17.863 "num_base_bdevs_discovered": 0, 00:09:17.863 "num_base_bdevs_operational": 4, 00:09:17.863 "base_bdevs_list": [ 00:09:17.863 { 00:09:17.863 "name": "BaseBdev1", 00:09:17.863 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:17.863 "is_configured": false, 00:09:17.863 "data_offset": 0, 00:09:17.863 "data_size": 0 00:09:17.863 }, 00:09:17.863 { 00:09:17.863 "name": "BaseBdev2", 00:09:17.863 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:17.863 "is_configured": false, 00:09:17.863 "data_offset": 0, 00:09:17.863 "data_size": 0 00:09:17.863 }, 00:09:17.863 { 00:09:17.863 "name": "BaseBdev3", 00:09:17.863 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:17.863 "is_configured": false, 00:09:17.863 "data_offset": 0, 00:09:17.863 "data_size": 0 00:09:17.863 }, 00:09:17.863 { 00:09:17.863 "name": "BaseBdev4", 00:09:17.863 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:17.863 "is_configured": false, 00:09:17.863 "data_offset": 0, 00:09:17.863 "data_size": 0 00:09:17.863 } 00:09:17.863 ] 00:09:17.863 }' 00:09:17.863 23:43:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:17.863 23:43:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:18.120 23:43:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:18.120 23:43:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:18.120 23:43:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:18.120 [2024-11-26 23:43:06.191826] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:18.120 [2024-11-26 23:43:06.192025] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:09:18.120 23:43:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:18.120 23:43:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:09:18.120 23:43:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:18.120 23:43:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:18.120 [2024-11-26 23:43:06.199831] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:18.120 [2024-11-26 23:43:06.199912] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:18.120 [2024-11-26 23:43:06.199940] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:18.120 [2024-11-26 23:43:06.199962] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:18.120 [2024-11-26 23:43:06.199979] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:18.120 [2024-11-26 23:43:06.200000] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:18.120 [2024-11-26 23:43:06.200017] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:09:18.120 [2024-11-26 23:43:06.200036] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:09:18.120 23:43:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:18.120 23:43:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:09:18.120 23:43:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:18.120 23:43:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:18.120 [2024-11-26 23:43:06.216617] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:18.120 BaseBdev1 00:09:18.120 23:43:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:18.120 23:43:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:09:18.120 23:43:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:09:18.120 23:43:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:18.120 23:43:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:09:18.120 23:43:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:18.120 23:43:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:18.120 23:43:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:18.120 23:43:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:18.120 23:43:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:18.120 23:43:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:18.120 23:43:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:09:18.120 23:43:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:18.120 23:43:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:18.120 [ 00:09:18.120 { 00:09:18.120 "name": "BaseBdev1", 00:09:18.120 "aliases": [ 00:09:18.120 "a4bcc1db-4c57-4977-8350-ee9ef0848f8d" 00:09:18.120 ], 00:09:18.120 "product_name": "Malloc disk", 00:09:18.120 "block_size": 512, 00:09:18.120 "num_blocks": 65536, 00:09:18.120 "uuid": "a4bcc1db-4c57-4977-8350-ee9ef0848f8d", 00:09:18.120 "assigned_rate_limits": { 00:09:18.120 "rw_ios_per_sec": 0, 00:09:18.120 "rw_mbytes_per_sec": 0, 00:09:18.120 "r_mbytes_per_sec": 0, 00:09:18.120 "w_mbytes_per_sec": 0 00:09:18.120 }, 00:09:18.120 "claimed": true, 00:09:18.120 "claim_type": "exclusive_write", 00:09:18.120 "zoned": false, 00:09:18.120 "supported_io_types": { 00:09:18.120 "read": true, 00:09:18.120 "write": true, 00:09:18.120 "unmap": true, 00:09:18.120 "flush": true, 00:09:18.120 "reset": true, 00:09:18.120 "nvme_admin": false, 00:09:18.120 "nvme_io": false, 00:09:18.120 "nvme_io_md": false, 00:09:18.120 "write_zeroes": true, 00:09:18.120 "zcopy": true, 00:09:18.120 "get_zone_info": false, 00:09:18.120 "zone_management": false, 00:09:18.120 "zone_append": false, 00:09:18.378 "compare": false, 00:09:18.378 "compare_and_write": false, 00:09:18.378 "abort": true, 00:09:18.378 "seek_hole": false, 00:09:18.378 "seek_data": false, 00:09:18.378 "copy": true, 00:09:18.378 "nvme_iov_md": false 00:09:18.378 }, 00:09:18.378 "memory_domains": [ 00:09:18.378 { 00:09:18.378 "dma_device_id": "system", 00:09:18.378 "dma_device_type": 1 00:09:18.378 }, 00:09:18.378 { 00:09:18.378 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:18.378 "dma_device_type": 2 00:09:18.378 } 00:09:18.378 ], 00:09:18.378 "driver_specific": {} 00:09:18.378 } 00:09:18.378 ] 00:09:18.378 23:43:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:18.378 23:43:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:09:18.378 23:43:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:09:18.378 23:43:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:18.378 23:43:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:18.378 23:43:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:18.378 23:43:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:18.378 23:43:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:18.378 23:43:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:18.378 23:43:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:18.378 23:43:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:18.378 23:43:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:18.378 23:43:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:18.378 23:43:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:18.378 23:43:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:18.378 23:43:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:18.378 23:43:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:18.378 23:43:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:18.378 "name": "Existed_Raid", 00:09:18.378 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:18.378 "strip_size_kb": 64, 00:09:18.378 "state": "configuring", 00:09:18.378 "raid_level": "concat", 00:09:18.378 "superblock": false, 00:09:18.378 "num_base_bdevs": 4, 00:09:18.378 "num_base_bdevs_discovered": 1, 00:09:18.378 "num_base_bdevs_operational": 4, 00:09:18.378 "base_bdevs_list": [ 00:09:18.378 { 00:09:18.378 "name": "BaseBdev1", 00:09:18.378 "uuid": "a4bcc1db-4c57-4977-8350-ee9ef0848f8d", 00:09:18.378 "is_configured": true, 00:09:18.378 "data_offset": 0, 00:09:18.378 "data_size": 65536 00:09:18.378 }, 00:09:18.378 { 00:09:18.378 "name": "BaseBdev2", 00:09:18.378 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:18.378 "is_configured": false, 00:09:18.378 "data_offset": 0, 00:09:18.378 "data_size": 0 00:09:18.378 }, 00:09:18.378 { 00:09:18.378 "name": "BaseBdev3", 00:09:18.378 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:18.378 "is_configured": false, 00:09:18.378 "data_offset": 0, 00:09:18.378 "data_size": 0 00:09:18.378 }, 00:09:18.378 { 00:09:18.378 "name": "BaseBdev4", 00:09:18.378 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:18.378 "is_configured": false, 00:09:18.378 "data_offset": 0, 00:09:18.378 "data_size": 0 00:09:18.378 } 00:09:18.378 ] 00:09:18.378 }' 00:09:18.378 23:43:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:18.378 23:43:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:18.637 23:43:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:18.637 23:43:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:18.637 23:43:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:18.637 [2024-11-26 23:43:06.687869] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:18.637 [2024-11-26 23:43:06.687976] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:09:18.637 23:43:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:18.637 23:43:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:09:18.637 23:43:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:18.637 23:43:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:18.637 [2024-11-26 23:43:06.699886] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:18.637 [2024-11-26 23:43:06.701765] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:18.637 [2024-11-26 23:43:06.701844] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:18.637 [2024-11-26 23:43:06.701871] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:18.637 [2024-11-26 23:43:06.701893] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:18.637 [2024-11-26 23:43:06.701911] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:09:18.637 [2024-11-26 23:43:06.701931] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:09:18.637 23:43:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:18.637 23:43:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:09:18.637 23:43:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:18.637 23:43:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:09:18.637 23:43:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:18.637 23:43:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:18.637 23:43:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:18.637 23:43:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:18.637 23:43:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:18.637 23:43:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:18.637 23:43:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:18.637 23:43:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:18.637 23:43:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:18.637 23:43:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:18.637 23:43:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:18.637 23:43:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:18.637 23:43:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:18.637 23:43:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:18.637 23:43:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:18.637 "name": "Existed_Raid", 00:09:18.637 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:18.637 "strip_size_kb": 64, 00:09:18.637 "state": "configuring", 00:09:18.637 "raid_level": "concat", 00:09:18.637 "superblock": false, 00:09:18.637 "num_base_bdevs": 4, 00:09:18.637 "num_base_bdevs_discovered": 1, 00:09:18.637 "num_base_bdevs_operational": 4, 00:09:18.637 "base_bdevs_list": [ 00:09:18.637 { 00:09:18.637 "name": "BaseBdev1", 00:09:18.637 "uuid": "a4bcc1db-4c57-4977-8350-ee9ef0848f8d", 00:09:18.637 "is_configured": true, 00:09:18.637 "data_offset": 0, 00:09:18.637 "data_size": 65536 00:09:18.637 }, 00:09:18.637 { 00:09:18.637 "name": "BaseBdev2", 00:09:18.637 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:18.637 "is_configured": false, 00:09:18.637 "data_offset": 0, 00:09:18.637 "data_size": 0 00:09:18.637 }, 00:09:18.637 { 00:09:18.637 "name": "BaseBdev3", 00:09:18.637 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:18.637 "is_configured": false, 00:09:18.637 "data_offset": 0, 00:09:18.637 "data_size": 0 00:09:18.637 }, 00:09:18.637 { 00:09:18.637 "name": "BaseBdev4", 00:09:18.637 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:18.637 "is_configured": false, 00:09:18.637 "data_offset": 0, 00:09:18.637 "data_size": 0 00:09:18.637 } 00:09:18.637 ] 00:09:18.637 }' 00:09:18.637 23:43:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:18.637 23:43:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:19.204 23:43:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:09:19.204 23:43:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:19.204 23:43:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:19.204 [2024-11-26 23:43:07.162121] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:19.204 BaseBdev2 00:09:19.204 23:43:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:19.204 23:43:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:09:19.204 23:43:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:09:19.204 23:43:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:19.204 23:43:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:09:19.204 23:43:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:19.204 23:43:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:19.204 23:43:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:19.204 23:43:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:19.204 23:43:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:19.204 23:43:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:19.204 23:43:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:09:19.204 23:43:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:19.204 23:43:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:19.204 [ 00:09:19.204 { 00:09:19.204 "name": "BaseBdev2", 00:09:19.204 "aliases": [ 00:09:19.204 "73b126d2-0b3a-4c84-ad92-00e5863adc8d" 00:09:19.204 ], 00:09:19.204 "product_name": "Malloc disk", 00:09:19.204 "block_size": 512, 00:09:19.204 "num_blocks": 65536, 00:09:19.204 "uuid": "73b126d2-0b3a-4c84-ad92-00e5863adc8d", 00:09:19.204 "assigned_rate_limits": { 00:09:19.204 "rw_ios_per_sec": 0, 00:09:19.204 "rw_mbytes_per_sec": 0, 00:09:19.204 "r_mbytes_per_sec": 0, 00:09:19.204 "w_mbytes_per_sec": 0 00:09:19.204 }, 00:09:19.204 "claimed": true, 00:09:19.204 "claim_type": "exclusive_write", 00:09:19.204 "zoned": false, 00:09:19.204 "supported_io_types": { 00:09:19.204 "read": true, 00:09:19.204 "write": true, 00:09:19.204 "unmap": true, 00:09:19.204 "flush": true, 00:09:19.204 "reset": true, 00:09:19.204 "nvme_admin": false, 00:09:19.204 "nvme_io": false, 00:09:19.204 "nvme_io_md": false, 00:09:19.204 "write_zeroes": true, 00:09:19.204 "zcopy": true, 00:09:19.204 "get_zone_info": false, 00:09:19.204 "zone_management": false, 00:09:19.204 "zone_append": false, 00:09:19.204 "compare": false, 00:09:19.204 "compare_and_write": false, 00:09:19.204 "abort": true, 00:09:19.204 "seek_hole": false, 00:09:19.204 "seek_data": false, 00:09:19.204 "copy": true, 00:09:19.204 "nvme_iov_md": false 00:09:19.204 }, 00:09:19.204 "memory_domains": [ 00:09:19.204 { 00:09:19.204 "dma_device_id": "system", 00:09:19.204 "dma_device_type": 1 00:09:19.204 }, 00:09:19.204 { 00:09:19.204 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:19.204 "dma_device_type": 2 00:09:19.204 } 00:09:19.204 ], 00:09:19.204 "driver_specific": {} 00:09:19.204 } 00:09:19.204 ] 00:09:19.204 23:43:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:19.204 23:43:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:09:19.204 23:43:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:09:19.204 23:43:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:19.204 23:43:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:09:19.204 23:43:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:19.204 23:43:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:19.204 23:43:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:19.204 23:43:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:19.204 23:43:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:19.204 23:43:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:19.204 23:43:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:19.204 23:43:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:19.204 23:43:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:19.204 23:43:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:19.204 23:43:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:19.204 23:43:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:19.204 23:43:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:19.204 23:43:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:19.204 23:43:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:19.204 "name": "Existed_Raid", 00:09:19.204 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:19.204 "strip_size_kb": 64, 00:09:19.204 "state": "configuring", 00:09:19.204 "raid_level": "concat", 00:09:19.204 "superblock": false, 00:09:19.204 "num_base_bdevs": 4, 00:09:19.204 "num_base_bdevs_discovered": 2, 00:09:19.204 "num_base_bdevs_operational": 4, 00:09:19.204 "base_bdevs_list": [ 00:09:19.204 { 00:09:19.204 "name": "BaseBdev1", 00:09:19.204 "uuid": "a4bcc1db-4c57-4977-8350-ee9ef0848f8d", 00:09:19.204 "is_configured": true, 00:09:19.204 "data_offset": 0, 00:09:19.204 "data_size": 65536 00:09:19.204 }, 00:09:19.204 { 00:09:19.204 "name": "BaseBdev2", 00:09:19.204 "uuid": "73b126d2-0b3a-4c84-ad92-00e5863adc8d", 00:09:19.204 "is_configured": true, 00:09:19.204 "data_offset": 0, 00:09:19.204 "data_size": 65536 00:09:19.204 }, 00:09:19.204 { 00:09:19.204 "name": "BaseBdev3", 00:09:19.205 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:19.205 "is_configured": false, 00:09:19.205 "data_offset": 0, 00:09:19.205 "data_size": 0 00:09:19.205 }, 00:09:19.205 { 00:09:19.205 "name": "BaseBdev4", 00:09:19.205 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:19.205 "is_configured": false, 00:09:19.205 "data_offset": 0, 00:09:19.205 "data_size": 0 00:09:19.205 } 00:09:19.205 ] 00:09:19.205 }' 00:09:19.205 23:43:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:19.205 23:43:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:19.770 23:43:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:09:19.770 23:43:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:19.770 23:43:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:19.770 [2024-11-26 23:43:07.659012] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:19.770 BaseBdev3 00:09:19.770 23:43:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:19.770 23:43:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:09:19.770 23:43:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:09:19.770 23:43:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:19.770 23:43:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:09:19.770 23:43:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:19.770 23:43:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:19.770 23:43:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:19.770 23:43:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:19.770 23:43:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:19.770 23:43:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:19.770 23:43:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:09:19.770 23:43:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:19.770 23:43:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:19.770 [ 00:09:19.770 { 00:09:19.770 "name": "BaseBdev3", 00:09:19.770 "aliases": [ 00:09:19.770 "c527e570-ff29-41ef-b225-06652aa021d3" 00:09:19.770 ], 00:09:19.770 "product_name": "Malloc disk", 00:09:19.770 "block_size": 512, 00:09:19.770 "num_blocks": 65536, 00:09:19.770 "uuid": "c527e570-ff29-41ef-b225-06652aa021d3", 00:09:19.770 "assigned_rate_limits": { 00:09:19.770 "rw_ios_per_sec": 0, 00:09:19.770 "rw_mbytes_per_sec": 0, 00:09:19.770 "r_mbytes_per_sec": 0, 00:09:19.770 "w_mbytes_per_sec": 0 00:09:19.770 }, 00:09:19.770 "claimed": true, 00:09:19.770 "claim_type": "exclusive_write", 00:09:19.770 "zoned": false, 00:09:19.770 "supported_io_types": { 00:09:19.770 "read": true, 00:09:19.770 "write": true, 00:09:19.770 "unmap": true, 00:09:19.770 "flush": true, 00:09:19.770 "reset": true, 00:09:19.770 "nvme_admin": false, 00:09:19.770 "nvme_io": false, 00:09:19.770 "nvme_io_md": false, 00:09:19.770 "write_zeroes": true, 00:09:19.770 "zcopy": true, 00:09:19.770 "get_zone_info": false, 00:09:19.770 "zone_management": false, 00:09:19.770 "zone_append": false, 00:09:19.770 "compare": false, 00:09:19.770 "compare_and_write": false, 00:09:19.770 "abort": true, 00:09:19.770 "seek_hole": false, 00:09:19.770 "seek_data": false, 00:09:19.770 "copy": true, 00:09:19.770 "nvme_iov_md": false 00:09:19.770 }, 00:09:19.770 "memory_domains": [ 00:09:19.770 { 00:09:19.770 "dma_device_id": "system", 00:09:19.770 "dma_device_type": 1 00:09:19.770 }, 00:09:19.770 { 00:09:19.770 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:19.770 "dma_device_type": 2 00:09:19.770 } 00:09:19.770 ], 00:09:19.770 "driver_specific": {} 00:09:19.770 } 00:09:19.770 ] 00:09:19.770 23:43:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:19.770 23:43:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:09:19.770 23:43:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:09:19.770 23:43:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:19.770 23:43:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:09:19.770 23:43:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:19.771 23:43:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:19.771 23:43:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:19.771 23:43:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:19.771 23:43:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:19.771 23:43:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:19.771 23:43:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:19.771 23:43:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:19.771 23:43:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:19.771 23:43:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:19.771 23:43:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:19.771 23:43:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:19.771 23:43:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:19.771 23:43:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:19.771 23:43:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:19.771 "name": "Existed_Raid", 00:09:19.771 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:19.771 "strip_size_kb": 64, 00:09:19.771 "state": "configuring", 00:09:19.771 "raid_level": "concat", 00:09:19.771 "superblock": false, 00:09:19.771 "num_base_bdevs": 4, 00:09:19.771 "num_base_bdevs_discovered": 3, 00:09:19.771 "num_base_bdevs_operational": 4, 00:09:19.771 "base_bdevs_list": [ 00:09:19.771 { 00:09:19.771 "name": "BaseBdev1", 00:09:19.771 "uuid": "a4bcc1db-4c57-4977-8350-ee9ef0848f8d", 00:09:19.771 "is_configured": true, 00:09:19.771 "data_offset": 0, 00:09:19.771 "data_size": 65536 00:09:19.771 }, 00:09:19.771 { 00:09:19.771 "name": "BaseBdev2", 00:09:19.771 "uuid": "73b126d2-0b3a-4c84-ad92-00e5863adc8d", 00:09:19.771 "is_configured": true, 00:09:19.771 "data_offset": 0, 00:09:19.771 "data_size": 65536 00:09:19.771 }, 00:09:19.771 { 00:09:19.771 "name": "BaseBdev3", 00:09:19.771 "uuid": "c527e570-ff29-41ef-b225-06652aa021d3", 00:09:19.771 "is_configured": true, 00:09:19.771 "data_offset": 0, 00:09:19.771 "data_size": 65536 00:09:19.771 }, 00:09:19.771 { 00:09:19.771 "name": "BaseBdev4", 00:09:19.771 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:19.771 "is_configured": false, 00:09:19.771 "data_offset": 0, 00:09:19.771 "data_size": 0 00:09:19.771 } 00:09:19.771 ] 00:09:19.771 }' 00:09:19.771 23:43:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:19.771 23:43:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:20.028 23:43:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:09:20.028 23:43:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:20.028 23:43:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:20.028 [2024-11-26 23:43:08.113235] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:09:20.028 [2024-11-26 23:43:08.113358] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:09:20.028 [2024-11-26 23:43:08.113379] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 262144, blocklen 512 00:09:20.028 [2024-11-26 23:43:08.113693] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:09:20.028 [2024-11-26 23:43:08.113826] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:09:20.028 [2024-11-26 23:43:08.113838] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:09:20.028 [2024-11-26 23:43:08.114025] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:20.028 BaseBdev4 00:09:20.028 23:43:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:20.028 23:43:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev4 00:09:20.028 23:43:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:09:20.029 23:43:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:20.029 23:43:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:09:20.029 23:43:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:20.029 23:43:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:20.029 23:43:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:20.029 23:43:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:20.029 23:43:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:20.029 23:43:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:20.029 23:43:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:09:20.029 23:43:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:20.029 23:43:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:20.029 [ 00:09:20.029 { 00:09:20.029 "name": "BaseBdev4", 00:09:20.029 "aliases": [ 00:09:20.029 "878b19a4-d105-462d-80da-f190df600f1d" 00:09:20.029 ], 00:09:20.029 "product_name": "Malloc disk", 00:09:20.029 "block_size": 512, 00:09:20.029 "num_blocks": 65536, 00:09:20.029 "uuid": "878b19a4-d105-462d-80da-f190df600f1d", 00:09:20.029 "assigned_rate_limits": { 00:09:20.029 "rw_ios_per_sec": 0, 00:09:20.029 "rw_mbytes_per_sec": 0, 00:09:20.029 "r_mbytes_per_sec": 0, 00:09:20.029 "w_mbytes_per_sec": 0 00:09:20.029 }, 00:09:20.029 "claimed": true, 00:09:20.029 "claim_type": "exclusive_write", 00:09:20.029 "zoned": false, 00:09:20.029 "supported_io_types": { 00:09:20.029 "read": true, 00:09:20.029 "write": true, 00:09:20.029 "unmap": true, 00:09:20.029 "flush": true, 00:09:20.029 "reset": true, 00:09:20.029 "nvme_admin": false, 00:09:20.029 "nvme_io": false, 00:09:20.029 "nvme_io_md": false, 00:09:20.029 "write_zeroes": true, 00:09:20.029 "zcopy": true, 00:09:20.029 "get_zone_info": false, 00:09:20.029 "zone_management": false, 00:09:20.029 "zone_append": false, 00:09:20.029 "compare": false, 00:09:20.029 "compare_and_write": false, 00:09:20.029 "abort": true, 00:09:20.029 "seek_hole": false, 00:09:20.029 "seek_data": false, 00:09:20.029 "copy": true, 00:09:20.029 "nvme_iov_md": false 00:09:20.029 }, 00:09:20.029 "memory_domains": [ 00:09:20.029 { 00:09:20.029 "dma_device_id": "system", 00:09:20.029 "dma_device_type": 1 00:09:20.029 }, 00:09:20.029 { 00:09:20.029 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:20.029 "dma_device_type": 2 00:09:20.029 } 00:09:20.029 ], 00:09:20.029 "driver_specific": {} 00:09:20.029 } 00:09:20.029 ] 00:09:20.029 23:43:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:20.029 23:43:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:09:20.029 23:43:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:09:20.029 23:43:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:20.029 23:43:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online concat 64 4 00:09:20.029 23:43:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:20.029 23:43:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:20.029 23:43:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:20.029 23:43:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:20.029 23:43:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:20.029 23:43:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:20.029 23:43:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:20.029 23:43:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:20.029 23:43:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:20.029 23:43:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:20.029 23:43:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:20.029 23:43:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:20.029 23:43:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:20.287 23:43:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:20.287 23:43:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:20.287 "name": "Existed_Raid", 00:09:20.287 "uuid": "366ab0ca-c39a-4554-8385-287fc736eaaf", 00:09:20.287 "strip_size_kb": 64, 00:09:20.287 "state": "online", 00:09:20.287 "raid_level": "concat", 00:09:20.287 "superblock": false, 00:09:20.287 "num_base_bdevs": 4, 00:09:20.287 "num_base_bdevs_discovered": 4, 00:09:20.287 "num_base_bdevs_operational": 4, 00:09:20.287 "base_bdevs_list": [ 00:09:20.287 { 00:09:20.287 "name": "BaseBdev1", 00:09:20.287 "uuid": "a4bcc1db-4c57-4977-8350-ee9ef0848f8d", 00:09:20.287 "is_configured": true, 00:09:20.287 "data_offset": 0, 00:09:20.287 "data_size": 65536 00:09:20.287 }, 00:09:20.287 { 00:09:20.287 "name": "BaseBdev2", 00:09:20.287 "uuid": "73b126d2-0b3a-4c84-ad92-00e5863adc8d", 00:09:20.287 "is_configured": true, 00:09:20.287 "data_offset": 0, 00:09:20.287 "data_size": 65536 00:09:20.287 }, 00:09:20.287 { 00:09:20.287 "name": "BaseBdev3", 00:09:20.287 "uuid": "c527e570-ff29-41ef-b225-06652aa021d3", 00:09:20.287 "is_configured": true, 00:09:20.287 "data_offset": 0, 00:09:20.287 "data_size": 65536 00:09:20.287 }, 00:09:20.287 { 00:09:20.287 "name": "BaseBdev4", 00:09:20.287 "uuid": "878b19a4-d105-462d-80da-f190df600f1d", 00:09:20.287 "is_configured": true, 00:09:20.287 "data_offset": 0, 00:09:20.287 "data_size": 65536 00:09:20.287 } 00:09:20.287 ] 00:09:20.287 }' 00:09:20.287 23:43:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:20.287 23:43:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:20.546 23:43:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:09:20.546 23:43:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:09:20.546 23:43:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:20.546 23:43:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:20.546 23:43:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:09:20.546 23:43:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:20.546 23:43:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:20.546 23:43:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:09:20.546 23:43:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:20.546 23:43:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:20.546 [2024-11-26 23:43:08.576892] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:20.546 23:43:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:20.546 23:43:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:20.546 "name": "Existed_Raid", 00:09:20.546 "aliases": [ 00:09:20.546 "366ab0ca-c39a-4554-8385-287fc736eaaf" 00:09:20.546 ], 00:09:20.546 "product_name": "Raid Volume", 00:09:20.546 "block_size": 512, 00:09:20.546 "num_blocks": 262144, 00:09:20.546 "uuid": "366ab0ca-c39a-4554-8385-287fc736eaaf", 00:09:20.546 "assigned_rate_limits": { 00:09:20.546 "rw_ios_per_sec": 0, 00:09:20.546 "rw_mbytes_per_sec": 0, 00:09:20.546 "r_mbytes_per_sec": 0, 00:09:20.546 "w_mbytes_per_sec": 0 00:09:20.546 }, 00:09:20.546 "claimed": false, 00:09:20.546 "zoned": false, 00:09:20.546 "supported_io_types": { 00:09:20.546 "read": true, 00:09:20.546 "write": true, 00:09:20.546 "unmap": true, 00:09:20.546 "flush": true, 00:09:20.546 "reset": true, 00:09:20.546 "nvme_admin": false, 00:09:20.546 "nvme_io": false, 00:09:20.546 "nvme_io_md": false, 00:09:20.546 "write_zeroes": true, 00:09:20.546 "zcopy": false, 00:09:20.546 "get_zone_info": false, 00:09:20.546 "zone_management": false, 00:09:20.546 "zone_append": false, 00:09:20.546 "compare": false, 00:09:20.546 "compare_and_write": false, 00:09:20.546 "abort": false, 00:09:20.546 "seek_hole": false, 00:09:20.546 "seek_data": false, 00:09:20.546 "copy": false, 00:09:20.546 "nvme_iov_md": false 00:09:20.546 }, 00:09:20.546 "memory_domains": [ 00:09:20.546 { 00:09:20.546 "dma_device_id": "system", 00:09:20.546 "dma_device_type": 1 00:09:20.546 }, 00:09:20.546 { 00:09:20.546 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:20.546 "dma_device_type": 2 00:09:20.546 }, 00:09:20.546 { 00:09:20.546 "dma_device_id": "system", 00:09:20.546 "dma_device_type": 1 00:09:20.546 }, 00:09:20.546 { 00:09:20.546 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:20.546 "dma_device_type": 2 00:09:20.546 }, 00:09:20.546 { 00:09:20.546 "dma_device_id": "system", 00:09:20.546 "dma_device_type": 1 00:09:20.546 }, 00:09:20.546 { 00:09:20.546 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:20.546 "dma_device_type": 2 00:09:20.546 }, 00:09:20.546 { 00:09:20.546 "dma_device_id": "system", 00:09:20.546 "dma_device_type": 1 00:09:20.546 }, 00:09:20.546 { 00:09:20.546 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:20.546 "dma_device_type": 2 00:09:20.546 } 00:09:20.546 ], 00:09:20.546 "driver_specific": { 00:09:20.546 "raid": { 00:09:20.546 "uuid": "366ab0ca-c39a-4554-8385-287fc736eaaf", 00:09:20.546 "strip_size_kb": 64, 00:09:20.546 "state": "online", 00:09:20.546 "raid_level": "concat", 00:09:20.546 "superblock": false, 00:09:20.546 "num_base_bdevs": 4, 00:09:20.546 "num_base_bdevs_discovered": 4, 00:09:20.546 "num_base_bdevs_operational": 4, 00:09:20.546 "base_bdevs_list": [ 00:09:20.546 { 00:09:20.546 "name": "BaseBdev1", 00:09:20.546 "uuid": "a4bcc1db-4c57-4977-8350-ee9ef0848f8d", 00:09:20.546 "is_configured": true, 00:09:20.546 "data_offset": 0, 00:09:20.546 "data_size": 65536 00:09:20.546 }, 00:09:20.546 { 00:09:20.546 "name": "BaseBdev2", 00:09:20.546 "uuid": "73b126d2-0b3a-4c84-ad92-00e5863adc8d", 00:09:20.546 "is_configured": true, 00:09:20.546 "data_offset": 0, 00:09:20.546 "data_size": 65536 00:09:20.546 }, 00:09:20.546 { 00:09:20.546 "name": "BaseBdev3", 00:09:20.546 "uuid": "c527e570-ff29-41ef-b225-06652aa021d3", 00:09:20.546 "is_configured": true, 00:09:20.546 "data_offset": 0, 00:09:20.546 "data_size": 65536 00:09:20.546 }, 00:09:20.546 { 00:09:20.546 "name": "BaseBdev4", 00:09:20.546 "uuid": "878b19a4-d105-462d-80da-f190df600f1d", 00:09:20.546 "is_configured": true, 00:09:20.546 "data_offset": 0, 00:09:20.546 "data_size": 65536 00:09:20.546 } 00:09:20.547 ] 00:09:20.547 } 00:09:20.547 } 00:09:20.547 }' 00:09:20.547 23:43:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:20.547 23:43:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:09:20.547 BaseBdev2 00:09:20.547 BaseBdev3 00:09:20.547 BaseBdev4' 00:09:20.547 23:43:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:20.547 23:43:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:20.547 23:43:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:20.547 23:43:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:09:20.547 23:43:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:20.547 23:43:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:20.547 23:43:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:20.806 23:43:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:20.806 23:43:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:20.806 23:43:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:20.806 23:43:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:20.806 23:43:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:09:20.806 23:43:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:20.806 23:43:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:20.806 23:43:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:20.806 23:43:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:20.806 23:43:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:20.806 23:43:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:20.806 23:43:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:20.806 23:43:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:20.806 23:43:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:09:20.806 23:43:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:20.806 23:43:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:20.806 23:43:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:20.806 23:43:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:20.806 23:43:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:20.806 23:43:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:20.806 23:43:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:09:20.806 23:43:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:20.806 23:43:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:20.806 23:43:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:20.806 23:43:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:20.806 23:43:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:20.806 23:43:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:20.806 23:43:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:09:20.806 23:43:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:20.806 23:43:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:20.806 [2024-11-26 23:43:08.852109] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:09:20.806 [2024-11-26 23:43:08.852137] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:20.806 [2024-11-26 23:43:08.852191] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:20.806 23:43:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:20.806 23:43:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:09:20.806 23:43:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy concat 00:09:20.806 23:43:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:09:20.806 23:43:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # return 1 00:09:20.806 23:43:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:09:20.806 23:43:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline concat 64 3 00:09:20.806 23:43:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:20.806 23:43:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:09:20.806 23:43:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:20.806 23:43:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:20.806 23:43:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:20.806 23:43:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:20.806 23:43:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:20.806 23:43:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:20.806 23:43:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:20.806 23:43:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:20.806 23:43:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:20.806 23:43:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:20.806 23:43:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:20.806 23:43:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:20.806 23:43:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:20.806 "name": "Existed_Raid", 00:09:20.806 "uuid": "366ab0ca-c39a-4554-8385-287fc736eaaf", 00:09:20.806 "strip_size_kb": 64, 00:09:20.806 "state": "offline", 00:09:20.806 "raid_level": "concat", 00:09:20.806 "superblock": false, 00:09:20.806 "num_base_bdevs": 4, 00:09:20.806 "num_base_bdevs_discovered": 3, 00:09:20.806 "num_base_bdevs_operational": 3, 00:09:20.806 "base_bdevs_list": [ 00:09:20.806 { 00:09:20.806 "name": null, 00:09:20.806 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:20.806 "is_configured": false, 00:09:20.806 "data_offset": 0, 00:09:20.806 "data_size": 65536 00:09:20.806 }, 00:09:20.806 { 00:09:20.806 "name": "BaseBdev2", 00:09:20.806 "uuid": "73b126d2-0b3a-4c84-ad92-00e5863adc8d", 00:09:20.806 "is_configured": true, 00:09:20.806 "data_offset": 0, 00:09:20.806 "data_size": 65536 00:09:20.806 }, 00:09:20.806 { 00:09:20.806 "name": "BaseBdev3", 00:09:20.806 "uuid": "c527e570-ff29-41ef-b225-06652aa021d3", 00:09:20.806 "is_configured": true, 00:09:20.806 "data_offset": 0, 00:09:20.806 "data_size": 65536 00:09:20.806 }, 00:09:20.806 { 00:09:20.806 "name": "BaseBdev4", 00:09:20.806 "uuid": "878b19a4-d105-462d-80da-f190df600f1d", 00:09:20.806 "is_configured": true, 00:09:20.806 "data_offset": 0, 00:09:20.806 "data_size": 65536 00:09:20.806 } 00:09:20.806 ] 00:09:20.806 }' 00:09:20.806 23:43:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:20.806 23:43:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:21.373 23:43:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:09:21.373 23:43:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:21.373 23:43:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:21.373 23:43:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:21.373 23:43:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:21.373 23:43:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:09:21.373 23:43:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:21.373 23:43:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:09:21.373 23:43:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:09:21.373 23:43:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:09:21.373 23:43:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:21.373 23:43:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:21.373 [2024-11-26 23:43:09.390454] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:09:21.373 23:43:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:21.373 23:43:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:09:21.373 23:43:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:21.373 23:43:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:09:21.373 23:43:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:21.373 23:43:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:21.373 23:43:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:21.373 23:43:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:21.373 23:43:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:09:21.373 23:43:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:09:21.373 23:43:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:09:21.373 23:43:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:21.373 23:43:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:21.373 [2024-11-26 23:43:09.465461] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:09:21.373 23:43:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:21.373 23:43:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:09:21.373 23:43:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:21.373 23:43:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:09:21.373 23:43:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:21.373 23:43:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:21.373 23:43:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:21.373 23:43:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:21.630 23:43:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:09:21.630 23:43:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:09:21.630 23:43:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev4 00:09:21.630 23:43:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:21.630 23:43:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:21.630 [2024-11-26 23:43:09.532746] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev4 00:09:21.630 [2024-11-26 23:43:09.532788] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:09:21.630 23:43:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:21.630 23:43:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:09:21.630 23:43:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:21.630 23:43:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:09:21.630 23:43:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:21.630 23:43:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:21.630 23:43:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:21.630 23:43:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:21.630 23:43:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:09:21.630 23:43:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:09:21.630 23:43:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 4 -gt 2 ']' 00:09:21.630 23:43:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:09:21.630 23:43:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:21.630 23:43:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:09:21.630 23:43:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:21.630 23:43:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:21.630 BaseBdev2 00:09:21.630 23:43:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:21.630 23:43:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:09:21.630 23:43:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:09:21.630 23:43:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:21.630 23:43:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:09:21.630 23:43:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:21.630 23:43:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:21.630 23:43:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:21.630 23:43:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:21.630 23:43:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:21.630 23:43:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:21.630 23:43:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:09:21.630 23:43:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:21.630 23:43:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:21.630 [ 00:09:21.630 { 00:09:21.630 "name": "BaseBdev2", 00:09:21.630 "aliases": [ 00:09:21.630 "9abcfd63-d5f1-43d9-95bf-fcfde94de69c" 00:09:21.630 ], 00:09:21.630 "product_name": "Malloc disk", 00:09:21.630 "block_size": 512, 00:09:21.630 "num_blocks": 65536, 00:09:21.630 "uuid": "9abcfd63-d5f1-43d9-95bf-fcfde94de69c", 00:09:21.630 "assigned_rate_limits": { 00:09:21.630 "rw_ios_per_sec": 0, 00:09:21.630 "rw_mbytes_per_sec": 0, 00:09:21.630 "r_mbytes_per_sec": 0, 00:09:21.630 "w_mbytes_per_sec": 0 00:09:21.630 }, 00:09:21.630 "claimed": false, 00:09:21.630 "zoned": false, 00:09:21.630 "supported_io_types": { 00:09:21.630 "read": true, 00:09:21.630 "write": true, 00:09:21.630 "unmap": true, 00:09:21.630 "flush": true, 00:09:21.630 "reset": true, 00:09:21.630 "nvme_admin": false, 00:09:21.630 "nvme_io": false, 00:09:21.630 "nvme_io_md": false, 00:09:21.630 "write_zeroes": true, 00:09:21.630 "zcopy": true, 00:09:21.630 "get_zone_info": false, 00:09:21.630 "zone_management": false, 00:09:21.630 "zone_append": false, 00:09:21.630 "compare": false, 00:09:21.630 "compare_and_write": false, 00:09:21.630 "abort": true, 00:09:21.630 "seek_hole": false, 00:09:21.630 "seek_data": false, 00:09:21.630 "copy": true, 00:09:21.630 "nvme_iov_md": false 00:09:21.630 }, 00:09:21.630 "memory_domains": [ 00:09:21.630 { 00:09:21.630 "dma_device_id": "system", 00:09:21.631 "dma_device_type": 1 00:09:21.631 }, 00:09:21.631 { 00:09:21.631 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:21.631 "dma_device_type": 2 00:09:21.631 } 00:09:21.631 ], 00:09:21.631 "driver_specific": {} 00:09:21.631 } 00:09:21.631 ] 00:09:21.631 23:43:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:21.631 23:43:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:09:21.631 23:43:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:09:21.631 23:43:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:21.631 23:43:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:09:21.631 23:43:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:21.631 23:43:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:21.631 BaseBdev3 00:09:21.631 23:43:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:21.631 23:43:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:09:21.631 23:43:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:09:21.631 23:43:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:21.631 23:43:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:09:21.631 23:43:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:21.631 23:43:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:21.631 23:43:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:21.631 23:43:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:21.631 23:43:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:21.631 23:43:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:21.631 23:43:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:09:21.631 23:43:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:21.631 23:43:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:21.631 [ 00:09:21.631 { 00:09:21.631 "name": "BaseBdev3", 00:09:21.631 "aliases": [ 00:09:21.631 "e3085b88-3f49-4fb1-8e2d-8ed7161b7600" 00:09:21.631 ], 00:09:21.631 "product_name": "Malloc disk", 00:09:21.631 "block_size": 512, 00:09:21.631 "num_blocks": 65536, 00:09:21.631 "uuid": "e3085b88-3f49-4fb1-8e2d-8ed7161b7600", 00:09:21.631 "assigned_rate_limits": { 00:09:21.631 "rw_ios_per_sec": 0, 00:09:21.631 "rw_mbytes_per_sec": 0, 00:09:21.631 "r_mbytes_per_sec": 0, 00:09:21.631 "w_mbytes_per_sec": 0 00:09:21.631 }, 00:09:21.631 "claimed": false, 00:09:21.631 "zoned": false, 00:09:21.631 "supported_io_types": { 00:09:21.631 "read": true, 00:09:21.631 "write": true, 00:09:21.631 "unmap": true, 00:09:21.631 "flush": true, 00:09:21.631 "reset": true, 00:09:21.631 "nvme_admin": false, 00:09:21.631 "nvme_io": false, 00:09:21.631 "nvme_io_md": false, 00:09:21.631 "write_zeroes": true, 00:09:21.631 "zcopy": true, 00:09:21.631 "get_zone_info": false, 00:09:21.631 "zone_management": false, 00:09:21.631 "zone_append": false, 00:09:21.631 "compare": false, 00:09:21.631 "compare_and_write": false, 00:09:21.631 "abort": true, 00:09:21.631 "seek_hole": false, 00:09:21.631 "seek_data": false, 00:09:21.631 "copy": true, 00:09:21.631 "nvme_iov_md": false 00:09:21.631 }, 00:09:21.631 "memory_domains": [ 00:09:21.631 { 00:09:21.631 "dma_device_id": "system", 00:09:21.631 "dma_device_type": 1 00:09:21.631 }, 00:09:21.631 { 00:09:21.631 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:21.631 "dma_device_type": 2 00:09:21.631 } 00:09:21.631 ], 00:09:21.631 "driver_specific": {} 00:09:21.631 } 00:09:21.631 ] 00:09:21.631 23:43:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:21.631 23:43:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:09:21.631 23:43:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:09:21.631 23:43:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:21.631 23:43:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:09:21.631 23:43:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:21.631 23:43:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:21.631 BaseBdev4 00:09:21.631 23:43:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:21.631 23:43:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev4 00:09:21.631 23:43:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:09:21.631 23:43:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:21.631 23:43:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:09:21.631 23:43:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:21.631 23:43:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:21.631 23:43:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:21.631 23:43:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:21.631 23:43:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:21.631 23:43:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:21.631 23:43:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:09:21.631 23:43:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:21.631 23:43:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:21.631 [ 00:09:21.631 { 00:09:21.631 "name": "BaseBdev4", 00:09:21.631 "aliases": [ 00:09:21.631 "d72953f0-67f4-451d-8f23-4de591eaf290" 00:09:21.631 ], 00:09:21.631 "product_name": "Malloc disk", 00:09:21.631 "block_size": 512, 00:09:21.631 "num_blocks": 65536, 00:09:21.631 "uuid": "d72953f0-67f4-451d-8f23-4de591eaf290", 00:09:21.631 "assigned_rate_limits": { 00:09:21.631 "rw_ios_per_sec": 0, 00:09:21.631 "rw_mbytes_per_sec": 0, 00:09:21.631 "r_mbytes_per_sec": 0, 00:09:21.631 "w_mbytes_per_sec": 0 00:09:21.631 }, 00:09:21.631 "claimed": false, 00:09:21.631 "zoned": false, 00:09:21.631 "supported_io_types": { 00:09:21.631 "read": true, 00:09:21.631 "write": true, 00:09:21.631 "unmap": true, 00:09:21.631 "flush": true, 00:09:21.631 "reset": true, 00:09:21.631 "nvme_admin": false, 00:09:21.631 "nvme_io": false, 00:09:21.631 "nvme_io_md": false, 00:09:21.631 "write_zeroes": true, 00:09:21.631 "zcopy": true, 00:09:21.631 "get_zone_info": false, 00:09:21.631 "zone_management": false, 00:09:21.631 "zone_append": false, 00:09:21.631 "compare": false, 00:09:21.631 "compare_and_write": false, 00:09:21.631 "abort": true, 00:09:21.631 "seek_hole": false, 00:09:21.631 "seek_data": false, 00:09:21.631 "copy": true, 00:09:21.631 "nvme_iov_md": false 00:09:21.631 }, 00:09:21.631 "memory_domains": [ 00:09:21.631 { 00:09:21.631 "dma_device_id": "system", 00:09:21.631 "dma_device_type": 1 00:09:21.631 }, 00:09:21.631 { 00:09:21.631 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:21.631 "dma_device_type": 2 00:09:21.631 } 00:09:21.889 ], 00:09:21.890 "driver_specific": {} 00:09:21.890 } 00:09:21.890 ] 00:09:21.890 23:43:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:21.890 23:43:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:09:21.890 23:43:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:09:21.890 23:43:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:21.890 23:43:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:09:21.890 23:43:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:21.890 23:43:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:21.890 [2024-11-26 23:43:09.769538] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:21.890 [2024-11-26 23:43:09.769585] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:21.890 [2024-11-26 23:43:09.769634] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:21.890 [2024-11-26 23:43:09.771456] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:21.890 [2024-11-26 23:43:09.771505] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:09:21.890 23:43:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:21.890 23:43:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:09:21.890 23:43:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:21.890 23:43:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:21.890 23:43:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:21.890 23:43:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:21.890 23:43:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:21.890 23:43:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:21.890 23:43:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:21.890 23:43:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:21.890 23:43:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:21.890 23:43:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:21.890 23:43:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:21.890 23:43:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:21.890 23:43:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:21.890 23:43:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:21.890 23:43:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:21.890 "name": "Existed_Raid", 00:09:21.890 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:21.890 "strip_size_kb": 64, 00:09:21.890 "state": "configuring", 00:09:21.890 "raid_level": "concat", 00:09:21.890 "superblock": false, 00:09:21.890 "num_base_bdevs": 4, 00:09:21.890 "num_base_bdevs_discovered": 3, 00:09:21.890 "num_base_bdevs_operational": 4, 00:09:21.890 "base_bdevs_list": [ 00:09:21.890 { 00:09:21.890 "name": "BaseBdev1", 00:09:21.890 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:21.890 "is_configured": false, 00:09:21.890 "data_offset": 0, 00:09:21.890 "data_size": 0 00:09:21.890 }, 00:09:21.890 { 00:09:21.890 "name": "BaseBdev2", 00:09:21.890 "uuid": "9abcfd63-d5f1-43d9-95bf-fcfde94de69c", 00:09:21.890 "is_configured": true, 00:09:21.890 "data_offset": 0, 00:09:21.890 "data_size": 65536 00:09:21.890 }, 00:09:21.890 { 00:09:21.890 "name": "BaseBdev3", 00:09:21.890 "uuid": "e3085b88-3f49-4fb1-8e2d-8ed7161b7600", 00:09:21.890 "is_configured": true, 00:09:21.890 "data_offset": 0, 00:09:21.890 "data_size": 65536 00:09:21.890 }, 00:09:21.890 { 00:09:21.890 "name": "BaseBdev4", 00:09:21.890 "uuid": "d72953f0-67f4-451d-8f23-4de591eaf290", 00:09:21.890 "is_configured": true, 00:09:21.890 "data_offset": 0, 00:09:21.890 "data_size": 65536 00:09:21.890 } 00:09:21.890 ] 00:09:21.890 }' 00:09:21.890 23:43:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:21.890 23:43:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:22.149 23:43:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:09:22.149 23:43:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:22.149 23:43:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:22.149 [2024-11-26 23:43:10.240755] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:09:22.149 23:43:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:22.149 23:43:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:09:22.149 23:43:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:22.149 23:43:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:22.149 23:43:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:22.149 23:43:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:22.149 23:43:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:22.149 23:43:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:22.149 23:43:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:22.149 23:43:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:22.149 23:43:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:22.149 23:43:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:22.149 23:43:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:22.149 23:43:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:22.149 23:43:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:22.149 23:43:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:22.407 23:43:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:22.407 "name": "Existed_Raid", 00:09:22.407 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:22.407 "strip_size_kb": 64, 00:09:22.407 "state": "configuring", 00:09:22.407 "raid_level": "concat", 00:09:22.407 "superblock": false, 00:09:22.407 "num_base_bdevs": 4, 00:09:22.407 "num_base_bdevs_discovered": 2, 00:09:22.407 "num_base_bdevs_operational": 4, 00:09:22.407 "base_bdevs_list": [ 00:09:22.407 { 00:09:22.407 "name": "BaseBdev1", 00:09:22.407 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:22.407 "is_configured": false, 00:09:22.407 "data_offset": 0, 00:09:22.407 "data_size": 0 00:09:22.407 }, 00:09:22.407 { 00:09:22.407 "name": null, 00:09:22.407 "uuid": "9abcfd63-d5f1-43d9-95bf-fcfde94de69c", 00:09:22.407 "is_configured": false, 00:09:22.407 "data_offset": 0, 00:09:22.407 "data_size": 65536 00:09:22.407 }, 00:09:22.407 { 00:09:22.407 "name": "BaseBdev3", 00:09:22.407 "uuid": "e3085b88-3f49-4fb1-8e2d-8ed7161b7600", 00:09:22.407 "is_configured": true, 00:09:22.407 "data_offset": 0, 00:09:22.407 "data_size": 65536 00:09:22.407 }, 00:09:22.407 { 00:09:22.407 "name": "BaseBdev4", 00:09:22.407 "uuid": "d72953f0-67f4-451d-8f23-4de591eaf290", 00:09:22.407 "is_configured": true, 00:09:22.407 "data_offset": 0, 00:09:22.407 "data_size": 65536 00:09:22.407 } 00:09:22.407 ] 00:09:22.407 }' 00:09:22.407 23:43:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:22.407 23:43:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:22.666 23:43:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:22.666 23:43:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:22.666 23:43:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:22.666 23:43:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:09:22.666 23:43:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:22.666 23:43:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:09:22.666 23:43:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:09:22.666 23:43:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:22.666 23:43:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:22.666 [2024-11-26 23:43:10.747002] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:22.666 BaseBdev1 00:09:22.666 23:43:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:22.666 23:43:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:09:22.666 23:43:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:09:22.666 23:43:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:22.666 23:43:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:09:22.666 23:43:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:22.666 23:43:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:22.666 23:43:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:22.666 23:43:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:22.666 23:43:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:22.666 23:43:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:22.666 23:43:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:09:22.666 23:43:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:22.666 23:43:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:22.666 [ 00:09:22.666 { 00:09:22.666 "name": "BaseBdev1", 00:09:22.666 "aliases": [ 00:09:22.666 "ce9228ff-2d5b-4e03-ac16-3e28fa151eb4" 00:09:22.666 ], 00:09:22.666 "product_name": "Malloc disk", 00:09:22.666 "block_size": 512, 00:09:22.666 "num_blocks": 65536, 00:09:22.666 "uuid": "ce9228ff-2d5b-4e03-ac16-3e28fa151eb4", 00:09:22.666 "assigned_rate_limits": { 00:09:22.666 "rw_ios_per_sec": 0, 00:09:22.666 "rw_mbytes_per_sec": 0, 00:09:22.666 "r_mbytes_per_sec": 0, 00:09:22.666 "w_mbytes_per_sec": 0 00:09:22.666 }, 00:09:22.666 "claimed": true, 00:09:22.666 "claim_type": "exclusive_write", 00:09:22.666 "zoned": false, 00:09:22.666 "supported_io_types": { 00:09:22.666 "read": true, 00:09:22.666 "write": true, 00:09:22.666 "unmap": true, 00:09:22.666 "flush": true, 00:09:22.666 "reset": true, 00:09:22.666 "nvme_admin": false, 00:09:22.666 "nvme_io": false, 00:09:22.666 "nvme_io_md": false, 00:09:22.666 "write_zeroes": true, 00:09:22.666 "zcopy": true, 00:09:22.666 "get_zone_info": false, 00:09:22.666 "zone_management": false, 00:09:22.666 "zone_append": false, 00:09:22.666 "compare": false, 00:09:22.666 "compare_and_write": false, 00:09:22.666 "abort": true, 00:09:22.666 "seek_hole": false, 00:09:22.666 "seek_data": false, 00:09:22.667 "copy": true, 00:09:22.667 "nvme_iov_md": false 00:09:22.667 }, 00:09:22.667 "memory_domains": [ 00:09:22.667 { 00:09:22.667 "dma_device_id": "system", 00:09:22.667 "dma_device_type": 1 00:09:22.667 }, 00:09:22.667 { 00:09:22.667 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:22.667 "dma_device_type": 2 00:09:22.667 } 00:09:22.667 ], 00:09:22.667 "driver_specific": {} 00:09:22.667 } 00:09:22.667 ] 00:09:22.667 23:43:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:22.667 23:43:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:09:22.667 23:43:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:09:22.667 23:43:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:22.667 23:43:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:22.667 23:43:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:22.667 23:43:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:22.667 23:43:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:22.667 23:43:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:22.667 23:43:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:22.667 23:43:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:22.667 23:43:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:22.667 23:43:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:22.667 23:43:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:22.667 23:43:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:22.667 23:43:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:22.926 23:43:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:22.926 23:43:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:22.926 "name": "Existed_Raid", 00:09:22.926 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:22.926 "strip_size_kb": 64, 00:09:22.926 "state": "configuring", 00:09:22.926 "raid_level": "concat", 00:09:22.926 "superblock": false, 00:09:22.926 "num_base_bdevs": 4, 00:09:22.926 "num_base_bdevs_discovered": 3, 00:09:22.926 "num_base_bdevs_operational": 4, 00:09:22.926 "base_bdevs_list": [ 00:09:22.926 { 00:09:22.926 "name": "BaseBdev1", 00:09:22.926 "uuid": "ce9228ff-2d5b-4e03-ac16-3e28fa151eb4", 00:09:22.926 "is_configured": true, 00:09:22.926 "data_offset": 0, 00:09:22.926 "data_size": 65536 00:09:22.926 }, 00:09:22.926 { 00:09:22.926 "name": null, 00:09:22.926 "uuid": "9abcfd63-d5f1-43d9-95bf-fcfde94de69c", 00:09:22.926 "is_configured": false, 00:09:22.926 "data_offset": 0, 00:09:22.926 "data_size": 65536 00:09:22.926 }, 00:09:22.926 { 00:09:22.926 "name": "BaseBdev3", 00:09:22.926 "uuid": "e3085b88-3f49-4fb1-8e2d-8ed7161b7600", 00:09:22.926 "is_configured": true, 00:09:22.926 "data_offset": 0, 00:09:22.926 "data_size": 65536 00:09:22.926 }, 00:09:22.926 { 00:09:22.926 "name": "BaseBdev4", 00:09:22.926 "uuid": "d72953f0-67f4-451d-8f23-4de591eaf290", 00:09:22.926 "is_configured": true, 00:09:22.926 "data_offset": 0, 00:09:22.926 "data_size": 65536 00:09:22.926 } 00:09:22.926 ] 00:09:22.926 }' 00:09:22.926 23:43:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:22.926 23:43:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:23.184 23:43:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:23.184 23:43:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:23.184 23:43:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:23.184 23:43:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:09:23.184 23:43:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:23.184 23:43:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:09:23.184 23:43:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:09:23.184 23:43:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:23.184 23:43:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:23.184 [2024-11-26 23:43:11.278151] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:09:23.184 23:43:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:23.184 23:43:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:09:23.184 23:43:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:23.184 23:43:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:23.184 23:43:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:23.184 23:43:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:23.184 23:43:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:23.184 23:43:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:23.184 23:43:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:23.184 23:43:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:23.184 23:43:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:23.184 23:43:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:23.184 23:43:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:23.184 23:43:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:23.184 23:43:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:23.184 23:43:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:23.441 23:43:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:23.441 "name": "Existed_Raid", 00:09:23.441 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:23.441 "strip_size_kb": 64, 00:09:23.441 "state": "configuring", 00:09:23.441 "raid_level": "concat", 00:09:23.441 "superblock": false, 00:09:23.441 "num_base_bdevs": 4, 00:09:23.441 "num_base_bdevs_discovered": 2, 00:09:23.441 "num_base_bdevs_operational": 4, 00:09:23.441 "base_bdevs_list": [ 00:09:23.441 { 00:09:23.441 "name": "BaseBdev1", 00:09:23.441 "uuid": "ce9228ff-2d5b-4e03-ac16-3e28fa151eb4", 00:09:23.441 "is_configured": true, 00:09:23.441 "data_offset": 0, 00:09:23.441 "data_size": 65536 00:09:23.441 }, 00:09:23.441 { 00:09:23.441 "name": null, 00:09:23.441 "uuid": "9abcfd63-d5f1-43d9-95bf-fcfde94de69c", 00:09:23.441 "is_configured": false, 00:09:23.441 "data_offset": 0, 00:09:23.441 "data_size": 65536 00:09:23.441 }, 00:09:23.441 { 00:09:23.441 "name": null, 00:09:23.441 "uuid": "e3085b88-3f49-4fb1-8e2d-8ed7161b7600", 00:09:23.441 "is_configured": false, 00:09:23.441 "data_offset": 0, 00:09:23.441 "data_size": 65536 00:09:23.441 }, 00:09:23.441 { 00:09:23.441 "name": "BaseBdev4", 00:09:23.441 "uuid": "d72953f0-67f4-451d-8f23-4de591eaf290", 00:09:23.441 "is_configured": true, 00:09:23.441 "data_offset": 0, 00:09:23.441 "data_size": 65536 00:09:23.441 } 00:09:23.441 ] 00:09:23.441 }' 00:09:23.441 23:43:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:23.441 23:43:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:23.698 23:43:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:23.698 23:43:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:09:23.698 23:43:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:23.698 23:43:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:23.698 23:43:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:23.698 23:43:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:09:23.698 23:43:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:09:23.698 23:43:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:23.698 23:43:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:23.698 [2024-11-26 23:43:11.793296] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:23.698 23:43:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:23.698 23:43:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:09:23.698 23:43:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:23.698 23:43:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:23.698 23:43:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:23.698 23:43:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:23.698 23:43:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:23.698 23:43:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:23.698 23:43:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:23.698 23:43:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:23.698 23:43:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:23.698 23:43:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:23.698 23:43:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:23.698 23:43:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:23.698 23:43:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:23.698 23:43:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:23.955 23:43:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:23.955 "name": "Existed_Raid", 00:09:23.955 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:23.955 "strip_size_kb": 64, 00:09:23.955 "state": "configuring", 00:09:23.955 "raid_level": "concat", 00:09:23.955 "superblock": false, 00:09:23.955 "num_base_bdevs": 4, 00:09:23.955 "num_base_bdevs_discovered": 3, 00:09:23.955 "num_base_bdevs_operational": 4, 00:09:23.955 "base_bdevs_list": [ 00:09:23.955 { 00:09:23.955 "name": "BaseBdev1", 00:09:23.955 "uuid": "ce9228ff-2d5b-4e03-ac16-3e28fa151eb4", 00:09:23.955 "is_configured": true, 00:09:23.955 "data_offset": 0, 00:09:23.955 "data_size": 65536 00:09:23.955 }, 00:09:23.955 { 00:09:23.955 "name": null, 00:09:23.955 "uuid": "9abcfd63-d5f1-43d9-95bf-fcfde94de69c", 00:09:23.955 "is_configured": false, 00:09:23.955 "data_offset": 0, 00:09:23.955 "data_size": 65536 00:09:23.955 }, 00:09:23.955 { 00:09:23.955 "name": "BaseBdev3", 00:09:23.955 "uuid": "e3085b88-3f49-4fb1-8e2d-8ed7161b7600", 00:09:23.955 "is_configured": true, 00:09:23.955 "data_offset": 0, 00:09:23.955 "data_size": 65536 00:09:23.955 }, 00:09:23.955 { 00:09:23.955 "name": "BaseBdev4", 00:09:23.955 "uuid": "d72953f0-67f4-451d-8f23-4de591eaf290", 00:09:23.955 "is_configured": true, 00:09:23.955 "data_offset": 0, 00:09:23.955 "data_size": 65536 00:09:23.955 } 00:09:23.955 ] 00:09:23.955 }' 00:09:23.955 23:43:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:23.955 23:43:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:24.213 23:43:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:24.213 23:43:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:24.213 23:43:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:24.213 23:43:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:09:24.213 23:43:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:24.213 23:43:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:09:24.213 23:43:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:09:24.213 23:43:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:24.213 23:43:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:24.213 [2024-11-26 23:43:12.260515] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:09:24.213 23:43:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:24.213 23:43:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:09:24.213 23:43:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:24.213 23:43:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:24.213 23:43:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:24.213 23:43:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:24.213 23:43:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:24.213 23:43:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:24.213 23:43:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:24.213 23:43:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:24.213 23:43:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:24.213 23:43:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:24.213 23:43:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:24.213 23:43:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:24.213 23:43:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:24.213 23:43:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:24.213 23:43:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:24.213 "name": "Existed_Raid", 00:09:24.213 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:24.213 "strip_size_kb": 64, 00:09:24.213 "state": "configuring", 00:09:24.213 "raid_level": "concat", 00:09:24.213 "superblock": false, 00:09:24.213 "num_base_bdevs": 4, 00:09:24.213 "num_base_bdevs_discovered": 2, 00:09:24.213 "num_base_bdevs_operational": 4, 00:09:24.213 "base_bdevs_list": [ 00:09:24.213 { 00:09:24.213 "name": null, 00:09:24.213 "uuid": "ce9228ff-2d5b-4e03-ac16-3e28fa151eb4", 00:09:24.213 "is_configured": false, 00:09:24.214 "data_offset": 0, 00:09:24.214 "data_size": 65536 00:09:24.214 }, 00:09:24.214 { 00:09:24.214 "name": null, 00:09:24.214 "uuid": "9abcfd63-d5f1-43d9-95bf-fcfde94de69c", 00:09:24.214 "is_configured": false, 00:09:24.214 "data_offset": 0, 00:09:24.214 "data_size": 65536 00:09:24.214 }, 00:09:24.214 { 00:09:24.214 "name": "BaseBdev3", 00:09:24.214 "uuid": "e3085b88-3f49-4fb1-8e2d-8ed7161b7600", 00:09:24.214 "is_configured": true, 00:09:24.214 "data_offset": 0, 00:09:24.214 "data_size": 65536 00:09:24.214 }, 00:09:24.214 { 00:09:24.214 "name": "BaseBdev4", 00:09:24.214 "uuid": "d72953f0-67f4-451d-8f23-4de591eaf290", 00:09:24.214 "is_configured": true, 00:09:24.214 "data_offset": 0, 00:09:24.214 "data_size": 65536 00:09:24.214 } 00:09:24.214 ] 00:09:24.214 }' 00:09:24.214 23:43:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:24.214 23:43:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:24.821 23:43:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:24.821 23:43:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:09:24.821 23:43:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:24.821 23:43:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:24.821 23:43:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:24.821 23:43:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:09:24.821 23:43:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:09:24.821 23:43:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:24.821 23:43:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:24.821 [2024-11-26 23:43:12.742343] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:24.821 23:43:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:24.821 23:43:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:09:24.821 23:43:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:24.821 23:43:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:24.821 23:43:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:24.821 23:43:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:24.821 23:43:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:24.821 23:43:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:24.821 23:43:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:24.821 23:43:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:24.821 23:43:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:24.821 23:43:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:24.821 23:43:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:24.821 23:43:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:24.821 23:43:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:24.821 23:43:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:24.821 23:43:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:24.821 "name": "Existed_Raid", 00:09:24.821 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:24.821 "strip_size_kb": 64, 00:09:24.821 "state": "configuring", 00:09:24.821 "raid_level": "concat", 00:09:24.821 "superblock": false, 00:09:24.821 "num_base_bdevs": 4, 00:09:24.821 "num_base_bdevs_discovered": 3, 00:09:24.821 "num_base_bdevs_operational": 4, 00:09:24.821 "base_bdevs_list": [ 00:09:24.821 { 00:09:24.821 "name": null, 00:09:24.821 "uuid": "ce9228ff-2d5b-4e03-ac16-3e28fa151eb4", 00:09:24.821 "is_configured": false, 00:09:24.821 "data_offset": 0, 00:09:24.821 "data_size": 65536 00:09:24.821 }, 00:09:24.821 { 00:09:24.821 "name": "BaseBdev2", 00:09:24.821 "uuid": "9abcfd63-d5f1-43d9-95bf-fcfde94de69c", 00:09:24.821 "is_configured": true, 00:09:24.821 "data_offset": 0, 00:09:24.821 "data_size": 65536 00:09:24.821 }, 00:09:24.821 { 00:09:24.821 "name": "BaseBdev3", 00:09:24.821 "uuid": "e3085b88-3f49-4fb1-8e2d-8ed7161b7600", 00:09:24.821 "is_configured": true, 00:09:24.821 "data_offset": 0, 00:09:24.821 "data_size": 65536 00:09:24.821 }, 00:09:24.821 { 00:09:24.821 "name": "BaseBdev4", 00:09:24.821 "uuid": "d72953f0-67f4-451d-8f23-4de591eaf290", 00:09:24.821 "is_configured": true, 00:09:24.821 "data_offset": 0, 00:09:24.821 "data_size": 65536 00:09:24.821 } 00:09:24.821 ] 00:09:24.821 }' 00:09:24.821 23:43:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:24.821 23:43:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:25.078 23:43:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:09:25.078 23:43:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:25.078 23:43:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:25.078 23:43:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:25.078 23:43:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:25.336 23:43:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:09:25.336 23:43:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:25.336 23:43:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:09:25.336 23:43:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:25.336 23:43:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:25.336 23:43:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:25.336 23:43:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u ce9228ff-2d5b-4e03-ac16-3e28fa151eb4 00:09:25.336 23:43:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:25.336 23:43:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:25.336 [2024-11-26 23:43:13.272335] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:09:25.336 [2024-11-26 23:43:13.272394] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:09:25.336 [2024-11-26 23:43:13.272401] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 262144, blocklen 512 00:09:25.336 [2024-11-26 23:43:13.272651] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002a10 00:09:25.336 [2024-11-26 23:43:13.272767] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:09:25.336 [2024-11-26 23:43:13.272782] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001c80 00:09:25.336 [2024-11-26 23:43:13.272942] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:25.336 NewBaseBdev 00:09:25.336 23:43:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:25.336 23:43:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:09:25.336 23:43:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:09:25.336 23:43:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:25.336 23:43:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:09:25.336 23:43:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:25.336 23:43:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:25.336 23:43:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:25.336 23:43:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:25.336 23:43:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:25.336 23:43:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:25.336 23:43:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:09:25.336 23:43:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:25.336 23:43:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:25.336 [ 00:09:25.336 { 00:09:25.336 "name": "NewBaseBdev", 00:09:25.336 "aliases": [ 00:09:25.336 "ce9228ff-2d5b-4e03-ac16-3e28fa151eb4" 00:09:25.336 ], 00:09:25.336 "product_name": "Malloc disk", 00:09:25.336 "block_size": 512, 00:09:25.336 "num_blocks": 65536, 00:09:25.336 "uuid": "ce9228ff-2d5b-4e03-ac16-3e28fa151eb4", 00:09:25.336 "assigned_rate_limits": { 00:09:25.336 "rw_ios_per_sec": 0, 00:09:25.336 "rw_mbytes_per_sec": 0, 00:09:25.336 "r_mbytes_per_sec": 0, 00:09:25.336 "w_mbytes_per_sec": 0 00:09:25.336 }, 00:09:25.336 "claimed": true, 00:09:25.336 "claim_type": "exclusive_write", 00:09:25.336 "zoned": false, 00:09:25.336 "supported_io_types": { 00:09:25.336 "read": true, 00:09:25.336 "write": true, 00:09:25.336 "unmap": true, 00:09:25.336 "flush": true, 00:09:25.336 "reset": true, 00:09:25.336 "nvme_admin": false, 00:09:25.336 "nvme_io": false, 00:09:25.336 "nvme_io_md": false, 00:09:25.336 "write_zeroes": true, 00:09:25.336 "zcopy": true, 00:09:25.336 "get_zone_info": false, 00:09:25.336 "zone_management": false, 00:09:25.336 "zone_append": false, 00:09:25.337 "compare": false, 00:09:25.337 "compare_and_write": false, 00:09:25.337 "abort": true, 00:09:25.337 "seek_hole": false, 00:09:25.337 "seek_data": false, 00:09:25.337 "copy": true, 00:09:25.337 "nvme_iov_md": false 00:09:25.337 }, 00:09:25.337 "memory_domains": [ 00:09:25.337 { 00:09:25.337 "dma_device_id": "system", 00:09:25.337 "dma_device_type": 1 00:09:25.337 }, 00:09:25.337 { 00:09:25.337 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:25.337 "dma_device_type": 2 00:09:25.337 } 00:09:25.337 ], 00:09:25.337 "driver_specific": {} 00:09:25.337 } 00:09:25.337 ] 00:09:25.337 23:43:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:25.337 23:43:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:09:25.337 23:43:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online concat 64 4 00:09:25.337 23:43:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:25.337 23:43:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:25.337 23:43:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:25.337 23:43:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:25.337 23:43:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:25.337 23:43:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:25.337 23:43:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:25.337 23:43:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:25.337 23:43:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:25.337 23:43:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:25.337 23:43:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:25.337 23:43:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:25.337 23:43:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:25.337 23:43:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:25.337 23:43:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:25.337 "name": "Existed_Raid", 00:09:25.337 "uuid": "793e3c17-7397-48d3-94cd-42f02f96b29e", 00:09:25.337 "strip_size_kb": 64, 00:09:25.337 "state": "online", 00:09:25.337 "raid_level": "concat", 00:09:25.337 "superblock": false, 00:09:25.337 "num_base_bdevs": 4, 00:09:25.337 "num_base_bdevs_discovered": 4, 00:09:25.337 "num_base_bdevs_operational": 4, 00:09:25.337 "base_bdevs_list": [ 00:09:25.337 { 00:09:25.337 "name": "NewBaseBdev", 00:09:25.337 "uuid": "ce9228ff-2d5b-4e03-ac16-3e28fa151eb4", 00:09:25.337 "is_configured": true, 00:09:25.337 "data_offset": 0, 00:09:25.337 "data_size": 65536 00:09:25.337 }, 00:09:25.337 { 00:09:25.337 "name": "BaseBdev2", 00:09:25.337 "uuid": "9abcfd63-d5f1-43d9-95bf-fcfde94de69c", 00:09:25.337 "is_configured": true, 00:09:25.337 "data_offset": 0, 00:09:25.337 "data_size": 65536 00:09:25.337 }, 00:09:25.337 { 00:09:25.337 "name": "BaseBdev3", 00:09:25.337 "uuid": "e3085b88-3f49-4fb1-8e2d-8ed7161b7600", 00:09:25.337 "is_configured": true, 00:09:25.337 "data_offset": 0, 00:09:25.337 "data_size": 65536 00:09:25.337 }, 00:09:25.337 { 00:09:25.337 "name": "BaseBdev4", 00:09:25.337 "uuid": "d72953f0-67f4-451d-8f23-4de591eaf290", 00:09:25.337 "is_configured": true, 00:09:25.337 "data_offset": 0, 00:09:25.337 "data_size": 65536 00:09:25.337 } 00:09:25.337 ] 00:09:25.337 }' 00:09:25.337 23:43:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:25.337 23:43:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:25.598 23:43:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:09:25.598 23:43:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:09:25.598 23:43:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:25.598 23:43:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:25.598 23:43:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:09:25.598 23:43:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:25.598 23:43:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:09:25.598 23:43:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:25.598 23:43:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:25.598 23:43:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:25.598 [2024-11-26 23:43:13.715933] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:25.863 23:43:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:25.863 23:43:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:25.863 "name": "Existed_Raid", 00:09:25.863 "aliases": [ 00:09:25.863 "793e3c17-7397-48d3-94cd-42f02f96b29e" 00:09:25.863 ], 00:09:25.863 "product_name": "Raid Volume", 00:09:25.863 "block_size": 512, 00:09:25.863 "num_blocks": 262144, 00:09:25.863 "uuid": "793e3c17-7397-48d3-94cd-42f02f96b29e", 00:09:25.863 "assigned_rate_limits": { 00:09:25.863 "rw_ios_per_sec": 0, 00:09:25.863 "rw_mbytes_per_sec": 0, 00:09:25.863 "r_mbytes_per_sec": 0, 00:09:25.863 "w_mbytes_per_sec": 0 00:09:25.863 }, 00:09:25.863 "claimed": false, 00:09:25.863 "zoned": false, 00:09:25.863 "supported_io_types": { 00:09:25.863 "read": true, 00:09:25.863 "write": true, 00:09:25.863 "unmap": true, 00:09:25.863 "flush": true, 00:09:25.863 "reset": true, 00:09:25.863 "nvme_admin": false, 00:09:25.863 "nvme_io": false, 00:09:25.863 "nvme_io_md": false, 00:09:25.863 "write_zeroes": true, 00:09:25.863 "zcopy": false, 00:09:25.863 "get_zone_info": false, 00:09:25.863 "zone_management": false, 00:09:25.863 "zone_append": false, 00:09:25.863 "compare": false, 00:09:25.863 "compare_and_write": false, 00:09:25.863 "abort": false, 00:09:25.863 "seek_hole": false, 00:09:25.863 "seek_data": false, 00:09:25.863 "copy": false, 00:09:25.863 "nvme_iov_md": false 00:09:25.863 }, 00:09:25.863 "memory_domains": [ 00:09:25.863 { 00:09:25.863 "dma_device_id": "system", 00:09:25.863 "dma_device_type": 1 00:09:25.863 }, 00:09:25.863 { 00:09:25.863 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:25.863 "dma_device_type": 2 00:09:25.863 }, 00:09:25.863 { 00:09:25.863 "dma_device_id": "system", 00:09:25.863 "dma_device_type": 1 00:09:25.863 }, 00:09:25.863 { 00:09:25.863 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:25.863 "dma_device_type": 2 00:09:25.863 }, 00:09:25.863 { 00:09:25.863 "dma_device_id": "system", 00:09:25.863 "dma_device_type": 1 00:09:25.863 }, 00:09:25.863 { 00:09:25.863 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:25.863 "dma_device_type": 2 00:09:25.863 }, 00:09:25.863 { 00:09:25.863 "dma_device_id": "system", 00:09:25.863 "dma_device_type": 1 00:09:25.863 }, 00:09:25.863 { 00:09:25.863 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:25.863 "dma_device_type": 2 00:09:25.863 } 00:09:25.863 ], 00:09:25.863 "driver_specific": { 00:09:25.863 "raid": { 00:09:25.863 "uuid": "793e3c17-7397-48d3-94cd-42f02f96b29e", 00:09:25.863 "strip_size_kb": 64, 00:09:25.863 "state": "online", 00:09:25.863 "raid_level": "concat", 00:09:25.863 "superblock": false, 00:09:25.863 "num_base_bdevs": 4, 00:09:25.863 "num_base_bdevs_discovered": 4, 00:09:25.863 "num_base_bdevs_operational": 4, 00:09:25.863 "base_bdevs_list": [ 00:09:25.863 { 00:09:25.863 "name": "NewBaseBdev", 00:09:25.863 "uuid": "ce9228ff-2d5b-4e03-ac16-3e28fa151eb4", 00:09:25.863 "is_configured": true, 00:09:25.863 "data_offset": 0, 00:09:25.863 "data_size": 65536 00:09:25.863 }, 00:09:25.863 { 00:09:25.863 "name": "BaseBdev2", 00:09:25.863 "uuid": "9abcfd63-d5f1-43d9-95bf-fcfde94de69c", 00:09:25.863 "is_configured": true, 00:09:25.863 "data_offset": 0, 00:09:25.863 "data_size": 65536 00:09:25.863 }, 00:09:25.863 { 00:09:25.863 "name": "BaseBdev3", 00:09:25.863 "uuid": "e3085b88-3f49-4fb1-8e2d-8ed7161b7600", 00:09:25.863 "is_configured": true, 00:09:25.863 "data_offset": 0, 00:09:25.863 "data_size": 65536 00:09:25.863 }, 00:09:25.863 { 00:09:25.863 "name": "BaseBdev4", 00:09:25.863 "uuid": "d72953f0-67f4-451d-8f23-4de591eaf290", 00:09:25.863 "is_configured": true, 00:09:25.863 "data_offset": 0, 00:09:25.863 "data_size": 65536 00:09:25.863 } 00:09:25.863 ] 00:09:25.863 } 00:09:25.863 } 00:09:25.863 }' 00:09:25.863 23:43:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:25.863 23:43:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:09:25.863 BaseBdev2 00:09:25.863 BaseBdev3 00:09:25.863 BaseBdev4' 00:09:25.863 23:43:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:25.863 23:43:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:25.863 23:43:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:25.863 23:43:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:09:25.863 23:43:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:25.863 23:43:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:25.863 23:43:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:25.863 23:43:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:25.863 23:43:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:25.863 23:43:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:25.863 23:43:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:25.863 23:43:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:09:25.863 23:43:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:25.863 23:43:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:25.863 23:43:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:25.863 23:43:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:25.864 23:43:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:25.864 23:43:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:25.864 23:43:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:25.864 23:43:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:25.864 23:43:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:09:25.864 23:43:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:25.864 23:43:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:25.864 23:43:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:25.864 23:43:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:25.864 23:43:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:25.864 23:43:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:25.864 23:43:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:09:25.864 23:43:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:25.864 23:43:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:25.864 23:43:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:26.130 23:43:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:26.131 23:43:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:26.131 23:43:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:26.131 23:43:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:26.131 23:43:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:26.131 23:43:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:26.131 [2024-11-26 23:43:14.023077] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:26.131 [2024-11-26 23:43:14.023110] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:26.131 [2024-11-26 23:43:14.023180] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:26.131 [2024-11-26 23:43:14.023245] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:26.131 [2024-11-26 23:43:14.023267] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name Existed_Raid, state offline 00:09:26.131 23:43:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:26.131 23:43:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 81882 00:09:26.131 23:43:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # '[' -z 81882 ']' 00:09:26.131 23:43:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@958 -- # kill -0 81882 00:09:26.131 23:43:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # uname 00:09:26.131 23:43:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:09:26.131 23:43:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 81882 00:09:26.131 killing process with pid 81882 00:09:26.131 23:43:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:09:26.131 23:43:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:09:26.131 23:43:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 81882' 00:09:26.131 23:43:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@973 -- # kill 81882 00:09:26.131 [2024-11-26 23:43:14.071993] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:09:26.131 23:43:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@978 -- # wait 81882 00:09:26.131 [2024-11-26 23:43:14.111973] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:09:26.390 ************************************ 00:09:26.390 END TEST raid_state_function_test 00:09:26.390 ************************************ 00:09:26.390 23:43:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:09:26.390 00:09:26.390 real 0m9.488s 00:09:26.390 user 0m16.300s 00:09:26.390 sys 0m1.905s 00:09:26.390 23:43:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:09:26.390 23:43:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:26.390 23:43:14 bdev_raid -- bdev/bdev_raid.sh@969 -- # run_test raid_state_function_test_sb raid_state_function_test concat 4 true 00:09:26.390 23:43:14 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:09:26.390 23:43:14 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:09:26.390 23:43:14 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:09:26.390 ************************************ 00:09:26.390 START TEST raid_state_function_test_sb 00:09:26.390 ************************************ 00:09:26.390 23:43:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1129 -- # raid_state_function_test concat 4 true 00:09:26.390 23:43:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=concat 00:09:26.390 23:43:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=4 00:09:26.390 23:43:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:09:26.390 23:43:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:09:26.390 23:43:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:09:26.390 23:43:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:26.390 23:43:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:09:26.390 23:43:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:26.391 23:43:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:26.391 23:43:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:09:26.391 23:43:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:26.391 23:43:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:26.391 23:43:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:09:26.391 23:43:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:26.391 23:43:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:26.391 23:43:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev4 00:09:26.391 23:43:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:26.391 23:43:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:26.391 23:43:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:09:26.391 23:43:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:09:26.391 23:43:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:09:26.391 23:43:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:09:26.391 23:43:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:09:26.391 23:43:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:09:26.391 23:43:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' concat '!=' raid1 ']' 00:09:26.391 23:43:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:09:26.391 23:43:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:09:26.391 23:43:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:09:26.391 23:43:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:09:26.391 23:43:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=82531 00:09:26.391 23:43:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 82531' 00:09:26.391 Process raid pid: 82531 00:09:26.391 23:43:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 82531 00:09:26.391 23:43:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:09:26.391 23:43:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@835 -- # '[' -z 82531 ']' 00:09:26.391 23:43:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:26.391 23:43:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@840 -- # local max_retries=100 00:09:26.391 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:26.391 23:43:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:26.391 23:43:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@844 -- # xtrace_disable 00:09:26.391 23:43:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:26.391 [2024-11-26 23:43:14.491154] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:09:26.391 [2024-11-26 23:43:14.491280] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:09:26.650 [2024-11-26 23:43:14.626657] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:26.650 [2024-11-26 23:43:14.651350] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:09:26.650 [2024-11-26 23:43:14.694151] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:26.650 [2024-11-26 23:43:14.694211] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:27.219 23:43:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:09:27.219 23:43:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@868 -- # return 0 00:09:27.219 23:43:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:09:27.219 23:43:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:27.219 23:43:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:27.219 [2024-11-26 23:43:15.320969] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:27.219 [2024-11-26 23:43:15.321021] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:27.219 [2024-11-26 23:43:15.321032] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:27.219 [2024-11-26 23:43:15.321041] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:27.219 [2024-11-26 23:43:15.321047] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:27.219 [2024-11-26 23:43:15.321060] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:27.219 [2024-11-26 23:43:15.321066] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:09:27.219 [2024-11-26 23:43:15.321074] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:09:27.219 23:43:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:27.219 23:43:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:09:27.219 23:43:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:27.219 23:43:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:27.219 23:43:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:27.219 23:43:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:27.219 23:43:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:27.219 23:43:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:27.219 23:43:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:27.219 23:43:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:27.219 23:43:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:27.219 23:43:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:27.219 23:43:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:27.219 23:43:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:27.219 23:43:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:27.479 23:43:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:27.479 23:43:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:27.479 "name": "Existed_Raid", 00:09:27.479 "uuid": "97632b6e-34c8-4d52-b562-48f6ea316807", 00:09:27.479 "strip_size_kb": 64, 00:09:27.479 "state": "configuring", 00:09:27.479 "raid_level": "concat", 00:09:27.479 "superblock": true, 00:09:27.479 "num_base_bdevs": 4, 00:09:27.479 "num_base_bdevs_discovered": 0, 00:09:27.479 "num_base_bdevs_operational": 4, 00:09:27.479 "base_bdevs_list": [ 00:09:27.479 { 00:09:27.479 "name": "BaseBdev1", 00:09:27.479 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:27.479 "is_configured": false, 00:09:27.479 "data_offset": 0, 00:09:27.479 "data_size": 0 00:09:27.479 }, 00:09:27.479 { 00:09:27.479 "name": "BaseBdev2", 00:09:27.479 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:27.479 "is_configured": false, 00:09:27.479 "data_offset": 0, 00:09:27.479 "data_size": 0 00:09:27.479 }, 00:09:27.479 { 00:09:27.479 "name": "BaseBdev3", 00:09:27.479 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:27.479 "is_configured": false, 00:09:27.479 "data_offset": 0, 00:09:27.479 "data_size": 0 00:09:27.479 }, 00:09:27.479 { 00:09:27.479 "name": "BaseBdev4", 00:09:27.479 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:27.479 "is_configured": false, 00:09:27.479 "data_offset": 0, 00:09:27.479 "data_size": 0 00:09:27.479 } 00:09:27.479 ] 00:09:27.479 }' 00:09:27.479 23:43:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:27.479 23:43:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:27.738 23:43:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:27.738 23:43:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:27.738 23:43:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:27.738 [2024-11-26 23:43:15.700245] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:27.738 [2024-11-26 23:43:15.700290] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:09:27.738 23:43:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:27.738 23:43:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:09:27.738 23:43:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:27.738 23:43:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:27.738 [2024-11-26 23:43:15.708260] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:27.738 [2024-11-26 23:43:15.708303] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:27.738 [2024-11-26 23:43:15.708311] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:27.738 [2024-11-26 23:43:15.708320] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:27.738 [2024-11-26 23:43:15.708326] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:27.738 [2024-11-26 23:43:15.708334] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:27.738 [2024-11-26 23:43:15.708340] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:09:27.738 [2024-11-26 23:43:15.708358] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:09:27.738 23:43:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:27.738 23:43:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:09:27.738 23:43:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:27.738 23:43:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:27.738 [2024-11-26 23:43:15.725082] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:27.738 BaseBdev1 00:09:27.738 23:43:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:27.738 23:43:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:09:27.738 23:43:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:09:27.738 23:43:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:27.738 23:43:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:09:27.738 23:43:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:27.738 23:43:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:27.738 23:43:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:27.738 23:43:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:27.738 23:43:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:27.738 23:43:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:27.738 23:43:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:09:27.738 23:43:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:27.738 23:43:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:27.739 [ 00:09:27.739 { 00:09:27.739 "name": "BaseBdev1", 00:09:27.739 "aliases": [ 00:09:27.739 "7302ff77-2d08-4754-a364-fd7e5fda2cd5" 00:09:27.739 ], 00:09:27.739 "product_name": "Malloc disk", 00:09:27.739 "block_size": 512, 00:09:27.739 "num_blocks": 65536, 00:09:27.739 "uuid": "7302ff77-2d08-4754-a364-fd7e5fda2cd5", 00:09:27.739 "assigned_rate_limits": { 00:09:27.739 "rw_ios_per_sec": 0, 00:09:27.739 "rw_mbytes_per_sec": 0, 00:09:27.739 "r_mbytes_per_sec": 0, 00:09:27.739 "w_mbytes_per_sec": 0 00:09:27.739 }, 00:09:27.739 "claimed": true, 00:09:27.739 "claim_type": "exclusive_write", 00:09:27.739 "zoned": false, 00:09:27.739 "supported_io_types": { 00:09:27.739 "read": true, 00:09:27.739 "write": true, 00:09:27.739 "unmap": true, 00:09:27.739 "flush": true, 00:09:27.739 "reset": true, 00:09:27.739 "nvme_admin": false, 00:09:27.739 "nvme_io": false, 00:09:27.739 "nvme_io_md": false, 00:09:27.739 "write_zeroes": true, 00:09:27.739 "zcopy": true, 00:09:27.739 "get_zone_info": false, 00:09:27.739 "zone_management": false, 00:09:27.739 "zone_append": false, 00:09:27.739 "compare": false, 00:09:27.739 "compare_and_write": false, 00:09:27.739 "abort": true, 00:09:27.739 "seek_hole": false, 00:09:27.739 "seek_data": false, 00:09:27.739 "copy": true, 00:09:27.739 "nvme_iov_md": false 00:09:27.739 }, 00:09:27.739 "memory_domains": [ 00:09:27.739 { 00:09:27.739 "dma_device_id": "system", 00:09:27.739 "dma_device_type": 1 00:09:27.739 }, 00:09:27.739 { 00:09:27.739 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:27.739 "dma_device_type": 2 00:09:27.739 } 00:09:27.739 ], 00:09:27.739 "driver_specific": {} 00:09:27.739 } 00:09:27.739 ] 00:09:27.739 23:43:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:27.739 23:43:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:09:27.739 23:43:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:09:27.739 23:43:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:27.739 23:43:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:27.739 23:43:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:27.739 23:43:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:27.739 23:43:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:27.739 23:43:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:27.739 23:43:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:27.739 23:43:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:27.739 23:43:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:27.739 23:43:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:27.739 23:43:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:27.739 23:43:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:27.739 23:43:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:27.739 23:43:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:27.739 23:43:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:27.739 "name": "Existed_Raid", 00:09:27.739 "uuid": "88643548-66f4-40d7-9806-9668c187f286", 00:09:27.739 "strip_size_kb": 64, 00:09:27.739 "state": "configuring", 00:09:27.739 "raid_level": "concat", 00:09:27.739 "superblock": true, 00:09:27.739 "num_base_bdevs": 4, 00:09:27.739 "num_base_bdevs_discovered": 1, 00:09:27.739 "num_base_bdevs_operational": 4, 00:09:27.739 "base_bdevs_list": [ 00:09:27.739 { 00:09:27.739 "name": "BaseBdev1", 00:09:27.739 "uuid": "7302ff77-2d08-4754-a364-fd7e5fda2cd5", 00:09:27.739 "is_configured": true, 00:09:27.739 "data_offset": 2048, 00:09:27.739 "data_size": 63488 00:09:27.739 }, 00:09:27.739 { 00:09:27.739 "name": "BaseBdev2", 00:09:27.739 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:27.739 "is_configured": false, 00:09:27.739 "data_offset": 0, 00:09:27.739 "data_size": 0 00:09:27.739 }, 00:09:27.739 { 00:09:27.739 "name": "BaseBdev3", 00:09:27.739 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:27.739 "is_configured": false, 00:09:27.739 "data_offset": 0, 00:09:27.739 "data_size": 0 00:09:27.739 }, 00:09:27.739 { 00:09:27.739 "name": "BaseBdev4", 00:09:27.739 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:27.739 "is_configured": false, 00:09:27.739 "data_offset": 0, 00:09:27.739 "data_size": 0 00:09:27.739 } 00:09:27.739 ] 00:09:27.739 }' 00:09:27.739 23:43:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:27.739 23:43:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:28.307 23:43:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:28.307 23:43:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:28.307 23:43:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:28.307 [2024-11-26 23:43:16.196370] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:28.307 [2024-11-26 23:43:16.196423] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:09:28.307 23:43:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:28.307 23:43:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:09:28.307 23:43:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:28.307 23:43:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:28.307 [2024-11-26 23:43:16.208394] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:28.307 [2024-11-26 23:43:16.210142] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:28.307 [2024-11-26 23:43:16.210197] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:28.307 [2024-11-26 23:43:16.210207] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:28.307 [2024-11-26 23:43:16.210216] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:28.307 [2024-11-26 23:43:16.210222] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:09:28.307 [2024-11-26 23:43:16.210230] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:09:28.307 23:43:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:28.307 23:43:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:09:28.307 23:43:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:28.307 23:43:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:09:28.307 23:43:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:28.307 23:43:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:28.307 23:43:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:28.307 23:43:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:28.307 23:43:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:28.307 23:43:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:28.307 23:43:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:28.307 23:43:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:28.307 23:43:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:28.307 23:43:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:28.307 23:43:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:28.307 23:43:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:28.307 23:43:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:28.307 23:43:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:28.307 23:43:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:28.307 "name": "Existed_Raid", 00:09:28.307 "uuid": "0c863eff-2e34-4d65-a4e8-791218b1bee9", 00:09:28.307 "strip_size_kb": 64, 00:09:28.307 "state": "configuring", 00:09:28.307 "raid_level": "concat", 00:09:28.307 "superblock": true, 00:09:28.307 "num_base_bdevs": 4, 00:09:28.307 "num_base_bdevs_discovered": 1, 00:09:28.307 "num_base_bdevs_operational": 4, 00:09:28.307 "base_bdevs_list": [ 00:09:28.307 { 00:09:28.307 "name": "BaseBdev1", 00:09:28.307 "uuid": "7302ff77-2d08-4754-a364-fd7e5fda2cd5", 00:09:28.307 "is_configured": true, 00:09:28.307 "data_offset": 2048, 00:09:28.307 "data_size": 63488 00:09:28.307 }, 00:09:28.307 { 00:09:28.307 "name": "BaseBdev2", 00:09:28.307 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:28.307 "is_configured": false, 00:09:28.307 "data_offset": 0, 00:09:28.307 "data_size": 0 00:09:28.307 }, 00:09:28.307 { 00:09:28.307 "name": "BaseBdev3", 00:09:28.307 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:28.307 "is_configured": false, 00:09:28.307 "data_offset": 0, 00:09:28.307 "data_size": 0 00:09:28.307 }, 00:09:28.307 { 00:09:28.307 "name": "BaseBdev4", 00:09:28.307 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:28.307 "is_configured": false, 00:09:28.307 "data_offset": 0, 00:09:28.307 "data_size": 0 00:09:28.307 } 00:09:28.307 ] 00:09:28.307 }' 00:09:28.307 23:43:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:28.307 23:43:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:28.567 23:43:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:09:28.567 23:43:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:28.567 23:43:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:28.567 [2024-11-26 23:43:16.638538] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:28.567 BaseBdev2 00:09:28.567 23:43:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:28.567 23:43:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:09:28.567 23:43:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:09:28.567 23:43:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:28.567 23:43:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:09:28.567 23:43:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:28.567 23:43:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:28.567 23:43:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:28.567 23:43:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:28.567 23:43:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:28.567 23:43:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:28.567 23:43:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:09:28.567 23:43:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:28.567 23:43:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:28.567 [ 00:09:28.567 { 00:09:28.567 "name": "BaseBdev2", 00:09:28.567 "aliases": [ 00:09:28.567 "96ab3edd-a4bb-433f-87cd-ffdb7d84ab8a" 00:09:28.567 ], 00:09:28.567 "product_name": "Malloc disk", 00:09:28.567 "block_size": 512, 00:09:28.567 "num_blocks": 65536, 00:09:28.567 "uuid": "96ab3edd-a4bb-433f-87cd-ffdb7d84ab8a", 00:09:28.567 "assigned_rate_limits": { 00:09:28.567 "rw_ios_per_sec": 0, 00:09:28.567 "rw_mbytes_per_sec": 0, 00:09:28.567 "r_mbytes_per_sec": 0, 00:09:28.567 "w_mbytes_per_sec": 0 00:09:28.567 }, 00:09:28.567 "claimed": true, 00:09:28.567 "claim_type": "exclusive_write", 00:09:28.567 "zoned": false, 00:09:28.567 "supported_io_types": { 00:09:28.567 "read": true, 00:09:28.567 "write": true, 00:09:28.567 "unmap": true, 00:09:28.567 "flush": true, 00:09:28.567 "reset": true, 00:09:28.567 "nvme_admin": false, 00:09:28.567 "nvme_io": false, 00:09:28.567 "nvme_io_md": false, 00:09:28.567 "write_zeroes": true, 00:09:28.567 "zcopy": true, 00:09:28.567 "get_zone_info": false, 00:09:28.567 "zone_management": false, 00:09:28.567 "zone_append": false, 00:09:28.567 "compare": false, 00:09:28.567 "compare_and_write": false, 00:09:28.567 "abort": true, 00:09:28.567 "seek_hole": false, 00:09:28.567 "seek_data": false, 00:09:28.567 "copy": true, 00:09:28.567 "nvme_iov_md": false 00:09:28.567 }, 00:09:28.567 "memory_domains": [ 00:09:28.567 { 00:09:28.567 "dma_device_id": "system", 00:09:28.567 "dma_device_type": 1 00:09:28.567 }, 00:09:28.567 { 00:09:28.567 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:28.567 "dma_device_type": 2 00:09:28.567 } 00:09:28.567 ], 00:09:28.567 "driver_specific": {} 00:09:28.567 } 00:09:28.567 ] 00:09:28.567 23:43:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:28.567 23:43:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:09:28.567 23:43:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:09:28.567 23:43:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:28.567 23:43:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:09:28.567 23:43:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:28.567 23:43:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:28.567 23:43:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:28.567 23:43:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:28.567 23:43:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:28.567 23:43:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:28.567 23:43:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:28.567 23:43:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:28.567 23:43:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:28.567 23:43:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:28.567 23:43:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:28.567 23:43:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:28.567 23:43:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:28.826 23:43:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:28.826 23:43:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:28.826 "name": "Existed_Raid", 00:09:28.826 "uuid": "0c863eff-2e34-4d65-a4e8-791218b1bee9", 00:09:28.826 "strip_size_kb": 64, 00:09:28.826 "state": "configuring", 00:09:28.826 "raid_level": "concat", 00:09:28.826 "superblock": true, 00:09:28.826 "num_base_bdevs": 4, 00:09:28.826 "num_base_bdevs_discovered": 2, 00:09:28.826 "num_base_bdevs_operational": 4, 00:09:28.826 "base_bdevs_list": [ 00:09:28.826 { 00:09:28.826 "name": "BaseBdev1", 00:09:28.826 "uuid": "7302ff77-2d08-4754-a364-fd7e5fda2cd5", 00:09:28.826 "is_configured": true, 00:09:28.826 "data_offset": 2048, 00:09:28.826 "data_size": 63488 00:09:28.826 }, 00:09:28.827 { 00:09:28.827 "name": "BaseBdev2", 00:09:28.827 "uuid": "96ab3edd-a4bb-433f-87cd-ffdb7d84ab8a", 00:09:28.827 "is_configured": true, 00:09:28.827 "data_offset": 2048, 00:09:28.827 "data_size": 63488 00:09:28.827 }, 00:09:28.827 { 00:09:28.827 "name": "BaseBdev3", 00:09:28.827 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:28.827 "is_configured": false, 00:09:28.827 "data_offset": 0, 00:09:28.827 "data_size": 0 00:09:28.827 }, 00:09:28.827 { 00:09:28.827 "name": "BaseBdev4", 00:09:28.827 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:28.827 "is_configured": false, 00:09:28.827 "data_offset": 0, 00:09:28.827 "data_size": 0 00:09:28.827 } 00:09:28.827 ] 00:09:28.827 }' 00:09:28.827 23:43:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:28.827 23:43:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:29.086 23:43:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:09:29.086 23:43:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:29.086 23:43:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:29.086 [2024-11-26 23:43:17.084291] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:29.086 BaseBdev3 00:09:29.086 23:43:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:29.086 23:43:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:09:29.086 23:43:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:09:29.086 23:43:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:29.086 23:43:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:09:29.086 23:43:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:29.086 23:43:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:29.086 23:43:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:29.086 23:43:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:29.086 23:43:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:29.086 23:43:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:29.086 23:43:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:09:29.086 23:43:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:29.086 23:43:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:29.086 [ 00:09:29.086 { 00:09:29.086 "name": "BaseBdev3", 00:09:29.086 "aliases": [ 00:09:29.086 "ae634f84-888f-4960-8386-e9abc44a9cb4" 00:09:29.086 ], 00:09:29.086 "product_name": "Malloc disk", 00:09:29.086 "block_size": 512, 00:09:29.086 "num_blocks": 65536, 00:09:29.086 "uuid": "ae634f84-888f-4960-8386-e9abc44a9cb4", 00:09:29.086 "assigned_rate_limits": { 00:09:29.086 "rw_ios_per_sec": 0, 00:09:29.086 "rw_mbytes_per_sec": 0, 00:09:29.086 "r_mbytes_per_sec": 0, 00:09:29.086 "w_mbytes_per_sec": 0 00:09:29.086 }, 00:09:29.086 "claimed": true, 00:09:29.086 "claim_type": "exclusive_write", 00:09:29.086 "zoned": false, 00:09:29.086 "supported_io_types": { 00:09:29.086 "read": true, 00:09:29.086 "write": true, 00:09:29.086 "unmap": true, 00:09:29.086 "flush": true, 00:09:29.086 "reset": true, 00:09:29.086 "nvme_admin": false, 00:09:29.086 "nvme_io": false, 00:09:29.086 "nvme_io_md": false, 00:09:29.086 "write_zeroes": true, 00:09:29.086 "zcopy": true, 00:09:29.086 "get_zone_info": false, 00:09:29.086 "zone_management": false, 00:09:29.086 "zone_append": false, 00:09:29.086 "compare": false, 00:09:29.086 "compare_and_write": false, 00:09:29.086 "abort": true, 00:09:29.086 "seek_hole": false, 00:09:29.086 "seek_data": false, 00:09:29.086 "copy": true, 00:09:29.086 "nvme_iov_md": false 00:09:29.086 }, 00:09:29.086 "memory_domains": [ 00:09:29.086 { 00:09:29.086 "dma_device_id": "system", 00:09:29.086 "dma_device_type": 1 00:09:29.086 }, 00:09:29.086 { 00:09:29.086 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:29.086 "dma_device_type": 2 00:09:29.086 } 00:09:29.086 ], 00:09:29.086 "driver_specific": {} 00:09:29.086 } 00:09:29.086 ] 00:09:29.086 23:43:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:29.086 23:43:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:09:29.086 23:43:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:09:29.086 23:43:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:29.086 23:43:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:09:29.086 23:43:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:29.086 23:43:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:29.087 23:43:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:29.087 23:43:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:29.087 23:43:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:29.087 23:43:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:29.087 23:43:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:29.087 23:43:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:29.087 23:43:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:29.087 23:43:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:29.087 23:43:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:29.087 23:43:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:29.087 23:43:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:29.087 23:43:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:29.087 23:43:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:29.087 "name": "Existed_Raid", 00:09:29.087 "uuid": "0c863eff-2e34-4d65-a4e8-791218b1bee9", 00:09:29.087 "strip_size_kb": 64, 00:09:29.087 "state": "configuring", 00:09:29.087 "raid_level": "concat", 00:09:29.087 "superblock": true, 00:09:29.087 "num_base_bdevs": 4, 00:09:29.087 "num_base_bdevs_discovered": 3, 00:09:29.087 "num_base_bdevs_operational": 4, 00:09:29.087 "base_bdevs_list": [ 00:09:29.087 { 00:09:29.087 "name": "BaseBdev1", 00:09:29.087 "uuid": "7302ff77-2d08-4754-a364-fd7e5fda2cd5", 00:09:29.087 "is_configured": true, 00:09:29.087 "data_offset": 2048, 00:09:29.087 "data_size": 63488 00:09:29.087 }, 00:09:29.087 { 00:09:29.087 "name": "BaseBdev2", 00:09:29.087 "uuid": "96ab3edd-a4bb-433f-87cd-ffdb7d84ab8a", 00:09:29.087 "is_configured": true, 00:09:29.087 "data_offset": 2048, 00:09:29.087 "data_size": 63488 00:09:29.087 }, 00:09:29.087 { 00:09:29.087 "name": "BaseBdev3", 00:09:29.087 "uuid": "ae634f84-888f-4960-8386-e9abc44a9cb4", 00:09:29.087 "is_configured": true, 00:09:29.087 "data_offset": 2048, 00:09:29.087 "data_size": 63488 00:09:29.087 }, 00:09:29.087 { 00:09:29.087 "name": "BaseBdev4", 00:09:29.087 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:29.087 "is_configured": false, 00:09:29.087 "data_offset": 0, 00:09:29.087 "data_size": 0 00:09:29.087 } 00:09:29.087 ] 00:09:29.087 }' 00:09:29.087 23:43:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:29.087 23:43:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:29.652 23:43:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:09:29.652 23:43:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:29.652 23:43:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:29.652 [2024-11-26 23:43:17.542636] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:09:29.652 [2024-11-26 23:43:17.542835] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:09:29.652 [2024-11-26 23:43:17.542854] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:09:29.652 [2024-11-26 23:43:17.543145] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:09:29.652 BaseBdev4 00:09:29.652 [2024-11-26 23:43:17.543281] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:09:29.652 [2024-11-26 23:43:17.543292] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:09:29.652 [2024-11-26 23:43:17.543421] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:29.652 23:43:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:29.652 23:43:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev4 00:09:29.652 23:43:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:09:29.652 23:43:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:29.652 23:43:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:09:29.652 23:43:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:29.652 23:43:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:29.652 23:43:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:29.652 23:43:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:29.652 23:43:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:29.652 23:43:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:29.652 23:43:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:09:29.652 23:43:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:29.652 23:43:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:29.652 [ 00:09:29.652 { 00:09:29.652 "name": "BaseBdev4", 00:09:29.652 "aliases": [ 00:09:29.652 "7b1de7b7-187f-4248-9762-1b667cec7124" 00:09:29.652 ], 00:09:29.652 "product_name": "Malloc disk", 00:09:29.652 "block_size": 512, 00:09:29.652 "num_blocks": 65536, 00:09:29.652 "uuid": "7b1de7b7-187f-4248-9762-1b667cec7124", 00:09:29.652 "assigned_rate_limits": { 00:09:29.652 "rw_ios_per_sec": 0, 00:09:29.652 "rw_mbytes_per_sec": 0, 00:09:29.652 "r_mbytes_per_sec": 0, 00:09:29.652 "w_mbytes_per_sec": 0 00:09:29.652 }, 00:09:29.652 "claimed": true, 00:09:29.652 "claim_type": "exclusive_write", 00:09:29.652 "zoned": false, 00:09:29.652 "supported_io_types": { 00:09:29.652 "read": true, 00:09:29.652 "write": true, 00:09:29.652 "unmap": true, 00:09:29.652 "flush": true, 00:09:29.652 "reset": true, 00:09:29.652 "nvme_admin": false, 00:09:29.652 "nvme_io": false, 00:09:29.652 "nvme_io_md": false, 00:09:29.652 "write_zeroes": true, 00:09:29.652 "zcopy": true, 00:09:29.652 "get_zone_info": false, 00:09:29.652 "zone_management": false, 00:09:29.652 "zone_append": false, 00:09:29.652 "compare": false, 00:09:29.652 "compare_and_write": false, 00:09:29.652 "abort": true, 00:09:29.652 "seek_hole": false, 00:09:29.652 "seek_data": false, 00:09:29.652 "copy": true, 00:09:29.652 "nvme_iov_md": false 00:09:29.652 }, 00:09:29.652 "memory_domains": [ 00:09:29.652 { 00:09:29.652 "dma_device_id": "system", 00:09:29.652 "dma_device_type": 1 00:09:29.652 }, 00:09:29.652 { 00:09:29.652 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:29.652 "dma_device_type": 2 00:09:29.652 } 00:09:29.652 ], 00:09:29.652 "driver_specific": {} 00:09:29.652 } 00:09:29.652 ] 00:09:29.652 23:43:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:29.652 23:43:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:09:29.653 23:43:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:09:29.653 23:43:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:29.653 23:43:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online concat 64 4 00:09:29.653 23:43:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:29.653 23:43:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:29.653 23:43:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:29.653 23:43:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:29.653 23:43:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:29.653 23:43:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:29.653 23:43:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:29.653 23:43:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:29.653 23:43:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:29.653 23:43:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:29.653 23:43:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:29.653 23:43:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:29.653 23:43:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:29.653 23:43:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:29.653 23:43:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:29.653 "name": "Existed_Raid", 00:09:29.653 "uuid": "0c863eff-2e34-4d65-a4e8-791218b1bee9", 00:09:29.653 "strip_size_kb": 64, 00:09:29.653 "state": "online", 00:09:29.653 "raid_level": "concat", 00:09:29.653 "superblock": true, 00:09:29.653 "num_base_bdevs": 4, 00:09:29.653 "num_base_bdevs_discovered": 4, 00:09:29.653 "num_base_bdevs_operational": 4, 00:09:29.653 "base_bdevs_list": [ 00:09:29.653 { 00:09:29.653 "name": "BaseBdev1", 00:09:29.653 "uuid": "7302ff77-2d08-4754-a364-fd7e5fda2cd5", 00:09:29.653 "is_configured": true, 00:09:29.653 "data_offset": 2048, 00:09:29.653 "data_size": 63488 00:09:29.653 }, 00:09:29.653 { 00:09:29.653 "name": "BaseBdev2", 00:09:29.653 "uuid": "96ab3edd-a4bb-433f-87cd-ffdb7d84ab8a", 00:09:29.653 "is_configured": true, 00:09:29.653 "data_offset": 2048, 00:09:29.653 "data_size": 63488 00:09:29.653 }, 00:09:29.653 { 00:09:29.653 "name": "BaseBdev3", 00:09:29.653 "uuid": "ae634f84-888f-4960-8386-e9abc44a9cb4", 00:09:29.653 "is_configured": true, 00:09:29.653 "data_offset": 2048, 00:09:29.653 "data_size": 63488 00:09:29.653 }, 00:09:29.653 { 00:09:29.653 "name": "BaseBdev4", 00:09:29.653 "uuid": "7b1de7b7-187f-4248-9762-1b667cec7124", 00:09:29.653 "is_configured": true, 00:09:29.653 "data_offset": 2048, 00:09:29.653 "data_size": 63488 00:09:29.653 } 00:09:29.653 ] 00:09:29.653 }' 00:09:29.653 23:43:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:29.653 23:43:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:29.911 23:43:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:09:29.911 23:43:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:09:29.911 23:43:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:29.911 23:43:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:29.911 23:43:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:09:29.911 23:43:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:29.911 23:43:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:29.912 23:43:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:09:29.912 23:43:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:29.912 23:43:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:29.912 [2024-11-26 23:43:18.006283] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:29.912 23:43:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:29.912 23:43:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:29.912 "name": "Existed_Raid", 00:09:29.912 "aliases": [ 00:09:29.912 "0c863eff-2e34-4d65-a4e8-791218b1bee9" 00:09:29.912 ], 00:09:29.912 "product_name": "Raid Volume", 00:09:29.912 "block_size": 512, 00:09:29.912 "num_blocks": 253952, 00:09:29.912 "uuid": "0c863eff-2e34-4d65-a4e8-791218b1bee9", 00:09:29.912 "assigned_rate_limits": { 00:09:29.912 "rw_ios_per_sec": 0, 00:09:29.912 "rw_mbytes_per_sec": 0, 00:09:29.912 "r_mbytes_per_sec": 0, 00:09:29.912 "w_mbytes_per_sec": 0 00:09:29.912 }, 00:09:29.912 "claimed": false, 00:09:29.912 "zoned": false, 00:09:29.912 "supported_io_types": { 00:09:29.912 "read": true, 00:09:29.912 "write": true, 00:09:29.912 "unmap": true, 00:09:29.912 "flush": true, 00:09:29.912 "reset": true, 00:09:29.912 "nvme_admin": false, 00:09:29.912 "nvme_io": false, 00:09:29.912 "nvme_io_md": false, 00:09:29.912 "write_zeroes": true, 00:09:29.912 "zcopy": false, 00:09:29.912 "get_zone_info": false, 00:09:29.912 "zone_management": false, 00:09:29.912 "zone_append": false, 00:09:29.912 "compare": false, 00:09:29.912 "compare_and_write": false, 00:09:29.912 "abort": false, 00:09:29.912 "seek_hole": false, 00:09:29.912 "seek_data": false, 00:09:29.912 "copy": false, 00:09:29.912 "nvme_iov_md": false 00:09:29.912 }, 00:09:29.912 "memory_domains": [ 00:09:29.912 { 00:09:29.912 "dma_device_id": "system", 00:09:29.912 "dma_device_type": 1 00:09:29.912 }, 00:09:29.912 { 00:09:29.912 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:29.912 "dma_device_type": 2 00:09:29.912 }, 00:09:29.912 { 00:09:29.912 "dma_device_id": "system", 00:09:29.912 "dma_device_type": 1 00:09:29.912 }, 00:09:29.912 { 00:09:29.912 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:29.912 "dma_device_type": 2 00:09:29.912 }, 00:09:29.912 { 00:09:29.912 "dma_device_id": "system", 00:09:29.912 "dma_device_type": 1 00:09:29.912 }, 00:09:29.912 { 00:09:29.912 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:29.912 "dma_device_type": 2 00:09:29.912 }, 00:09:29.912 { 00:09:29.912 "dma_device_id": "system", 00:09:29.912 "dma_device_type": 1 00:09:29.912 }, 00:09:29.912 { 00:09:29.912 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:29.912 "dma_device_type": 2 00:09:29.912 } 00:09:29.912 ], 00:09:29.912 "driver_specific": { 00:09:29.912 "raid": { 00:09:29.912 "uuid": "0c863eff-2e34-4d65-a4e8-791218b1bee9", 00:09:29.912 "strip_size_kb": 64, 00:09:29.912 "state": "online", 00:09:29.912 "raid_level": "concat", 00:09:29.912 "superblock": true, 00:09:29.912 "num_base_bdevs": 4, 00:09:29.912 "num_base_bdevs_discovered": 4, 00:09:29.912 "num_base_bdevs_operational": 4, 00:09:29.912 "base_bdevs_list": [ 00:09:29.912 { 00:09:29.912 "name": "BaseBdev1", 00:09:29.912 "uuid": "7302ff77-2d08-4754-a364-fd7e5fda2cd5", 00:09:29.912 "is_configured": true, 00:09:29.912 "data_offset": 2048, 00:09:29.912 "data_size": 63488 00:09:29.912 }, 00:09:29.912 { 00:09:29.912 "name": "BaseBdev2", 00:09:29.912 "uuid": "96ab3edd-a4bb-433f-87cd-ffdb7d84ab8a", 00:09:29.912 "is_configured": true, 00:09:29.912 "data_offset": 2048, 00:09:29.912 "data_size": 63488 00:09:29.912 }, 00:09:29.912 { 00:09:29.912 "name": "BaseBdev3", 00:09:29.912 "uuid": "ae634f84-888f-4960-8386-e9abc44a9cb4", 00:09:29.912 "is_configured": true, 00:09:29.912 "data_offset": 2048, 00:09:29.912 "data_size": 63488 00:09:29.912 }, 00:09:29.912 { 00:09:29.912 "name": "BaseBdev4", 00:09:29.912 "uuid": "7b1de7b7-187f-4248-9762-1b667cec7124", 00:09:29.912 "is_configured": true, 00:09:29.912 "data_offset": 2048, 00:09:29.912 "data_size": 63488 00:09:29.912 } 00:09:29.912 ] 00:09:29.912 } 00:09:29.912 } 00:09:29.912 }' 00:09:29.912 23:43:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:30.171 23:43:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:09:30.171 BaseBdev2 00:09:30.171 BaseBdev3 00:09:30.171 BaseBdev4' 00:09:30.171 23:43:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:30.171 23:43:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:30.171 23:43:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:30.171 23:43:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:09:30.171 23:43:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:30.171 23:43:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:30.171 23:43:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:30.171 23:43:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:30.171 23:43:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:30.171 23:43:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:30.171 23:43:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:30.171 23:43:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:09:30.171 23:43:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:30.171 23:43:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:30.171 23:43:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:30.171 23:43:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:30.171 23:43:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:30.171 23:43:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:30.171 23:43:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:30.171 23:43:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:09:30.171 23:43:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:30.171 23:43:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:30.171 23:43:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:30.171 23:43:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:30.171 23:43:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:30.171 23:43:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:30.171 23:43:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:30.171 23:43:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:30.171 23:43:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:09:30.171 23:43:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:30.171 23:43:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:30.171 23:43:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:30.171 23:43:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:30.171 23:43:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:30.172 23:43:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:09:30.172 23:43:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:30.172 23:43:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:30.172 [2024-11-26 23:43:18.281512] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:09:30.172 [2024-11-26 23:43:18.281545] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:30.172 [2024-11-26 23:43:18.281605] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:30.172 23:43:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:30.172 23:43:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:09:30.172 23:43:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy concat 00:09:30.172 23:43:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:09:30.172 23:43:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # return 1 00:09:30.172 23:43:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:09:30.172 23:43:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline concat 64 3 00:09:30.172 23:43:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:30.172 23:43:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:09:30.172 23:43:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:30.172 23:43:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:30.172 23:43:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:30.172 23:43:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:30.172 23:43:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:30.172 23:43:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:30.172 23:43:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:30.430 23:43:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:30.430 23:43:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:30.430 23:43:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:30.431 23:43:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:30.431 23:43:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:30.431 23:43:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:30.431 "name": "Existed_Raid", 00:09:30.431 "uuid": "0c863eff-2e34-4d65-a4e8-791218b1bee9", 00:09:30.431 "strip_size_kb": 64, 00:09:30.431 "state": "offline", 00:09:30.431 "raid_level": "concat", 00:09:30.431 "superblock": true, 00:09:30.431 "num_base_bdevs": 4, 00:09:30.431 "num_base_bdevs_discovered": 3, 00:09:30.431 "num_base_bdevs_operational": 3, 00:09:30.431 "base_bdevs_list": [ 00:09:30.431 { 00:09:30.431 "name": null, 00:09:30.431 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:30.431 "is_configured": false, 00:09:30.431 "data_offset": 0, 00:09:30.431 "data_size": 63488 00:09:30.431 }, 00:09:30.431 { 00:09:30.431 "name": "BaseBdev2", 00:09:30.431 "uuid": "96ab3edd-a4bb-433f-87cd-ffdb7d84ab8a", 00:09:30.431 "is_configured": true, 00:09:30.431 "data_offset": 2048, 00:09:30.431 "data_size": 63488 00:09:30.431 }, 00:09:30.431 { 00:09:30.431 "name": "BaseBdev3", 00:09:30.431 "uuid": "ae634f84-888f-4960-8386-e9abc44a9cb4", 00:09:30.431 "is_configured": true, 00:09:30.431 "data_offset": 2048, 00:09:30.431 "data_size": 63488 00:09:30.431 }, 00:09:30.431 { 00:09:30.431 "name": "BaseBdev4", 00:09:30.431 "uuid": "7b1de7b7-187f-4248-9762-1b667cec7124", 00:09:30.431 "is_configured": true, 00:09:30.431 "data_offset": 2048, 00:09:30.431 "data_size": 63488 00:09:30.431 } 00:09:30.431 ] 00:09:30.431 }' 00:09:30.431 23:43:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:30.431 23:43:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:30.689 23:43:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:09:30.689 23:43:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:30.689 23:43:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:09:30.689 23:43:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:30.689 23:43:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:30.689 23:43:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:30.689 23:43:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:30.689 23:43:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:09:30.689 23:43:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:09:30.689 23:43:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:09:30.689 23:43:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:30.689 23:43:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:30.689 [2024-11-26 23:43:18.715920] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:09:30.689 23:43:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:30.689 23:43:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:09:30.689 23:43:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:30.689 23:43:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:30.689 23:43:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:30.689 23:43:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:30.689 23:43:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:09:30.689 23:43:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:30.689 23:43:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:09:30.689 23:43:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:09:30.689 23:43:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:09:30.689 23:43:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:30.689 23:43:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:30.689 [2024-11-26 23:43:18.783026] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:09:30.689 23:43:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:30.689 23:43:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:09:30.689 23:43:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:30.689 23:43:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:09:30.689 23:43:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:30.689 23:43:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:30.689 23:43:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:30.950 23:43:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:30.950 23:43:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:09:30.950 23:43:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:09:30.950 23:43:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev4 00:09:30.950 23:43:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:30.950 23:43:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:30.950 [2024-11-26 23:43:18.841838] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev4 00:09:30.950 [2024-11-26 23:43:18.841884] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:09:30.950 23:43:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:30.950 23:43:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:09:30.950 23:43:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:30.950 23:43:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:30.950 23:43:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:30.950 23:43:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:09:30.950 23:43:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:30.950 23:43:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:30.950 23:43:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:09:30.950 23:43:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:09:30.950 23:43:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 4 -gt 2 ']' 00:09:30.950 23:43:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:09:30.950 23:43:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:30.950 23:43:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:09:30.950 23:43:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:30.950 23:43:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:30.950 BaseBdev2 00:09:30.950 23:43:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:30.950 23:43:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:09:30.950 23:43:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:09:30.950 23:43:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:30.950 23:43:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:09:30.950 23:43:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:30.950 23:43:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:30.950 23:43:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:30.950 23:43:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:30.950 23:43:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:30.950 23:43:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:30.950 23:43:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:09:30.950 23:43:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:30.950 23:43:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:30.950 [ 00:09:30.950 { 00:09:30.950 "name": "BaseBdev2", 00:09:30.950 "aliases": [ 00:09:30.950 "b0ba44a4-bafa-4b62-9a9b-bc3538366303" 00:09:30.950 ], 00:09:30.950 "product_name": "Malloc disk", 00:09:30.950 "block_size": 512, 00:09:30.950 "num_blocks": 65536, 00:09:30.950 "uuid": "b0ba44a4-bafa-4b62-9a9b-bc3538366303", 00:09:30.950 "assigned_rate_limits": { 00:09:30.950 "rw_ios_per_sec": 0, 00:09:30.950 "rw_mbytes_per_sec": 0, 00:09:30.950 "r_mbytes_per_sec": 0, 00:09:30.950 "w_mbytes_per_sec": 0 00:09:30.950 }, 00:09:30.950 "claimed": false, 00:09:30.950 "zoned": false, 00:09:30.950 "supported_io_types": { 00:09:30.950 "read": true, 00:09:30.950 "write": true, 00:09:30.950 "unmap": true, 00:09:30.950 "flush": true, 00:09:30.950 "reset": true, 00:09:30.950 "nvme_admin": false, 00:09:30.950 "nvme_io": false, 00:09:30.950 "nvme_io_md": false, 00:09:30.950 "write_zeroes": true, 00:09:30.950 "zcopy": true, 00:09:30.950 "get_zone_info": false, 00:09:30.950 "zone_management": false, 00:09:30.950 "zone_append": false, 00:09:30.950 "compare": false, 00:09:30.950 "compare_and_write": false, 00:09:30.950 "abort": true, 00:09:30.950 "seek_hole": false, 00:09:30.950 "seek_data": false, 00:09:30.950 "copy": true, 00:09:30.950 "nvme_iov_md": false 00:09:30.950 }, 00:09:30.950 "memory_domains": [ 00:09:30.950 { 00:09:30.950 "dma_device_id": "system", 00:09:30.950 "dma_device_type": 1 00:09:30.950 }, 00:09:30.950 { 00:09:30.950 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:30.950 "dma_device_type": 2 00:09:30.950 } 00:09:30.950 ], 00:09:30.950 "driver_specific": {} 00:09:30.950 } 00:09:30.950 ] 00:09:30.950 23:43:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:30.950 23:43:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:09:30.950 23:43:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:09:30.950 23:43:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:30.950 23:43:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:09:30.950 23:43:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:30.950 23:43:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:30.950 BaseBdev3 00:09:30.950 23:43:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:30.950 23:43:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:09:30.950 23:43:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:09:30.950 23:43:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:30.950 23:43:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:09:30.951 23:43:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:30.951 23:43:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:30.951 23:43:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:30.951 23:43:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:30.951 23:43:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:30.951 23:43:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:30.951 23:43:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:09:30.951 23:43:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:30.951 23:43:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:30.951 [ 00:09:30.951 { 00:09:30.951 "name": "BaseBdev3", 00:09:30.951 "aliases": [ 00:09:30.951 "df1c05d3-aec8-490a-aba7-ebac08556835" 00:09:30.951 ], 00:09:30.951 "product_name": "Malloc disk", 00:09:30.951 "block_size": 512, 00:09:30.951 "num_blocks": 65536, 00:09:30.951 "uuid": "df1c05d3-aec8-490a-aba7-ebac08556835", 00:09:30.951 "assigned_rate_limits": { 00:09:30.951 "rw_ios_per_sec": 0, 00:09:30.951 "rw_mbytes_per_sec": 0, 00:09:30.951 "r_mbytes_per_sec": 0, 00:09:30.951 "w_mbytes_per_sec": 0 00:09:30.951 }, 00:09:30.951 "claimed": false, 00:09:30.951 "zoned": false, 00:09:30.951 "supported_io_types": { 00:09:30.951 "read": true, 00:09:30.951 "write": true, 00:09:30.951 "unmap": true, 00:09:30.951 "flush": true, 00:09:30.951 "reset": true, 00:09:30.951 "nvme_admin": false, 00:09:30.951 "nvme_io": false, 00:09:30.951 "nvme_io_md": false, 00:09:30.951 "write_zeroes": true, 00:09:30.951 "zcopy": true, 00:09:30.951 "get_zone_info": false, 00:09:30.951 "zone_management": false, 00:09:30.951 "zone_append": false, 00:09:30.951 "compare": false, 00:09:30.951 "compare_and_write": false, 00:09:30.951 "abort": true, 00:09:30.951 "seek_hole": false, 00:09:30.951 "seek_data": false, 00:09:30.951 "copy": true, 00:09:30.951 "nvme_iov_md": false 00:09:30.951 }, 00:09:30.951 "memory_domains": [ 00:09:30.951 { 00:09:30.951 "dma_device_id": "system", 00:09:30.951 "dma_device_type": 1 00:09:30.951 }, 00:09:30.951 { 00:09:30.951 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:30.951 "dma_device_type": 2 00:09:30.951 } 00:09:30.951 ], 00:09:30.951 "driver_specific": {} 00:09:30.951 } 00:09:30.951 ] 00:09:30.951 23:43:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:30.951 23:43:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:09:30.951 23:43:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:09:30.951 23:43:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:30.951 23:43:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:09:30.951 23:43:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:30.951 23:43:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:30.951 BaseBdev4 00:09:30.951 23:43:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:30.951 23:43:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev4 00:09:30.951 23:43:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:09:30.951 23:43:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:30.951 23:43:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:09:30.951 23:43:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:30.951 23:43:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:30.951 23:43:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:30.951 23:43:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:30.951 23:43:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:30.951 23:43:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:30.951 23:43:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:09:30.951 23:43:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:30.951 23:43:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:30.951 [ 00:09:30.951 { 00:09:30.951 "name": "BaseBdev4", 00:09:30.951 "aliases": [ 00:09:30.951 "a244b1a4-8a1b-47d0-b1b5-74bd778ade01" 00:09:30.951 ], 00:09:30.951 "product_name": "Malloc disk", 00:09:30.951 "block_size": 512, 00:09:30.951 "num_blocks": 65536, 00:09:30.951 "uuid": "a244b1a4-8a1b-47d0-b1b5-74bd778ade01", 00:09:30.951 "assigned_rate_limits": { 00:09:30.951 "rw_ios_per_sec": 0, 00:09:30.951 "rw_mbytes_per_sec": 0, 00:09:30.951 "r_mbytes_per_sec": 0, 00:09:30.951 "w_mbytes_per_sec": 0 00:09:30.951 }, 00:09:30.951 "claimed": false, 00:09:30.951 "zoned": false, 00:09:30.951 "supported_io_types": { 00:09:30.951 "read": true, 00:09:30.951 "write": true, 00:09:30.951 "unmap": true, 00:09:30.951 "flush": true, 00:09:30.951 "reset": true, 00:09:30.951 "nvme_admin": false, 00:09:30.951 "nvme_io": false, 00:09:30.951 "nvme_io_md": false, 00:09:30.951 "write_zeroes": true, 00:09:30.951 "zcopy": true, 00:09:30.951 "get_zone_info": false, 00:09:30.951 "zone_management": false, 00:09:30.951 "zone_append": false, 00:09:30.951 "compare": false, 00:09:30.951 "compare_and_write": false, 00:09:30.951 "abort": true, 00:09:30.951 "seek_hole": false, 00:09:30.951 "seek_data": false, 00:09:30.951 "copy": true, 00:09:30.951 "nvme_iov_md": false 00:09:30.951 }, 00:09:30.951 "memory_domains": [ 00:09:30.951 { 00:09:30.951 "dma_device_id": "system", 00:09:30.951 "dma_device_type": 1 00:09:30.951 }, 00:09:30.951 { 00:09:30.951 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:30.951 "dma_device_type": 2 00:09:30.951 } 00:09:30.951 ], 00:09:30.951 "driver_specific": {} 00:09:30.951 } 00:09:30.951 ] 00:09:30.951 23:43:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:30.951 23:43:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:09:30.951 23:43:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:09:30.951 23:43:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:30.951 23:43:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:09:30.951 23:43:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:30.951 23:43:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:30.951 [2024-11-26 23:43:19.065481] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:30.951 [2024-11-26 23:43:19.065567] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:30.951 [2024-11-26 23:43:19.065648] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:30.951 [2024-11-26 23:43:19.067392] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:30.951 [2024-11-26 23:43:19.067479] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:09:30.951 23:43:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:30.951 23:43:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:09:30.951 23:43:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:30.951 23:43:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:30.951 23:43:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:30.951 23:43:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:30.951 23:43:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:30.951 23:43:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:30.951 23:43:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:30.951 23:43:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:30.951 23:43:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:30.951 23:43:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:31.211 23:43:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:31.211 23:43:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:31.211 23:43:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:31.211 23:43:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:31.211 23:43:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:31.211 "name": "Existed_Raid", 00:09:31.211 "uuid": "f96537f8-4615-4817-9be2-dd55a602b4c2", 00:09:31.211 "strip_size_kb": 64, 00:09:31.211 "state": "configuring", 00:09:31.211 "raid_level": "concat", 00:09:31.211 "superblock": true, 00:09:31.211 "num_base_bdevs": 4, 00:09:31.211 "num_base_bdevs_discovered": 3, 00:09:31.211 "num_base_bdevs_operational": 4, 00:09:31.211 "base_bdevs_list": [ 00:09:31.211 { 00:09:31.211 "name": "BaseBdev1", 00:09:31.211 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:31.211 "is_configured": false, 00:09:31.211 "data_offset": 0, 00:09:31.211 "data_size": 0 00:09:31.211 }, 00:09:31.211 { 00:09:31.211 "name": "BaseBdev2", 00:09:31.211 "uuid": "b0ba44a4-bafa-4b62-9a9b-bc3538366303", 00:09:31.211 "is_configured": true, 00:09:31.211 "data_offset": 2048, 00:09:31.211 "data_size": 63488 00:09:31.211 }, 00:09:31.211 { 00:09:31.211 "name": "BaseBdev3", 00:09:31.211 "uuid": "df1c05d3-aec8-490a-aba7-ebac08556835", 00:09:31.211 "is_configured": true, 00:09:31.211 "data_offset": 2048, 00:09:31.211 "data_size": 63488 00:09:31.211 }, 00:09:31.211 { 00:09:31.211 "name": "BaseBdev4", 00:09:31.211 "uuid": "a244b1a4-8a1b-47d0-b1b5-74bd778ade01", 00:09:31.211 "is_configured": true, 00:09:31.211 "data_offset": 2048, 00:09:31.211 "data_size": 63488 00:09:31.211 } 00:09:31.211 ] 00:09:31.211 }' 00:09:31.211 23:43:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:31.211 23:43:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:31.471 23:43:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:09:31.471 23:43:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:31.471 23:43:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:31.471 [2024-11-26 23:43:19.468843] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:09:31.471 23:43:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:31.471 23:43:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:09:31.471 23:43:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:31.471 23:43:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:31.471 23:43:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:31.471 23:43:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:31.471 23:43:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:31.471 23:43:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:31.471 23:43:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:31.471 23:43:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:31.471 23:43:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:31.471 23:43:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:31.471 23:43:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:31.471 23:43:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:31.471 23:43:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:31.471 23:43:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:31.471 23:43:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:31.471 "name": "Existed_Raid", 00:09:31.471 "uuid": "f96537f8-4615-4817-9be2-dd55a602b4c2", 00:09:31.471 "strip_size_kb": 64, 00:09:31.471 "state": "configuring", 00:09:31.471 "raid_level": "concat", 00:09:31.471 "superblock": true, 00:09:31.471 "num_base_bdevs": 4, 00:09:31.471 "num_base_bdevs_discovered": 2, 00:09:31.472 "num_base_bdevs_operational": 4, 00:09:31.472 "base_bdevs_list": [ 00:09:31.472 { 00:09:31.472 "name": "BaseBdev1", 00:09:31.472 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:31.472 "is_configured": false, 00:09:31.472 "data_offset": 0, 00:09:31.472 "data_size": 0 00:09:31.472 }, 00:09:31.472 { 00:09:31.472 "name": null, 00:09:31.472 "uuid": "b0ba44a4-bafa-4b62-9a9b-bc3538366303", 00:09:31.472 "is_configured": false, 00:09:31.472 "data_offset": 0, 00:09:31.472 "data_size": 63488 00:09:31.472 }, 00:09:31.472 { 00:09:31.472 "name": "BaseBdev3", 00:09:31.472 "uuid": "df1c05d3-aec8-490a-aba7-ebac08556835", 00:09:31.472 "is_configured": true, 00:09:31.472 "data_offset": 2048, 00:09:31.472 "data_size": 63488 00:09:31.472 }, 00:09:31.472 { 00:09:31.472 "name": "BaseBdev4", 00:09:31.472 "uuid": "a244b1a4-8a1b-47d0-b1b5-74bd778ade01", 00:09:31.472 "is_configured": true, 00:09:31.472 "data_offset": 2048, 00:09:31.472 "data_size": 63488 00:09:31.472 } 00:09:31.472 ] 00:09:31.472 }' 00:09:31.472 23:43:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:31.472 23:43:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:32.039 23:43:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:32.039 23:43:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:32.039 23:43:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:32.039 23:43:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:09:32.039 23:43:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:32.039 23:43:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:09:32.039 23:43:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:09:32.039 23:43:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:32.039 23:43:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:32.039 [2024-11-26 23:43:19.946898] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:32.039 BaseBdev1 00:09:32.039 23:43:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:32.039 23:43:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:09:32.039 23:43:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:09:32.039 23:43:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:32.039 23:43:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:09:32.039 23:43:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:32.039 23:43:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:32.039 23:43:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:32.039 23:43:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:32.039 23:43:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:32.039 23:43:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:32.039 23:43:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:09:32.039 23:43:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:32.039 23:43:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:32.039 [ 00:09:32.039 { 00:09:32.039 "name": "BaseBdev1", 00:09:32.039 "aliases": [ 00:09:32.039 "d6c970d1-8138-4dc0-aa8d-581cf5b98584" 00:09:32.039 ], 00:09:32.039 "product_name": "Malloc disk", 00:09:32.039 "block_size": 512, 00:09:32.039 "num_blocks": 65536, 00:09:32.039 "uuid": "d6c970d1-8138-4dc0-aa8d-581cf5b98584", 00:09:32.039 "assigned_rate_limits": { 00:09:32.039 "rw_ios_per_sec": 0, 00:09:32.039 "rw_mbytes_per_sec": 0, 00:09:32.039 "r_mbytes_per_sec": 0, 00:09:32.039 "w_mbytes_per_sec": 0 00:09:32.039 }, 00:09:32.039 "claimed": true, 00:09:32.039 "claim_type": "exclusive_write", 00:09:32.039 "zoned": false, 00:09:32.039 "supported_io_types": { 00:09:32.039 "read": true, 00:09:32.039 "write": true, 00:09:32.039 "unmap": true, 00:09:32.039 "flush": true, 00:09:32.039 "reset": true, 00:09:32.039 "nvme_admin": false, 00:09:32.039 "nvme_io": false, 00:09:32.039 "nvme_io_md": false, 00:09:32.039 "write_zeroes": true, 00:09:32.039 "zcopy": true, 00:09:32.039 "get_zone_info": false, 00:09:32.039 "zone_management": false, 00:09:32.039 "zone_append": false, 00:09:32.039 "compare": false, 00:09:32.039 "compare_and_write": false, 00:09:32.039 "abort": true, 00:09:32.039 "seek_hole": false, 00:09:32.039 "seek_data": false, 00:09:32.039 "copy": true, 00:09:32.039 "nvme_iov_md": false 00:09:32.039 }, 00:09:32.039 "memory_domains": [ 00:09:32.039 { 00:09:32.039 "dma_device_id": "system", 00:09:32.039 "dma_device_type": 1 00:09:32.039 }, 00:09:32.039 { 00:09:32.039 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:32.039 "dma_device_type": 2 00:09:32.039 } 00:09:32.039 ], 00:09:32.039 "driver_specific": {} 00:09:32.039 } 00:09:32.039 ] 00:09:32.039 23:43:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:32.039 23:43:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:09:32.039 23:43:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:09:32.039 23:43:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:32.039 23:43:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:32.039 23:43:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:32.039 23:43:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:32.039 23:43:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:32.039 23:43:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:32.039 23:43:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:32.039 23:43:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:32.039 23:43:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:32.039 23:43:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:32.039 23:43:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:32.039 23:43:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:32.039 23:43:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:32.039 23:43:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:32.039 23:43:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:32.039 "name": "Existed_Raid", 00:09:32.039 "uuid": "f96537f8-4615-4817-9be2-dd55a602b4c2", 00:09:32.039 "strip_size_kb": 64, 00:09:32.039 "state": "configuring", 00:09:32.039 "raid_level": "concat", 00:09:32.039 "superblock": true, 00:09:32.039 "num_base_bdevs": 4, 00:09:32.039 "num_base_bdevs_discovered": 3, 00:09:32.039 "num_base_bdevs_operational": 4, 00:09:32.039 "base_bdevs_list": [ 00:09:32.039 { 00:09:32.039 "name": "BaseBdev1", 00:09:32.039 "uuid": "d6c970d1-8138-4dc0-aa8d-581cf5b98584", 00:09:32.039 "is_configured": true, 00:09:32.039 "data_offset": 2048, 00:09:32.039 "data_size": 63488 00:09:32.039 }, 00:09:32.039 { 00:09:32.039 "name": null, 00:09:32.039 "uuid": "b0ba44a4-bafa-4b62-9a9b-bc3538366303", 00:09:32.039 "is_configured": false, 00:09:32.039 "data_offset": 0, 00:09:32.039 "data_size": 63488 00:09:32.039 }, 00:09:32.039 { 00:09:32.039 "name": "BaseBdev3", 00:09:32.039 "uuid": "df1c05d3-aec8-490a-aba7-ebac08556835", 00:09:32.039 "is_configured": true, 00:09:32.039 "data_offset": 2048, 00:09:32.039 "data_size": 63488 00:09:32.039 }, 00:09:32.039 { 00:09:32.039 "name": "BaseBdev4", 00:09:32.039 "uuid": "a244b1a4-8a1b-47d0-b1b5-74bd778ade01", 00:09:32.039 "is_configured": true, 00:09:32.039 "data_offset": 2048, 00:09:32.039 "data_size": 63488 00:09:32.039 } 00:09:32.039 ] 00:09:32.039 }' 00:09:32.039 23:43:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:32.039 23:43:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:32.298 23:43:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:32.298 23:43:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:32.298 23:43:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:09:32.298 23:43:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:32.298 23:43:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:32.556 23:43:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:09:32.556 23:43:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:09:32.556 23:43:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:32.556 23:43:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:32.557 [2024-11-26 23:43:20.446110] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:09:32.557 23:43:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:32.557 23:43:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:09:32.557 23:43:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:32.557 23:43:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:32.557 23:43:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:32.557 23:43:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:32.557 23:43:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:32.557 23:43:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:32.557 23:43:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:32.557 23:43:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:32.557 23:43:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:32.557 23:43:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:32.557 23:43:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:32.557 23:43:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:32.557 23:43:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:32.557 23:43:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:32.557 23:43:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:32.557 "name": "Existed_Raid", 00:09:32.557 "uuid": "f96537f8-4615-4817-9be2-dd55a602b4c2", 00:09:32.557 "strip_size_kb": 64, 00:09:32.557 "state": "configuring", 00:09:32.557 "raid_level": "concat", 00:09:32.557 "superblock": true, 00:09:32.557 "num_base_bdevs": 4, 00:09:32.557 "num_base_bdevs_discovered": 2, 00:09:32.557 "num_base_bdevs_operational": 4, 00:09:32.557 "base_bdevs_list": [ 00:09:32.557 { 00:09:32.557 "name": "BaseBdev1", 00:09:32.557 "uuid": "d6c970d1-8138-4dc0-aa8d-581cf5b98584", 00:09:32.557 "is_configured": true, 00:09:32.557 "data_offset": 2048, 00:09:32.557 "data_size": 63488 00:09:32.557 }, 00:09:32.557 { 00:09:32.557 "name": null, 00:09:32.557 "uuid": "b0ba44a4-bafa-4b62-9a9b-bc3538366303", 00:09:32.557 "is_configured": false, 00:09:32.557 "data_offset": 0, 00:09:32.557 "data_size": 63488 00:09:32.557 }, 00:09:32.557 { 00:09:32.557 "name": null, 00:09:32.557 "uuid": "df1c05d3-aec8-490a-aba7-ebac08556835", 00:09:32.557 "is_configured": false, 00:09:32.557 "data_offset": 0, 00:09:32.557 "data_size": 63488 00:09:32.557 }, 00:09:32.557 { 00:09:32.557 "name": "BaseBdev4", 00:09:32.557 "uuid": "a244b1a4-8a1b-47d0-b1b5-74bd778ade01", 00:09:32.557 "is_configured": true, 00:09:32.557 "data_offset": 2048, 00:09:32.557 "data_size": 63488 00:09:32.557 } 00:09:32.557 ] 00:09:32.557 }' 00:09:32.557 23:43:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:32.557 23:43:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:32.816 23:43:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:32.816 23:43:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:32.816 23:43:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:32.816 23:43:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:09:32.816 23:43:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:32.816 23:43:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:09:32.816 23:43:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:09:32.816 23:43:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:32.816 23:43:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:32.816 [2024-11-26 23:43:20.921332] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:32.816 23:43:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:32.816 23:43:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:09:32.816 23:43:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:32.816 23:43:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:32.816 23:43:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:32.816 23:43:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:32.816 23:43:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:32.816 23:43:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:32.816 23:43:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:32.816 23:43:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:32.816 23:43:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:32.816 23:43:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:32.816 23:43:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:32.816 23:43:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:32.816 23:43:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:33.075 23:43:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:33.075 23:43:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:33.075 "name": "Existed_Raid", 00:09:33.075 "uuid": "f96537f8-4615-4817-9be2-dd55a602b4c2", 00:09:33.075 "strip_size_kb": 64, 00:09:33.075 "state": "configuring", 00:09:33.075 "raid_level": "concat", 00:09:33.075 "superblock": true, 00:09:33.075 "num_base_bdevs": 4, 00:09:33.075 "num_base_bdevs_discovered": 3, 00:09:33.076 "num_base_bdevs_operational": 4, 00:09:33.076 "base_bdevs_list": [ 00:09:33.076 { 00:09:33.076 "name": "BaseBdev1", 00:09:33.076 "uuid": "d6c970d1-8138-4dc0-aa8d-581cf5b98584", 00:09:33.076 "is_configured": true, 00:09:33.076 "data_offset": 2048, 00:09:33.076 "data_size": 63488 00:09:33.076 }, 00:09:33.076 { 00:09:33.076 "name": null, 00:09:33.076 "uuid": "b0ba44a4-bafa-4b62-9a9b-bc3538366303", 00:09:33.076 "is_configured": false, 00:09:33.076 "data_offset": 0, 00:09:33.076 "data_size": 63488 00:09:33.076 }, 00:09:33.076 { 00:09:33.076 "name": "BaseBdev3", 00:09:33.076 "uuid": "df1c05d3-aec8-490a-aba7-ebac08556835", 00:09:33.076 "is_configured": true, 00:09:33.076 "data_offset": 2048, 00:09:33.076 "data_size": 63488 00:09:33.076 }, 00:09:33.076 { 00:09:33.076 "name": "BaseBdev4", 00:09:33.076 "uuid": "a244b1a4-8a1b-47d0-b1b5-74bd778ade01", 00:09:33.076 "is_configured": true, 00:09:33.076 "data_offset": 2048, 00:09:33.076 "data_size": 63488 00:09:33.076 } 00:09:33.076 ] 00:09:33.076 }' 00:09:33.076 23:43:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:33.076 23:43:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:33.335 23:43:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:33.335 23:43:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:09:33.335 23:43:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:33.335 23:43:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:33.335 23:43:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:33.335 23:43:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:09:33.335 23:43:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:09:33.335 23:43:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:33.335 23:43:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:33.335 [2024-11-26 23:43:21.388563] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:09:33.335 23:43:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:33.335 23:43:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:09:33.335 23:43:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:33.335 23:43:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:33.335 23:43:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:33.335 23:43:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:33.335 23:43:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:33.335 23:43:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:33.335 23:43:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:33.335 23:43:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:33.335 23:43:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:33.335 23:43:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:33.335 23:43:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:33.335 23:43:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:33.335 23:43:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:33.335 23:43:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:33.335 23:43:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:33.335 "name": "Existed_Raid", 00:09:33.335 "uuid": "f96537f8-4615-4817-9be2-dd55a602b4c2", 00:09:33.335 "strip_size_kb": 64, 00:09:33.335 "state": "configuring", 00:09:33.335 "raid_level": "concat", 00:09:33.335 "superblock": true, 00:09:33.335 "num_base_bdevs": 4, 00:09:33.335 "num_base_bdevs_discovered": 2, 00:09:33.335 "num_base_bdevs_operational": 4, 00:09:33.335 "base_bdevs_list": [ 00:09:33.335 { 00:09:33.335 "name": null, 00:09:33.335 "uuid": "d6c970d1-8138-4dc0-aa8d-581cf5b98584", 00:09:33.335 "is_configured": false, 00:09:33.335 "data_offset": 0, 00:09:33.335 "data_size": 63488 00:09:33.335 }, 00:09:33.335 { 00:09:33.335 "name": null, 00:09:33.335 "uuid": "b0ba44a4-bafa-4b62-9a9b-bc3538366303", 00:09:33.335 "is_configured": false, 00:09:33.335 "data_offset": 0, 00:09:33.335 "data_size": 63488 00:09:33.335 }, 00:09:33.336 { 00:09:33.336 "name": "BaseBdev3", 00:09:33.336 "uuid": "df1c05d3-aec8-490a-aba7-ebac08556835", 00:09:33.336 "is_configured": true, 00:09:33.336 "data_offset": 2048, 00:09:33.336 "data_size": 63488 00:09:33.336 }, 00:09:33.336 { 00:09:33.336 "name": "BaseBdev4", 00:09:33.336 "uuid": "a244b1a4-8a1b-47d0-b1b5-74bd778ade01", 00:09:33.336 "is_configured": true, 00:09:33.336 "data_offset": 2048, 00:09:33.336 "data_size": 63488 00:09:33.336 } 00:09:33.336 ] 00:09:33.336 }' 00:09:33.336 23:43:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:33.336 23:43:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:33.921 23:43:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:33.921 23:43:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:09:33.921 23:43:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:33.921 23:43:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:33.921 23:43:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:33.921 23:43:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:09:33.921 23:43:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:09:33.921 23:43:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:33.921 23:43:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:33.921 [2024-11-26 23:43:21.902240] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:33.921 23:43:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:33.921 23:43:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:09:33.921 23:43:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:33.921 23:43:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:33.921 23:43:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:33.921 23:43:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:33.921 23:43:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:33.921 23:43:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:33.921 23:43:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:33.921 23:43:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:33.921 23:43:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:33.921 23:43:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:33.921 23:43:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:33.921 23:43:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:33.921 23:43:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:33.921 23:43:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:33.921 23:43:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:33.921 "name": "Existed_Raid", 00:09:33.921 "uuid": "f96537f8-4615-4817-9be2-dd55a602b4c2", 00:09:33.921 "strip_size_kb": 64, 00:09:33.921 "state": "configuring", 00:09:33.921 "raid_level": "concat", 00:09:33.921 "superblock": true, 00:09:33.921 "num_base_bdevs": 4, 00:09:33.921 "num_base_bdevs_discovered": 3, 00:09:33.921 "num_base_bdevs_operational": 4, 00:09:33.921 "base_bdevs_list": [ 00:09:33.921 { 00:09:33.921 "name": null, 00:09:33.921 "uuid": "d6c970d1-8138-4dc0-aa8d-581cf5b98584", 00:09:33.921 "is_configured": false, 00:09:33.921 "data_offset": 0, 00:09:33.921 "data_size": 63488 00:09:33.921 }, 00:09:33.921 { 00:09:33.921 "name": "BaseBdev2", 00:09:33.921 "uuid": "b0ba44a4-bafa-4b62-9a9b-bc3538366303", 00:09:33.921 "is_configured": true, 00:09:33.921 "data_offset": 2048, 00:09:33.921 "data_size": 63488 00:09:33.921 }, 00:09:33.921 { 00:09:33.921 "name": "BaseBdev3", 00:09:33.921 "uuid": "df1c05d3-aec8-490a-aba7-ebac08556835", 00:09:33.921 "is_configured": true, 00:09:33.921 "data_offset": 2048, 00:09:33.921 "data_size": 63488 00:09:33.921 }, 00:09:33.921 { 00:09:33.921 "name": "BaseBdev4", 00:09:33.921 "uuid": "a244b1a4-8a1b-47d0-b1b5-74bd778ade01", 00:09:33.921 "is_configured": true, 00:09:33.921 "data_offset": 2048, 00:09:33.921 "data_size": 63488 00:09:33.921 } 00:09:33.921 ] 00:09:33.921 }' 00:09:33.921 23:43:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:33.921 23:43:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:34.180 23:43:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:34.181 23:43:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:09:34.181 23:43:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:34.181 23:43:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:34.440 23:43:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:34.440 23:43:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:09:34.440 23:43:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:34.440 23:43:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:34.440 23:43:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:34.440 23:43:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:09:34.440 23:43:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:34.440 23:43:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u d6c970d1-8138-4dc0-aa8d-581cf5b98584 00:09:34.440 23:43:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:34.440 23:43:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:34.440 [2024-11-26 23:43:22.404235] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:09:34.440 [2024-11-26 23:43:22.404492] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:09:34.440 [2024-11-26 23:43:22.404540] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:09:34.440 NewBaseBdev 00:09:34.440 [2024-11-26 23:43:22.404805] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002a10 00:09:34.440 [2024-11-26 23:43:22.404920] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:09:34.440 [2024-11-26 23:43:22.404936] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001c80 00:09:34.440 [2024-11-26 23:43:22.405026] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:34.440 23:43:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:34.440 23:43:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:09:34.440 23:43:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:09:34.440 23:43:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:34.440 23:43:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:09:34.440 23:43:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:34.440 23:43:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:34.440 23:43:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:34.440 23:43:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:34.440 23:43:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:34.441 23:43:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:34.441 23:43:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:09:34.441 23:43:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:34.441 23:43:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:34.441 [ 00:09:34.441 { 00:09:34.441 "name": "NewBaseBdev", 00:09:34.441 "aliases": [ 00:09:34.441 "d6c970d1-8138-4dc0-aa8d-581cf5b98584" 00:09:34.441 ], 00:09:34.441 "product_name": "Malloc disk", 00:09:34.441 "block_size": 512, 00:09:34.441 "num_blocks": 65536, 00:09:34.441 "uuid": "d6c970d1-8138-4dc0-aa8d-581cf5b98584", 00:09:34.441 "assigned_rate_limits": { 00:09:34.441 "rw_ios_per_sec": 0, 00:09:34.441 "rw_mbytes_per_sec": 0, 00:09:34.441 "r_mbytes_per_sec": 0, 00:09:34.441 "w_mbytes_per_sec": 0 00:09:34.441 }, 00:09:34.441 "claimed": true, 00:09:34.441 "claim_type": "exclusive_write", 00:09:34.441 "zoned": false, 00:09:34.441 "supported_io_types": { 00:09:34.441 "read": true, 00:09:34.441 "write": true, 00:09:34.441 "unmap": true, 00:09:34.441 "flush": true, 00:09:34.441 "reset": true, 00:09:34.441 "nvme_admin": false, 00:09:34.441 "nvme_io": false, 00:09:34.441 "nvme_io_md": false, 00:09:34.441 "write_zeroes": true, 00:09:34.441 "zcopy": true, 00:09:34.441 "get_zone_info": false, 00:09:34.441 "zone_management": false, 00:09:34.441 "zone_append": false, 00:09:34.441 "compare": false, 00:09:34.441 "compare_and_write": false, 00:09:34.441 "abort": true, 00:09:34.441 "seek_hole": false, 00:09:34.441 "seek_data": false, 00:09:34.441 "copy": true, 00:09:34.441 "nvme_iov_md": false 00:09:34.441 }, 00:09:34.441 "memory_domains": [ 00:09:34.441 { 00:09:34.441 "dma_device_id": "system", 00:09:34.441 "dma_device_type": 1 00:09:34.441 }, 00:09:34.441 { 00:09:34.441 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:34.441 "dma_device_type": 2 00:09:34.441 } 00:09:34.441 ], 00:09:34.441 "driver_specific": {} 00:09:34.441 } 00:09:34.441 ] 00:09:34.441 23:43:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:34.441 23:43:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:09:34.441 23:43:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online concat 64 4 00:09:34.441 23:43:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:34.441 23:43:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:34.441 23:43:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:34.441 23:43:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:34.441 23:43:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:34.441 23:43:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:34.441 23:43:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:34.441 23:43:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:34.441 23:43:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:34.441 23:43:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:34.441 23:43:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:34.441 23:43:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:34.441 23:43:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:34.441 23:43:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:34.441 23:43:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:34.441 "name": "Existed_Raid", 00:09:34.441 "uuid": "f96537f8-4615-4817-9be2-dd55a602b4c2", 00:09:34.441 "strip_size_kb": 64, 00:09:34.441 "state": "online", 00:09:34.441 "raid_level": "concat", 00:09:34.441 "superblock": true, 00:09:34.441 "num_base_bdevs": 4, 00:09:34.441 "num_base_bdevs_discovered": 4, 00:09:34.441 "num_base_bdevs_operational": 4, 00:09:34.441 "base_bdevs_list": [ 00:09:34.441 { 00:09:34.441 "name": "NewBaseBdev", 00:09:34.441 "uuid": "d6c970d1-8138-4dc0-aa8d-581cf5b98584", 00:09:34.441 "is_configured": true, 00:09:34.441 "data_offset": 2048, 00:09:34.441 "data_size": 63488 00:09:34.441 }, 00:09:34.441 { 00:09:34.441 "name": "BaseBdev2", 00:09:34.441 "uuid": "b0ba44a4-bafa-4b62-9a9b-bc3538366303", 00:09:34.441 "is_configured": true, 00:09:34.441 "data_offset": 2048, 00:09:34.441 "data_size": 63488 00:09:34.441 }, 00:09:34.441 { 00:09:34.441 "name": "BaseBdev3", 00:09:34.441 "uuid": "df1c05d3-aec8-490a-aba7-ebac08556835", 00:09:34.441 "is_configured": true, 00:09:34.441 "data_offset": 2048, 00:09:34.441 "data_size": 63488 00:09:34.441 }, 00:09:34.441 { 00:09:34.441 "name": "BaseBdev4", 00:09:34.441 "uuid": "a244b1a4-8a1b-47d0-b1b5-74bd778ade01", 00:09:34.441 "is_configured": true, 00:09:34.441 "data_offset": 2048, 00:09:34.441 "data_size": 63488 00:09:34.441 } 00:09:34.441 ] 00:09:34.441 }' 00:09:34.441 23:43:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:34.441 23:43:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:35.009 23:43:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:09:35.009 23:43:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:09:35.009 23:43:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:35.009 23:43:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:35.009 23:43:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:09:35.009 23:43:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:35.009 23:43:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:09:35.009 23:43:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:35.009 23:43:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:35.009 23:43:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:35.009 [2024-11-26 23:43:22.891816] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:35.009 23:43:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:35.009 23:43:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:35.009 "name": "Existed_Raid", 00:09:35.009 "aliases": [ 00:09:35.009 "f96537f8-4615-4817-9be2-dd55a602b4c2" 00:09:35.009 ], 00:09:35.009 "product_name": "Raid Volume", 00:09:35.009 "block_size": 512, 00:09:35.009 "num_blocks": 253952, 00:09:35.009 "uuid": "f96537f8-4615-4817-9be2-dd55a602b4c2", 00:09:35.009 "assigned_rate_limits": { 00:09:35.009 "rw_ios_per_sec": 0, 00:09:35.009 "rw_mbytes_per_sec": 0, 00:09:35.009 "r_mbytes_per_sec": 0, 00:09:35.009 "w_mbytes_per_sec": 0 00:09:35.009 }, 00:09:35.009 "claimed": false, 00:09:35.009 "zoned": false, 00:09:35.009 "supported_io_types": { 00:09:35.009 "read": true, 00:09:35.009 "write": true, 00:09:35.009 "unmap": true, 00:09:35.009 "flush": true, 00:09:35.009 "reset": true, 00:09:35.009 "nvme_admin": false, 00:09:35.009 "nvme_io": false, 00:09:35.009 "nvme_io_md": false, 00:09:35.009 "write_zeroes": true, 00:09:35.009 "zcopy": false, 00:09:35.009 "get_zone_info": false, 00:09:35.009 "zone_management": false, 00:09:35.009 "zone_append": false, 00:09:35.009 "compare": false, 00:09:35.009 "compare_and_write": false, 00:09:35.009 "abort": false, 00:09:35.009 "seek_hole": false, 00:09:35.009 "seek_data": false, 00:09:35.009 "copy": false, 00:09:35.009 "nvme_iov_md": false 00:09:35.009 }, 00:09:35.009 "memory_domains": [ 00:09:35.009 { 00:09:35.009 "dma_device_id": "system", 00:09:35.009 "dma_device_type": 1 00:09:35.009 }, 00:09:35.009 { 00:09:35.009 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:35.009 "dma_device_type": 2 00:09:35.009 }, 00:09:35.009 { 00:09:35.009 "dma_device_id": "system", 00:09:35.009 "dma_device_type": 1 00:09:35.009 }, 00:09:35.009 { 00:09:35.009 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:35.009 "dma_device_type": 2 00:09:35.009 }, 00:09:35.009 { 00:09:35.009 "dma_device_id": "system", 00:09:35.009 "dma_device_type": 1 00:09:35.009 }, 00:09:35.009 { 00:09:35.009 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:35.009 "dma_device_type": 2 00:09:35.009 }, 00:09:35.009 { 00:09:35.009 "dma_device_id": "system", 00:09:35.009 "dma_device_type": 1 00:09:35.009 }, 00:09:35.009 { 00:09:35.009 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:35.009 "dma_device_type": 2 00:09:35.009 } 00:09:35.009 ], 00:09:35.009 "driver_specific": { 00:09:35.009 "raid": { 00:09:35.009 "uuid": "f96537f8-4615-4817-9be2-dd55a602b4c2", 00:09:35.009 "strip_size_kb": 64, 00:09:35.009 "state": "online", 00:09:35.009 "raid_level": "concat", 00:09:35.009 "superblock": true, 00:09:35.009 "num_base_bdevs": 4, 00:09:35.009 "num_base_bdevs_discovered": 4, 00:09:35.009 "num_base_bdevs_operational": 4, 00:09:35.009 "base_bdevs_list": [ 00:09:35.009 { 00:09:35.009 "name": "NewBaseBdev", 00:09:35.009 "uuid": "d6c970d1-8138-4dc0-aa8d-581cf5b98584", 00:09:35.010 "is_configured": true, 00:09:35.010 "data_offset": 2048, 00:09:35.010 "data_size": 63488 00:09:35.010 }, 00:09:35.010 { 00:09:35.010 "name": "BaseBdev2", 00:09:35.010 "uuid": "b0ba44a4-bafa-4b62-9a9b-bc3538366303", 00:09:35.010 "is_configured": true, 00:09:35.010 "data_offset": 2048, 00:09:35.010 "data_size": 63488 00:09:35.010 }, 00:09:35.010 { 00:09:35.010 "name": "BaseBdev3", 00:09:35.010 "uuid": "df1c05d3-aec8-490a-aba7-ebac08556835", 00:09:35.010 "is_configured": true, 00:09:35.010 "data_offset": 2048, 00:09:35.010 "data_size": 63488 00:09:35.010 }, 00:09:35.010 { 00:09:35.010 "name": "BaseBdev4", 00:09:35.010 "uuid": "a244b1a4-8a1b-47d0-b1b5-74bd778ade01", 00:09:35.010 "is_configured": true, 00:09:35.010 "data_offset": 2048, 00:09:35.010 "data_size": 63488 00:09:35.010 } 00:09:35.010 ] 00:09:35.010 } 00:09:35.010 } 00:09:35.010 }' 00:09:35.010 23:43:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:35.010 23:43:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:09:35.010 BaseBdev2 00:09:35.010 BaseBdev3 00:09:35.010 BaseBdev4' 00:09:35.010 23:43:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:35.010 23:43:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:35.010 23:43:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:35.010 23:43:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:09:35.010 23:43:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:35.010 23:43:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:35.010 23:43:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:35.010 23:43:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:35.010 23:43:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:35.010 23:43:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:35.010 23:43:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:35.010 23:43:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:35.010 23:43:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:09:35.010 23:43:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:35.010 23:43:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:35.010 23:43:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:35.010 23:43:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:35.010 23:43:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:35.010 23:43:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:35.010 23:43:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:09:35.010 23:43:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:35.010 23:43:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:35.010 23:43:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:35.010 23:43:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:35.269 23:43:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:35.269 23:43:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:35.269 23:43:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:35.269 23:43:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:35.269 23:43:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:09:35.269 23:43:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:35.269 23:43:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:35.269 23:43:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:35.269 23:43:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:35.269 23:43:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:35.269 23:43:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:35.269 23:43:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:35.269 23:43:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:35.269 [2024-11-26 23:43:23.206893] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:35.269 [2024-11-26 23:43:23.206966] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:35.269 [2024-11-26 23:43:23.207060] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:35.269 [2024-11-26 23:43:23.207151] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:35.269 [2024-11-26 23:43:23.207196] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name Existed_Raid, state offline 00:09:35.269 23:43:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:35.269 23:43:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 82531 00:09:35.269 23:43:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # '[' -z 82531 ']' 00:09:35.269 23:43:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@958 -- # kill -0 82531 00:09:35.269 23:43:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # uname 00:09:35.269 23:43:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:09:35.269 23:43:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 82531 00:09:35.269 23:43:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:09:35.269 killing process with pid 82531 00:09:35.269 23:43:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:09:35.269 23:43:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@972 -- # echo 'killing process with pid 82531' 00:09:35.269 23:43:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@973 -- # kill 82531 00:09:35.269 [2024-11-26 23:43:23.254740] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:09:35.269 23:43:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@978 -- # wait 82531 00:09:35.269 [2024-11-26 23:43:23.294644] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:09:35.550 23:43:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:09:35.550 00:09:35.550 real 0m9.114s 00:09:35.550 user 0m15.633s 00:09:35.550 sys 0m1.821s 00:09:35.550 23:43:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1130 -- # xtrace_disable 00:09:35.550 23:43:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:35.550 ************************************ 00:09:35.550 END TEST raid_state_function_test_sb 00:09:35.550 ************************************ 00:09:35.550 23:43:23 bdev_raid -- bdev/bdev_raid.sh@970 -- # run_test raid_superblock_test raid_superblock_test concat 4 00:09:35.550 23:43:23 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:09:35.550 23:43:23 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:09:35.550 23:43:23 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:09:35.550 ************************************ 00:09:35.550 START TEST raid_superblock_test 00:09:35.550 ************************************ 00:09:35.550 23:43:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1129 -- # raid_superblock_test concat 4 00:09:35.550 23:43:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=concat 00:09:35.550 23:43:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=4 00:09:35.550 23:43:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:09:35.550 23:43:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:09:35.550 23:43:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:09:35.550 23:43:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:09:35.550 23:43:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:09:35.550 23:43:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:09:35.551 23:43:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:09:35.551 23:43:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:09:35.551 23:43:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:09:35.551 23:43:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:09:35.551 23:43:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:09:35.551 23:43:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' concat '!=' raid1 ']' 00:09:35.551 23:43:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@405 -- # strip_size=64 00:09:35.551 23:43:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@406 -- # strip_size_create_arg='-z 64' 00:09:35.551 23:43:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=83175 00:09:35.551 23:43:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:09:35.551 23:43:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 83175 00:09:35.551 23:43:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@835 -- # '[' -z 83175 ']' 00:09:35.551 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:35.551 23:43:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:35.551 23:43:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:09:35.551 23:43:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:35.551 23:43:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:09:35.551 23:43:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:35.832 [2024-11-26 23:43:23.692035] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:09:35.832 [2024-11-26 23:43:23.692216] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid83175 ] 00:09:35.832 [2024-11-26 23:43:23.852607] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:35.832 [2024-11-26 23:43:23.877473] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:09:35.832 [2024-11-26 23:43:23.919741] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:35.832 [2024-11-26 23:43:23.919778] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:36.765 23:43:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:09:36.766 23:43:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@868 -- # return 0 00:09:36.766 23:43:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:09:36.766 23:43:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:09:36.766 23:43:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:09:36.766 23:43:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:09:36.766 23:43:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:09:36.766 23:43:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:09:36.766 23:43:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:09:36.766 23:43:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:09:36.766 23:43:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:09:36.766 23:43:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:36.766 23:43:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:36.766 malloc1 00:09:36.766 23:43:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:36.766 23:43:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:09:36.766 23:43:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:36.766 23:43:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:36.766 [2024-11-26 23:43:24.559164] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:09:36.766 [2024-11-26 23:43:24.559282] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:36.766 [2024-11-26 23:43:24.559323] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:09:36.766 [2024-11-26 23:43:24.559375] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:36.766 [2024-11-26 23:43:24.561392] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:36.766 [2024-11-26 23:43:24.561473] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:09:36.766 pt1 00:09:36.766 23:43:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:36.766 23:43:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:09:36.766 23:43:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:09:36.766 23:43:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:09:36.766 23:43:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:09:36.766 23:43:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:09:36.766 23:43:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:09:36.766 23:43:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:09:36.766 23:43:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:09:36.766 23:43:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:09:36.766 23:43:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:36.766 23:43:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:36.766 malloc2 00:09:36.766 23:43:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:36.766 23:43:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:09:36.766 23:43:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:36.766 23:43:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:36.766 [2024-11-26 23:43:24.591624] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:09:36.766 [2024-11-26 23:43:24.591720] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:36.766 [2024-11-26 23:43:24.591772] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:09:36.766 [2024-11-26 23:43:24.591802] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:36.766 [2024-11-26 23:43:24.593821] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:36.766 [2024-11-26 23:43:24.593887] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:09:36.766 pt2 00:09:36.766 23:43:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:36.766 23:43:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:09:36.766 23:43:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:09:36.766 23:43:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc3 00:09:36.766 23:43:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt3 00:09:36.766 23:43:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000003 00:09:36.766 23:43:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:09:36.766 23:43:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:09:36.766 23:43:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:09:36.766 23:43:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc3 00:09:36.766 23:43:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:36.766 23:43:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:36.766 malloc3 00:09:36.766 23:43:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:36.766 23:43:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:09:36.766 23:43:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:36.766 23:43:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:36.766 [2024-11-26 23:43:24.624101] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:09:36.766 [2024-11-26 23:43:24.624196] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:36.766 [2024-11-26 23:43:24.624245] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:09:36.766 [2024-11-26 23:43:24.624274] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:36.766 [2024-11-26 23:43:24.626298] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:36.766 [2024-11-26 23:43:24.626385] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:09:36.766 pt3 00:09:36.766 23:43:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:36.766 23:43:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:09:36.766 23:43:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:09:36.766 23:43:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc4 00:09:36.766 23:43:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt4 00:09:36.766 23:43:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000004 00:09:36.766 23:43:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:09:36.766 23:43:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:09:36.766 23:43:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:09:36.766 23:43:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc4 00:09:36.766 23:43:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:36.766 23:43:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:36.766 malloc4 00:09:36.766 23:43:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:36.766 23:43:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:09:36.766 23:43:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:36.766 23:43:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:36.766 [2024-11-26 23:43:24.664400] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:09:36.766 [2024-11-26 23:43:24.664504] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:36.766 [2024-11-26 23:43:24.664542] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:09:36.766 [2024-11-26 23:43:24.664580] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:36.766 [2024-11-26 23:43:24.666750] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:36.766 [2024-11-26 23:43:24.666822] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:09:36.766 pt4 00:09:36.766 23:43:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:36.766 23:43:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:09:36.766 23:43:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:09:36.766 23:43:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''pt1 pt2 pt3 pt4'\''' -n raid_bdev1 -s 00:09:36.766 23:43:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:36.766 23:43:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:36.766 [2024-11-26 23:43:24.680387] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:09:36.766 [2024-11-26 23:43:24.682246] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:09:36.766 [2024-11-26 23:43:24.682367] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:09:36.766 [2024-11-26 23:43:24.682478] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:09:36.766 [2024-11-26 23:43:24.682673] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:09:36.766 [2024-11-26 23:43:24.682721] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:09:36.766 [2024-11-26 23:43:24.683006] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:09:36.766 [2024-11-26 23:43:24.683181] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:09:36.766 [2024-11-26 23:43:24.683225] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:09:36.766 [2024-11-26 23:43:24.683404] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:36.766 23:43:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:36.766 23:43:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online concat 64 4 00:09:36.766 23:43:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:36.767 23:43:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:36.767 23:43:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:36.767 23:43:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:36.767 23:43:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:36.767 23:43:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:36.767 23:43:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:36.767 23:43:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:36.767 23:43:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:36.767 23:43:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:36.767 23:43:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:36.767 23:43:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:36.767 23:43:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:36.767 23:43:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:36.767 23:43:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:36.767 "name": "raid_bdev1", 00:09:36.767 "uuid": "950cb8e7-29ba-4e34-b7be-7453678af9e0", 00:09:36.767 "strip_size_kb": 64, 00:09:36.767 "state": "online", 00:09:36.767 "raid_level": "concat", 00:09:36.767 "superblock": true, 00:09:36.767 "num_base_bdevs": 4, 00:09:36.767 "num_base_bdevs_discovered": 4, 00:09:36.767 "num_base_bdevs_operational": 4, 00:09:36.767 "base_bdevs_list": [ 00:09:36.767 { 00:09:36.767 "name": "pt1", 00:09:36.767 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:36.767 "is_configured": true, 00:09:36.767 "data_offset": 2048, 00:09:36.767 "data_size": 63488 00:09:36.767 }, 00:09:36.767 { 00:09:36.767 "name": "pt2", 00:09:36.767 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:36.767 "is_configured": true, 00:09:36.767 "data_offset": 2048, 00:09:36.767 "data_size": 63488 00:09:36.767 }, 00:09:36.767 { 00:09:36.767 "name": "pt3", 00:09:36.767 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:36.767 "is_configured": true, 00:09:36.767 "data_offset": 2048, 00:09:36.767 "data_size": 63488 00:09:36.767 }, 00:09:36.767 { 00:09:36.767 "name": "pt4", 00:09:36.767 "uuid": "00000000-0000-0000-0000-000000000004", 00:09:36.767 "is_configured": true, 00:09:36.767 "data_offset": 2048, 00:09:36.767 "data_size": 63488 00:09:36.767 } 00:09:36.767 ] 00:09:36.767 }' 00:09:36.767 23:43:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:36.767 23:43:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:37.336 23:43:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:09:37.336 23:43:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:09:37.336 23:43:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:37.336 23:43:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:37.336 23:43:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:09:37.336 23:43:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:37.336 23:43:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:37.336 23:43:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:09:37.336 23:43:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:37.336 23:43:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:37.336 [2024-11-26 23:43:25.175840] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:37.336 23:43:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:37.336 23:43:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:37.336 "name": "raid_bdev1", 00:09:37.336 "aliases": [ 00:09:37.336 "950cb8e7-29ba-4e34-b7be-7453678af9e0" 00:09:37.336 ], 00:09:37.336 "product_name": "Raid Volume", 00:09:37.336 "block_size": 512, 00:09:37.336 "num_blocks": 253952, 00:09:37.336 "uuid": "950cb8e7-29ba-4e34-b7be-7453678af9e0", 00:09:37.336 "assigned_rate_limits": { 00:09:37.336 "rw_ios_per_sec": 0, 00:09:37.336 "rw_mbytes_per_sec": 0, 00:09:37.336 "r_mbytes_per_sec": 0, 00:09:37.336 "w_mbytes_per_sec": 0 00:09:37.336 }, 00:09:37.336 "claimed": false, 00:09:37.336 "zoned": false, 00:09:37.336 "supported_io_types": { 00:09:37.336 "read": true, 00:09:37.336 "write": true, 00:09:37.336 "unmap": true, 00:09:37.336 "flush": true, 00:09:37.336 "reset": true, 00:09:37.336 "nvme_admin": false, 00:09:37.336 "nvme_io": false, 00:09:37.336 "nvme_io_md": false, 00:09:37.336 "write_zeroes": true, 00:09:37.336 "zcopy": false, 00:09:37.336 "get_zone_info": false, 00:09:37.336 "zone_management": false, 00:09:37.336 "zone_append": false, 00:09:37.336 "compare": false, 00:09:37.336 "compare_and_write": false, 00:09:37.336 "abort": false, 00:09:37.336 "seek_hole": false, 00:09:37.336 "seek_data": false, 00:09:37.336 "copy": false, 00:09:37.336 "nvme_iov_md": false 00:09:37.336 }, 00:09:37.336 "memory_domains": [ 00:09:37.336 { 00:09:37.336 "dma_device_id": "system", 00:09:37.336 "dma_device_type": 1 00:09:37.336 }, 00:09:37.336 { 00:09:37.336 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:37.336 "dma_device_type": 2 00:09:37.336 }, 00:09:37.336 { 00:09:37.336 "dma_device_id": "system", 00:09:37.336 "dma_device_type": 1 00:09:37.336 }, 00:09:37.336 { 00:09:37.336 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:37.336 "dma_device_type": 2 00:09:37.336 }, 00:09:37.336 { 00:09:37.336 "dma_device_id": "system", 00:09:37.336 "dma_device_type": 1 00:09:37.336 }, 00:09:37.336 { 00:09:37.336 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:37.336 "dma_device_type": 2 00:09:37.336 }, 00:09:37.336 { 00:09:37.336 "dma_device_id": "system", 00:09:37.336 "dma_device_type": 1 00:09:37.336 }, 00:09:37.336 { 00:09:37.336 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:37.336 "dma_device_type": 2 00:09:37.336 } 00:09:37.336 ], 00:09:37.336 "driver_specific": { 00:09:37.336 "raid": { 00:09:37.336 "uuid": "950cb8e7-29ba-4e34-b7be-7453678af9e0", 00:09:37.336 "strip_size_kb": 64, 00:09:37.336 "state": "online", 00:09:37.336 "raid_level": "concat", 00:09:37.336 "superblock": true, 00:09:37.336 "num_base_bdevs": 4, 00:09:37.336 "num_base_bdevs_discovered": 4, 00:09:37.336 "num_base_bdevs_operational": 4, 00:09:37.336 "base_bdevs_list": [ 00:09:37.336 { 00:09:37.337 "name": "pt1", 00:09:37.337 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:37.337 "is_configured": true, 00:09:37.337 "data_offset": 2048, 00:09:37.337 "data_size": 63488 00:09:37.337 }, 00:09:37.337 { 00:09:37.337 "name": "pt2", 00:09:37.337 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:37.337 "is_configured": true, 00:09:37.337 "data_offset": 2048, 00:09:37.337 "data_size": 63488 00:09:37.337 }, 00:09:37.337 { 00:09:37.337 "name": "pt3", 00:09:37.337 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:37.337 "is_configured": true, 00:09:37.337 "data_offset": 2048, 00:09:37.337 "data_size": 63488 00:09:37.337 }, 00:09:37.337 { 00:09:37.337 "name": "pt4", 00:09:37.337 "uuid": "00000000-0000-0000-0000-000000000004", 00:09:37.337 "is_configured": true, 00:09:37.337 "data_offset": 2048, 00:09:37.337 "data_size": 63488 00:09:37.337 } 00:09:37.337 ] 00:09:37.337 } 00:09:37.337 } 00:09:37.337 }' 00:09:37.337 23:43:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:37.337 23:43:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:09:37.337 pt2 00:09:37.337 pt3 00:09:37.337 pt4' 00:09:37.337 23:43:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:37.337 23:43:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:37.337 23:43:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:37.337 23:43:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:09:37.337 23:43:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:37.337 23:43:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:37.337 23:43:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:37.337 23:43:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:37.337 23:43:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:37.337 23:43:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:37.337 23:43:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:37.337 23:43:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:09:37.337 23:43:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:37.337 23:43:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:37.337 23:43:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:37.337 23:43:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:37.337 23:43:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:37.337 23:43:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:37.337 23:43:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:37.337 23:43:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:09:37.337 23:43:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:37.337 23:43:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:37.337 23:43:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:37.337 23:43:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:37.337 23:43:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:37.337 23:43:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:37.337 23:43:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:37.337 23:43:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt4 00:09:37.337 23:43:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:37.337 23:43:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:37.337 23:43:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:37.337 23:43:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:37.597 23:43:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:37.597 23:43:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:37.597 23:43:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:09:37.597 23:43:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:09:37.597 23:43:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:37.597 23:43:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:37.597 [2024-11-26 23:43:25.495239] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:37.597 23:43:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:37.597 23:43:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=950cb8e7-29ba-4e34-b7be-7453678af9e0 00:09:37.597 23:43:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z 950cb8e7-29ba-4e34-b7be-7453678af9e0 ']' 00:09:37.597 23:43:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:09:37.597 23:43:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:37.597 23:43:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:37.597 [2024-11-26 23:43:25.534884] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:37.597 [2024-11-26 23:43:25.534916] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:37.597 [2024-11-26 23:43:25.534991] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:37.597 [2024-11-26 23:43:25.535074] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:37.597 [2024-11-26 23:43:25.535089] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:09:37.597 23:43:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:37.597 23:43:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:37.597 23:43:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:09:37.597 23:43:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:37.597 23:43:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:37.597 23:43:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:37.597 23:43:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:09:37.597 23:43:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:09:37.597 23:43:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:09:37.597 23:43:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:09:37.597 23:43:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:37.597 23:43:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:37.597 23:43:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:37.597 23:43:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:09:37.597 23:43:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:09:37.598 23:43:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:37.598 23:43:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:37.598 23:43:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:37.598 23:43:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:09:37.598 23:43:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt3 00:09:37.598 23:43:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:37.598 23:43:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:37.598 23:43:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:37.598 23:43:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:09:37.598 23:43:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt4 00:09:37.598 23:43:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:37.598 23:43:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:37.598 23:43:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:37.598 23:43:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:09:37.598 23:43:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:09:37.598 23:43:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:37.598 23:43:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:37.598 23:43:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:37.598 23:43:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:09:37.598 23:43:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:09:37.598 23:43:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@652 -- # local es=0 00:09:37.598 23:43:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:09:37.598 23:43:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:09:37.598 23:43:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:09:37.598 23:43:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:09:37.598 23:43:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:09:37.598 23:43:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:09:37.598 23:43:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:37.598 23:43:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:37.598 [2024-11-26 23:43:25.698620] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:09:37.598 [2024-11-26 23:43:25.700456] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:09:37.598 [2024-11-26 23:43:25.700557] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc3 is claimed 00:09:37.598 [2024-11-26 23:43:25.700604] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc4 is claimed 00:09:37.598 [2024-11-26 23:43:25.700678] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:09:37.598 [2024-11-26 23:43:25.700722] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:09:37.598 [2024-11-26 23:43:25.700740] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc3 00:09:37.598 [2024-11-26 23:43:25.700755] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc4 00:09:37.598 [2024-11-26 23:43:25.700768] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:37.598 [2024-11-26 23:43:25.700777] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state configuring 00:09:37.598 request: 00:09:37.598 { 00:09:37.598 "name": "raid_bdev1", 00:09:37.598 "raid_level": "concat", 00:09:37.598 "base_bdevs": [ 00:09:37.598 "malloc1", 00:09:37.598 "malloc2", 00:09:37.598 "malloc3", 00:09:37.598 "malloc4" 00:09:37.598 ], 00:09:37.598 "strip_size_kb": 64, 00:09:37.598 "superblock": false, 00:09:37.598 "method": "bdev_raid_create", 00:09:37.598 "req_id": 1 00:09:37.598 } 00:09:37.598 Got JSON-RPC error response 00:09:37.598 response: 00:09:37.598 { 00:09:37.598 "code": -17, 00:09:37.598 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:09:37.598 } 00:09:37.598 23:43:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:09:37.598 23:43:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # es=1 00:09:37.598 23:43:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:09:37.598 23:43:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:09:37.598 23:43:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:09:37.598 23:43:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:37.598 23:43:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:37.598 23:43:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:37.598 23:43:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:09:37.598 23:43:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:37.857 23:43:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:09:37.857 23:43:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:09:37.857 23:43:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:09:37.857 23:43:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:37.858 23:43:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:37.858 [2024-11-26 23:43:25.766483] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:09:37.858 [2024-11-26 23:43:25.766574] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:37.858 [2024-11-26 23:43:25.766614] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:09:37.858 [2024-11-26 23:43:25.766664] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:37.858 [2024-11-26 23:43:25.768812] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:37.858 [2024-11-26 23:43:25.768879] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:09:37.858 [2024-11-26 23:43:25.768967] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:09:37.858 [2024-11-26 23:43:25.769041] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:09:37.858 pt1 00:09:37.858 23:43:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:37.858 23:43:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring concat 64 4 00:09:37.858 23:43:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:37.858 23:43:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:37.858 23:43:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:37.858 23:43:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:37.858 23:43:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:37.858 23:43:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:37.858 23:43:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:37.858 23:43:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:37.858 23:43:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:37.858 23:43:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:37.858 23:43:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:37.858 23:43:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:37.858 23:43:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:37.858 23:43:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:37.858 23:43:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:37.858 "name": "raid_bdev1", 00:09:37.858 "uuid": "950cb8e7-29ba-4e34-b7be-7453678af9e0", 00:09:37.858 "strip_size_kb": 64, 00:09:37.858 "state": "configuring", 00:09:37.858 "raid_level": "concat", 00:09:37.858 "superblock": true, 00:09:37.858 "num_base_bdevs": 4, 00:09:37.858 "num_base_bdevs_discovered": 1, 00:09:37.858 "num_base_bdevs_operational": 4, 00:09:37.858 "base_bdevs_list": [ 00:09:37.858 { 00:09:37.858 "name": "pt1", 00:09:37.858 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:37.858 "is_configured": true, 00:09:37.858 "data_offset": 2048, 00:09:37.858 "data_size": 63488 00:09:37.858 }, 00:09:37.858 { 00:09:37.858 "name": null, 00:09:37.858 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:37.858 "is_configured": false, 00:09:37.858 "data_offset": 2048, 00:09:37.858 "data_size": 63488 00:09:37.858 }, 00:09:37.858 { 00:09:37.858 "name": null, 00:09:37.858 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:37.858 "is_configured": false, 00:09:37.858 "data_offset": 2048, 00:09:37.858 "data_size": 63488 00:09:37.858 }, 00:09:37.858 { 00:09:37.858 "name": null, 00:09:37.858 "uuid": "00000000-0000-0000-0000-000000000004", 00:09:37.858 "is_configured": false, 00:09:37.858 "data_offset": 2048, 00:09:37.858 "data_size": 63488 00:09:37.858 } 00:09:37.858 ] 00:09:37.858 }' 00:09:37.858 23:43:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:37.858 23:43:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:38.117 23:43:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 4 -gt 2 ']' 00:09:38.117 23:43:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@472 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:09:38.117 23:43:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:38.117 23:43:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:38.117 [2024-11-26 23:43:26.237708] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:09:38.117 [2024-11-26 23:43:26.237777] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:38.117 [2024-11-26 23:43:26.237799] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009c80 00:09:38.117 [2024-11-26 23:43:26.237809] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:38.117 [2024-11-26 23:43:26.238243] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:38.117 [2024-11-26 23:43:26.238263] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:09:38.117 [2024-11-26 23:43:26.238350] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:09:38.117 [2024-11-26 23:43:26.238386] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:09:38.117 pt2 00:09:38.117 23:43:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:38.117 23:43:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@473 -- # rpc_cmd bdev_passthru_delete pt2 00:09:38.117 23:43:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:38.118 23:43:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:38.118 [2024-11-26 23:43:26.245706] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt2 00:09:38.377 23:43:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:38.377 23:43:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@474 -- # verify_raid_bdev_state raid_bdev1 configuring concat 64 4 00:09:38.377 23:43:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:38.377 23:43:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:38.377 23:43:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:38.377 23:43:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:38.377 23:43:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:38.377 23:43:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:38.377 23:43:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:38.377 23:43:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:38.377 23:43:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:38.377 23:43:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:38.377 23:43:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:38.377 23:43:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:38.377 23:43:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:38.377 23:43:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:38.377 23:43:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:38.377 "name": "raid_bdev1", 00:09:38.377 "uuid": "950cb8e7-29ba-4e34-b7be-7453678af9e0", 00:09:38.377 "strip_size_kb": 64, 00:09:38.377 "state": "configuring", 00:09:38.377 "raid_level": "concat", 00:09:38.377 "superblock": true, 00:09:38.377 "num_base_bdevs": 4, 00:09:38.377 "num_base_bdevs_discovered": 1, 00:09:38.377 "num_base_bdevs_operational": 4, 00:09:38.377 "base_bdevs_list": [ 00:09:38.377 { 00:09:38.377 "name": "pt1", 00:09:38.377 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:38.377 "is_configured": true, 00:09:38.377 "data_offset": 2048, 00:09:38.377 "data_size": 63488 00:09:38.377 }, 00:09:38.377 { 00:09:38.377 "name": null, 00:09:38.377 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:38.377 "is_configured": false, 00:09:38.377 "data_offset": 0, 00:09:38.377 "data_size": 63488 00:09:38.377 }, 00:09:38.377 { 00:09:38.377 "name": null, 00:09:38.377 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:38.377 "is_configured": false, 00:09:38.377 "data_offset": 2048, 00:09:38.377 "data_size": 63488 00:09:38.377 }, 00:09:38.377 { 00:09:38.377 "name": null, 00:09:38.377 "uuid": "00000000-0000-0000-0000-000000000004", 00:09:38.377 "is_configured": false, 00:09:38.377 "data_offset": 2048, 00:09:38.377 "data_size": 63488 00:09:38.377 } 00:09:38.377 ] 00:09:38.377 }' 00:09:38.377 23:43:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:38.377 23:43:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:38.635 23:43:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:09:38.635 23:43:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:09:38.635 23:43:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:09:38.635 23:43:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:38.635 23:43:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:38.636 [2024-11-26 23:43:26.645059] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:09:38.636 [2024-11-26 23:43:26.645183] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:38.636 [2024-11-26 23:43:26.645218] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009f80 00:09:38.636 [2024-11-26 23:43:26.645252] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:38.636 [2024-11-26 23:43:26.645677] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:38.636 [2024-11-26 23:43:26.645742] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:09:38.636 [2024-11-26 23:43:26.645847] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:09:38.636 [2024-11-26 23:43:26.645909] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:09:38.636 pt2 00:09:38.636 23:43:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:38.636 23:43:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:09:38.636 23:43:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:09:38.636 23:43:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:09:38.636 23:43:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:38.636 23:43:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:38.636 [2024-11-26 23:43:26.656983] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:09:38.636 [2024-11-26 23:43:26.657068] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:38.636 [2024-11-26 23:43:26.657098] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:09:38.636 [2024-11-26 23:43:26.657128] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:38.636 [2024-11-26 23:43:26.657518] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:38.636 [2024-11-26 23:43:26.657580] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:09:38.636 [2024-11-26 23:43:26.657658] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:09:38.636 [2024-11-26 23:43:26.657705] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:09:38.636 pt3 00:09:38.636 23:43:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:38.636 23:43:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:09:38.636 23:43:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:09:38.636 23:43:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:09:38.636 23:43:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:38.636 23:43:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:38.636 [2024-11-26 23:43:26.668980] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:09:38.636 [2024-11-26 23:43:26.669027] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:38.636 [2024-11-26 23:43:26.669040] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a580 00:09:38.636 [2024-11-26 23:43:26.669048] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:38.636 [2024-11-26 23:43:26.669326] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:38.636 [2024-11-26 23:43:26.669362] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:09:38.636 [2024-11-26 23:43:26.669428] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt4 00:09:38.636 [2024-11-26 23:43:26.669447] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:09:38.636 [2024-11-26 23:43:26.669541] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:09:38.636 [2024-11-26 23:43:26.669551] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:09:38.636 [2024-11-26 23:43:26.669765] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:09:38.636 [2024-11-26 23:43:26.669884] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:09:38.636 [2024-11-26 23:43:26.669892] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:09:38.636 [2024-11-26 23:43:26.669983] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:38.636 pt4 00:09:38.636 23:43:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:38.636 23:43:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:09:38.636 23:43:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:09:38.636 23:43:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online concat 64 4 00:09:38.636 23:43:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:38.636 23:43:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:38.636 23:43:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:38.636 23:43:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:38.636 23:43:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:38.636 23:43:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:38.636 23:43:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:38.636 23:43:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:38.636 23:43:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:38.636 23:43:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:38.636 23:43:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:38.636 23:43:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:38.636 23:43:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:38.636 23:43:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:38.636 23:43:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:38.636 "name": "raid_bdev1", 00:09:38.636 "uuid": "950cb8e7-29ba-4e34-b7be-7453678af9e0", 00:09:38.636 "strip_size_kb": 64, 00:09:38.636 "state": "online", 00:09:38.636 "raid_level": "concat", 00:09:38.636 "superblock": true, 00:09:38.636 "num_base_bdevs": 4, 00:09:38.636 "num_base_bdevs_discovered": 4, 00:09:38.636 "num_base_bdevs_operational": 4, 00:09:38.636 "base_bdevs_list": [ 00:09:38.636 { 00:09:38.636 "name": "pt1", 00:09:38.636 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:38.636 "is_configured": true, 00:09:38.636 "data_offset": 2048, 00:09:38.636 "data_size": 63488 00:09:38.636 }, 00:09:38.636 { 00:09:38.636 "name": "pt2", 00:09:38.636 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:38.636 "is_configured": true, 00:09:38.636 "data_offset": 2048, 00:09:38.636 "data_size": 63488 00:09:38.636 }, 00:09:38.636 { 00:09:38.636 "name": "pt3", 00:09:38.636 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:38.636 "is_configured": true, 00:09:38.636 "data_offset": 2048, 00:09:38.636 "data_size": 63488 00:09:38.636 }, 00:09:38.636 { 00:09:38.636 "name": "pt4", 00:09:38.636 "uuid": "00000000-0000-0000-0000-000000000004", 00:09:38.636 "is_configured": true, 00:09:38.636 "data_offset": 2048, 00:09:38.636 "data_size": 63488 00:09:38.636 } 00:09:38.636 ] 00:09:38.636 }' 00:09:38.636 23:43:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:38.636 23:43:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:39.203 23:43:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:09:39.203 23:43:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:09:39.203 23:43:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:39.203 23:43:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:39.203 23:43:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:09:39.203 23:43:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:39.203 23:43:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:09:39.203 23:43:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:39.203 23:43:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:39.203 23:43:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:39.203 [2024-11-26 23:43:27.144535] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:39.203 23:43:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:39.203 23:43:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:39.203 "name": "raid_bdev1", 00:09:39.203 "aliases": [ 00:09:39.203 "950cb8e7-29ba-4e34-b7be-7453678af9e0" 00:09:39.203 ], 00:09:39.203 "product_name": "Raid Volume", 00:09:39.203 "block_size": 512, 00:09:39.203 "num_blocks": 253952, 00:09:39.203 "uuid": "950cb8e7-29ba-4e34-b7be-7453678af9e0", 00:09:39.203 "assigned_rate_limits": { 00:09:39.203 "rw_ios_per_sec": 0, 00:09:39.203 "rw_mbytes_per_sec": 0, 00:09:39.203 "r_mbytes_per_sec": 0, 00:09:39.203 "w_mbytes_per_sec": 0 00:09:39.203 }, 00:09:39.203 "claimed": false, 00:09:39.203 "zoned": false, 00:09:39.203 "supported_io_types": { 00:09:39.203 "read": true, 00:09:39.203 "write": true, 00:09:39.203 "unmap": true, 00:09:39.203 "flush": true, 00:09:39.203 "reset": true, 00:09:39.203 "nvme_admin": false, 00:09:39.203 "nvme_io": false, 00:09:39.203 "nvme_io_md": false, 00:09:39.203 "write_zeroes": true, 00:09:39.203 "zcopy": false, 00:09:39.204 "get_zone_info": false, 00:09:39.204 "zone_management": false, 00:09:39.204 "zone_append": false, 00:09:39.204 "compare": false, 00:09:39.204 "compare_and_write": false, 00:09:39.204 "abort": false, 00:09:39.204 "seek_hole": false, 00:09:39.204 "seek_data": false, 00:09:39.204 "copy": false, 00:09:39.204 "nvme_iov_md": false 00:09:39.204 }, 00:09:39.204 "memory_domains": [ 00:09:39.204 { 00:09:39.204 "dma_device_id": "system", 00:09:39.204 "dma_device_type": 1 00:09:39.204 }, 00:09:39.204 { 00:09:39.204 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:39.204 "dma_device_type": 2 00:09:39.204 }, 00:09:39.204 { 00:09:39.204 "dma_device_id": "system", 00:09:39.204 "dma_device_type": 1 00:09:39.204 }, 00:09:39.204 { 00:09:39.204 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:39.204 "dma_device_type": 2 00:09:39.204 }, 00:09:39.204 { 00:09:39.204 "dma_device_id": "system", 00:09:39.204 "dma_device_type": 1 00:09:39.204 }, 00:09:39.204 { 00:09:39.204 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:39.204 "dma_device_type": 2 00:09:39.204 }, 00:09:39.204 { 00:09:39.204 "dma_device_id": "system", 00:09:39.204 "dma_device_type": 1 00:09:39.204 }, 00:09:39.204 { 00:09:39.204 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:39.204 "dma_device_type": 2 00:09:39.204 } 00:09:39.204 ], 00:09:39.204 "driver_specific": { 00:09:39.204 "raid": { 00:09:39.204 "uuid": "950cb8e7-29ba-4e34-b7be-7453678af9e0", 00:09:39.204 "strip_size_kb": 64, 00:09:39.204 "state": "online", 00:09:39.204 "raid_level": "concat", 00:09:39.204 "superblock": true, 00:09:39.204 "num_base_bdevs": 4, 00:09:39.204 "num_base_bdevs_discovered": 4, 00:09:39.204 "num_base_bdevs_operational": 4, 00:09:39.204 "base_bdevs_list": [ 00:09:39.204 { 00:09:39.204 "name": "pt1", 00:09:39.204 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:39.204 "is_configured": true, 00:09:39.204 "data_offset": 2048, 00:09:39.204 "data_size": 63488 00:09:39.204 }, 00:09:39.204 { 00:09:39.204 "name": "pt2", 00:09:39.204 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:39.204 "is_configured": true, 00:09:39.204 "data_offset": 2048, 00:09:39.204 "data_size": 63488 00:09:39.204 }, 00:09:39.204 { 00:09:39.204 "name": "pt3", 00:09:39.204 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:39.204 "is_configured": true, 00:09:39.204 "data_offset": 2048, 00:09:39.204 "data_size": 63488 00:09:39.204 }, 00:09:39.204 { 00:09:39.204 "name": "pt4", 00:09:39.204 "uuid": "00000000-0000-0000-0000-000000000004", 00:09:39.204 "is_configured": true, 00:09:39.204 "data_offset": 2048, 00:09:39.204 "data_size": 63488 00:09:39.204 } 00:09:39.204 ] 00:09:39.204 } 00:09:39.204 } 00:09:39.204 }' 00:09:39.204 23:43:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:39.204 23:43:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:09:39.204 pt2 00:09:39.204 pt3 00:09:39.204 pt4' 00:09:39.204 23:43:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:39.204 23:43:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:39.204 23:43:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:39.204 23:43:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:09:39.204 23:43:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:39.204 23:43:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:39.204 23:43:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:39.204 23:43:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:39.204 23:43:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:39.204 23:43:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:39.204 23:43:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:39.204 23:43:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:09:39.204 23:43:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:39.204 23:43:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:39.204 23:43:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:39.462 23:43:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:39.462 23:43:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:39.462 23:43:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:39.462 23:43:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:39.462 23:43:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:39.462 23:43:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:09:39.462 23:43:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:39.462 23:43:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:39.462 23:43:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:39.462 23:43:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:39.462 23:43:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:39.462 23:43:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:39.462 23:43:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt4 00:09:39.462 23:43:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:39.462 23:43:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:39.462 23:43:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:39.462 23:43:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:39.462 23:43:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:39.462 23:43:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:39.462 23:43:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:09:39.462 23:43:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:09:39.462 23:43:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:39.462 23:43:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:39.462 [2024-11-26 23:43:27.455905] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:39.462 23:43:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:39.462 23:43:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' 950cb8e7-29ba-4e34-b7be-7453678af9e0 '!=' 950cb8e7-29ba-4e34-b7be-7453678af9e0 ']' 00:09:39.462 23:43:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy concat 00:09:39.462 23:43:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:09:39.462 23:43:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # return 1 00:09:39.462 23:43:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 83175 00:09:39.462 23:43:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # '[' -z 83175 ']' 00:09:39.462 23:43:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@958 -- # kill -0 83175 00:09:39.462 23:43:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # uname 00:09:39.462 23:43:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:09:39.462 23:43:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 83175 00:09:39.462 23:43:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:09:39.462 killing process with pid 83175 00:09:39.462 23:43:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:09:39.462 23:43:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 83175' 00:09:39.462 23:43:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@973 -- # kill 83175 00:09:39.462 [2024-11-26 23:43:27.524631] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:09:39.462 [2024-11-26 23:43:27.524722] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:39.462 [2024-11-26 23:43:27.524788] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:39.462 [2024-11-26 23:43:27.524800] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:09:39.462 23:43:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@978 -- # wait 83175 00:09:39.462 [2024-11-26 23:43:27.567991] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:09:39.721 23:43:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:09:39.721 00:09:39.721 real 0m4.186s 00:09:39.721 user 0m6.657s 00:09:39.721 sys 0m0.875s 00:09:39.721 23:43:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:09:39.721 ************************************ 00:09:39.721 END TEST raid_superblock_test 00:09:39.721 23:43:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:39.721 ************************************ 00:09:39.721 23:43:27 bdev_raid -- bdev/bdev_raid.sh@971 -- # run_test raid_read_error_test raid_io_error_test concat 4 read 00:09:39.721 23:43:27 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:09:39.721 23:43:27 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:09:39.721 23:43:27 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:09:39.721 ************************************ 00:09:39.721 START TEST raid_read_error_test 00:09:39.721 ************************************ 00:09:39.721 23:43:27 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test concat 4 read 00:09:39.721 23:43:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=concat 00:09:39.721 23:43:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=4 00:09:39.721 23:43:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=read 00:09:39.721 23:43:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:09:39.721 23:43:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:39.721 23:43:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:09:39.721 23:43:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:39.721 23:43:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:39.721 23:43:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:09:39.721 23:43:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:39.721 23:43:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:39.721 23:43:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:09:39.721 23:43:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:39.721 23:43:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:39.722 23:43:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev4 00:09:39.722 23:43:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:39.722 23:43:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:39.722 23:43:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:09:39.722 23:43:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:09:39.722 23:43:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:09:39.722 23:43:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:09:39.722 23:43:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:09:39.722 23:43:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:09:39.722 23:43:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:09:39.722 23:43:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # '[' concat '!=' raid1 ']' 00:09:39.722 23:43:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:09:39.980 23:43:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:09:39.980 23:43:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:09:39.980 23:43:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.8o9y1fLthF 00:09:39.980 23:43:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=83422 00:09:39.980 23:43:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:09:39.980 23:43:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 83422 00:09:39.980 23:43:27 bdev_raid.raid_read_error_test -- common/autotest_common.sh@835 -- # '[' -z 83422 ']' 00:09:39.980 23:43:27 bdev_raid.raid_read_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:39.980 23:43:27 bdev_raid.raid_read_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:09:39.980 23:43:27 bdev_raid.raid_read_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:39.980 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:39.980 23:43:27 bdev_raid.raid_read_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:09:39.980 23:43:27 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:39.981 [2024-11-26 23:43:27.940720] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:09:39.981 [2024-11-26 23:43:27.940833] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid83422 ] 00:09:39.981 [2024-11-26 23:43:28.096691] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:40.240 [2024-11-26 23:43:28.122011] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:09:40.240 [2024-11-26 23:43:28.164445] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:40.240 [2024-11-26 23:43:28.164486] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:40.808 23:43:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:09:40.808 23:43:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@868 -- # return 0 00:09:40.808 23:43:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:40.808 23:43:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:09:40.808 23:43:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:40.808 23:43:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:40.808 BaseBdev1_malloc 00:09:40.808 23:43:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:40.808 23:43:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:09:40.808 23:43:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:40.808 23:43:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:40.808 true 00:09:40.808 23:43:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:40.808 23:43:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:09:40.808 23:43:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:40.808 23:43:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:40.808 [2024-11-26 23:43:28.799881] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:09:40.808 [2024-11-26 23:43:28.799982] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:40.808 [2024-11-26 23:43:28.800036] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006980 00:09:40.808 [2024-11-26 23:43:28.800064] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:40.808 [2024-11-26 23:43:28.802134] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:40.808 [2024-11-26 23:43:28.802212] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:09:40.808 BaseBdev1 00:09:40.808 23:43:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:40.808 23:43:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:40.808 23:43:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:09:40.808 23:43:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:40.808 23:43:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:40.808 BaseBdev2_malloc 00:09:40.808 23:43:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:40.808 23:43:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:09:40.808 23:43:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:40.808 23:43:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:40.808 true 00:09:40.808 23:43:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:40.808 23:43:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:09:40.808 23:43:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:40.808 23:43:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:40.808 [2024-11-26 23:43:28.840341] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:09:40.808 [2024-11-26 23:43:28.840400] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:40.808 [2024-11-26 23:43:28.840418] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:09:40.808 [2024-11-26 23:43:28.840450] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:40.808 [2024-11-26 23:43:28.842463] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:40.808 [2024-11-26 23:43:28.842501] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:09:40.808 BaseBdev2 00:09:40.808 23:43:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:40.808 23:43:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:40.808 23:43:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:09:40.808 23:43:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:40.808 23:43:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:40.808 BaseBdev3_malloc 00:09:40.808 23:43:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:40.808 23:43:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:09:40.808 23:43:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:40.808 23:43:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:40.808 true 00:09:40.808 23:43:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:40.808 23:43:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:09:40.808 23:43:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:40.808 23:43:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:40.808 [2024-11-26 23:43:28.880842] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:09:40.808 [2024-11-26 23:43:28.880891] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:40.808 [2024-11-26 23:43:28.880908] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008780 00:09:40.808 [2024-11-26 23:43:28.880916] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:40.808 [2024-11-26 23:43:28.882946] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:40.808 [2024-11-26 23:43:28.883035] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:09:40.808 BaseBdev3 00:09:40.808 23:43:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:40.808 23:43:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:40.808 23:43:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:09:40.808 23:43:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:40.808 23:43:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:40.808 BaseBdev4_malloc 00:09:40.808 23:43:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:40.808 23:43:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev4_malloc 00:09:40.808 23:43:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:40.808 23:43:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:41.067 true 00:09:41.067 23:43:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:41.067 23:43:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev4_malloc -p BaseBdev4 00:09:41.067 23:43:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:41.067 23:43:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:41.067 [2024-11-26 23:43:28.943416] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev4_malloc 00:09:41.067 [2024-11-26 23:43:28.943468] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:41.067 [2024-11-26 23:43:28.943491] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:09:41.067 [2024-11-26 23:43:28.943501] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:41.067 [2024-11-26 23:43:28.945771] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:41.067 [2024-11-26 23:43:28.945850] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:09:41.067 BaseBdev4 00:09:41.067 23:43:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:41.067 23:43:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 -s 00:09:41.067 23:43:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:41.067 23:43:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:41.067 [2024-11-26 23:43:28.955427] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:41.067 [2024-11-26 23:43:28.957176] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:41.067 [2024-11-26 23:43:28.957253] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:41.067 [2024-11-26 23:43:28.957304] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:09:41.067 [2024-11-26 23:43:28.957510] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000002000 00:09:41.067 [2024-11-26 23:43:28.957522] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:09:41.067 [2024-11-26 23:43:28.957777] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002ef0 00:09:41.067 [2024-11-26 23:43:28.957910] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000002000 00:09:41.067 [2024-11-26 23:43:28.957922] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000002000 00:09:41.067 [2024-11-26 23:43:28.958055] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:41.067 23:43:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:41.067 23:43:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online concat 64 4 00:09:41.067 23:43:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:41.067 23:43:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:41.067 23:43:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:41.067 23:43:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:41.067 23:43:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:41.067 23:43:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:41.067 23:43:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:41.067 23:43:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:41.067 23:43:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:41.067 23:43:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:41.067 23:43:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:41.067 23:43:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:41.067 23:43:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:41.067 23:43:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:41.067 23:43:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:41.067 "name": "raid_bdev1", 00:09:41.067 "uuid": "be50fa0c-a23f-4d04-92b9-04aca713ca96", 00:09:41.067 "strip_size_kb": 64, 00:09:41.067 "state": "online", 00:09:41.067 "raid_level": "concat", 00:09:41.067 "superblock": true, 00:09:41.067 "num_base_bdevs": 4, 00:09:41.067 "num_base_bdevs_discovered": 4, 00:09:41.067 "num_base_bdevs_operational": 4, 00:09:41.067 "base_bdevs_list": [ 00:09:41.067 { 00:09:41.067 "name": "BaseBdev1", 00:09:41.067 "uuid": "1d058525-6b8e-56d4-9c78-d43780ecfd93", 00:09:41.067 "is_configured": true, 00:09:41.067 "data_offset": 2048, 00:09:41.067 "data_size": 63488 00:09:41.067 }, 00:09:41.067 { 00:09:41.067 "name": "BaseBdev2", 00:09:41.067 "uuid": "1b6f15e0-1c10-5273-a159-eeb4896f4b6d", 00:09:41.068 "is_configured": true, 00:09:41.068 "data_offset": 2048, 00:09:41.068 "data_size": 63488 00:09:41.068 }, 00:09:41.068 { 00:09:41.068 "name": "BaseBdev3", 00:09:41.068 "uuid": "fdb86441-60d3-5af9-9ac5-ce6010cdfb5a", 00:09:41.068 "is_configured": true, 00:09:41.068 "data_offset": 2048, 00:09:41.068 "data_size": 63488 00:09:41.068 }, 00:09:41.068 { 00:09:41.068 "name": "BaseBdev4", 00:09:41.068 "uuid": "2530b04d-9a6c-5ec5-8579-278917e04a44", 00:09:41.068 "is_configured": true, 00:09:41.068 "data_offset": 2048, 00:09:41.068 "data_size": 63488 00:09:41.068 } 00:09:41.068 ] 00:09:41.068 }' 00:09:41.068 23:43:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:41.068 23:43:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:41.325 23:43:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:09:41.325 23:43:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:09:41.584 [2024-11-26 23:43:29.522857] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000003090 00:09:42.519 23:43:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:09:42.519 23:43:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:42.519 23:43:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:42.519 23:43:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:42.519 23:43:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:09:42.519 23:43:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ concat = \r\a\i\d\1 ]] 00:09:42.519 23:43:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=4 00:09:42.519 23:43:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online concat 64 4 00:09:42.519 23:43:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:42.519 23:43:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:42.519 23:43:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:42.519 23:43:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:42.519 23:43:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:42.519 23:43:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:42.519 23:43:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:42.519 23:43:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:42.519 23:43:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:42.519 23:43:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:42.519 23:43:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:42.519 23:43:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:42.519 23:43:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:42.519 23:43:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:42.519 23:43:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:42.519 "name": "raid_bdev1", 00:09:42.519 "uuid": "be50fa0c-a23f-4d04-92b9-04aca713ca96", 00:09:42.519 "strip_size_kb": 64, 00:09:42.519 "state": "online", 00:09:42.519 "raid_level": "concat", 00:09:42.519 "superblock": true, 00:09:42.519 "num_base_bdevs": 4, 00:09:42.519 "num_base_bdevs_discovered": 4, 00:09:42.519 "num_base_bdevs_operational": 4, 00:09:42.519 "base_bdevs_list": [ 00:09:42.519 { 00:09:42.519 "name": "BaseBdev1", 00:09:42.519 "uuid": "1d058525-6b8e-56d4-9c78-d43780ecfd93", 00:09:42.519 "is_configured": true, 00:09:42.519 "data_offset": 2048, 00:09:42.519 "data_size": 63488 00:09:42.519 }, 00:09:42.519 { 00:09:42.519 "name": "BaseBdev2", 00:09:42.519 "uuid": "1b6f15e0-1c10-5273-a159-eeb4896f4b6d", 00:09:42.519 "is_configured": true, 00:09:42.519 "data_offset": 2048, 00:09:42.519 "data_size": 63488 00:09:42.519 }, 00:09:42.519 { 00:09:42.519 "name": "BaseBdev3", 00:09:42.519 "uuid": "fdb86441-60d3-5af9-9ac5-ce6010cdfb5a", 00:09:42.519 "is_configured": true, 00:09:42.519 "data_offset": 2048, 00:09:42.519 "data_size": 63488 00:09:42.519 }, 00:09:42.519 { 00:09:42.519 "name": "BaseBdev4", 00:09:42.519 "uuid": "2530b04d-9a6c-5ec5-8579-278917e04a44", 00:09:42.519 "is_configured": true, 00:09:42.519 "data_offset": 2048, 00:09:42.519 "data_size": 63488 00:09:42.519 } 00:09:42.519 ] 00:09:42.519 }' 00:09:42.519 23:43:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:42.519 23:43:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:42.778 23:43:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:09:42.778 23:43:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:42.778 23:43:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:42.778 [2024-11-26 23:43:30.866539] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:42.778 [2024-11-26 23:43:30.866615] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:42.778 [2024-11-26 23:43:30.869089] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:42.778 [2024-11-26 23:43:30.869194] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:42.778 [2024-11-26 23:43:30.869260] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:42.778 [2024-11-26 23:43:30.869313] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002000 name raid_bdev1, state offline 00:09:42.778 { 00:09:42.778 "results": [ 00:09:42.778 { 00:09:42.778 "job": "raid_bdev1", 00:09:42.778 "core_mask": "0x1", 00:09:42.778 "workload": "randrw", 00:09:42.778 "percentage": 50, 00:09:42.778 "status": "finished", 00:09:42.778 "queue_depth": 1, 00:09:42.778 "io_size": 131072, 00:09:42.778 "runtime": 1.34451, 00:09:42.778 "iops": 16419.36467560673, 00:09:42.778 "mibps": 2052.420584450841, 00:09:42.778 "io_failed": 1, 00:09:42.778 "io_timeout": 0, 00:09:42.778 "avg_latency_us": 84.11076610980265, 00:09:42.778 "min_latency_us": 25.3764192139738, 00:09:42.778 "max_latency_us": 1380.8349344978167 00:09:42.778 } 00:09:42.778 ], 00:09:42.778 "core_count": 1 00:09:42.778 } 00:09:42.778 23:43:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:42.778 23:43:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 83422 00:09:42.778 23:43:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # '[' -z 83422 ']' 00:09:42.778 23:43:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@958 -- # kill -0 83422 00:09:42.778 23:43:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # uname 00:09:42.778 23:43:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:09:42.778 23:43:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 83422 00:09:43.037 23:43:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:09:43.037 23:43:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:09:43.037 23:43:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 83422' 00:09:43.037 killing process with pid 83422 00:09:43.037 23:43:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@973 -- # kill 83422 00:09:43.037 [2024-11-26 23:43:30.916814] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:09:43.037 23:43:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@978 -- # wait 83422 00:09:43.037 [2024-11-26 23:43:30.952743] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:09:43.037 23:43:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.8o9y1fLthF 00:09:43.037 23:43:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:09:43.037 23:43:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:09:43.037 23:43:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.74 00:09:43.037 23:43:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy concat 00:09:43.037 23:43:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:09:43.037 23:43:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:09:43.037 23:43:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.74 != \0\.\0\0 ]] 00:09:43.037 00:09:43.037 real 0m3.329s 00:09:43.037 user 0m4.204s 00:09:43.037 sys 0m0.524s 00:09:43.296 23:43:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:09:43.296 23:43:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:43.296 ************************************ 00:09:43.296 END TEST raid_read_error_test 00:09:43.296 ************************************ 00:09:43.296 23:43:31 bdev_raid -- bdev/bdev_raid.sh@972 -- # run_test raid_write_error_test raid_io_error_test concat 4 write 00:09:43.296 23:43:31 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:09:43.296 23:43:31 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:09:43.296 23:43:31 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:09:43.296 ************************************ 00:09:43.296 START TEST raid_write_error_test 00:09:43.296 ************************************ 00:09:43.296 23:43:31 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test concat 4 write 00:09:43.296 23:43:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=concat 00:09:43.296 23:43:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=4 00:09:43.296 23:43:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=write 00:09:43.296 23:43:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:09:43.296 23:43:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:43.296 23:43:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:09:43.296 23:43:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:43.296 23:43:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:43.296 23:43:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:09:43.296 23:43:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:43.296 23:43:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:43.296 23:43:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:09:43.297 23:43:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:43.297 23:43:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:43.297 23:43:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev4 00:09:43.297 23:43:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:43.297 23:43:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:43.297 23:43:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:09:43.297 23:43:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:09:43.297 23:43:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:09:43.297 23:43:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:09:43.297 23:43:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:09:43.297 23:43:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:09:43.297 23:43:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:09:43.297 23:43:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # '[' concat '!=' raid1 ']' 00:09:43.297 23:43:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:09:43.297 23:43:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:09:43.297 23:43:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:09:43.297 23:43:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.0gt49KbknK 00:09:43.297 23:43:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=83563 00:09:43.297 23:43:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:09:43.297 23:43:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 83563 00:09:43.297 23:43:31 bdev_raid.raid_write_error_test -- common/autotest_common.sh@835 -- # '[' -z 83563 ']' 00:09:43.297 23:43:31 bdev_raid.raid_write_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:43.297 23:43:31 bdev_raid.raid_write_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:09:43.297 23:43:31 bdev_raid.raid_write_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:43.297 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:43.297 23:43:31 bdev_raid.raid_write_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:09:43.297 23:43:31 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:43.297 [2024-11-26 23:43:31.337454] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:09:43.297 [2024-11-26 23:43:31.337587] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid83563 ] 00:09:43.556 [2024-11-26 23:43:31.491780] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:43.556 [2024-11-26 23:43:31.516236] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:09:43.556 [2024-11-26 23:43:31.558833] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:43.556 [2024-11-26 23:43:31.558867] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:44.128 23:43:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:09:44.128 23:43:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@868 -- # return 0 00:09:44.128 23:43:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:44.128 23:43:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:09:44.128 23:43:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:44.128 23:43:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:44.128 BaseBdev1_malloc 00:09:44.128 23:43:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:44.128 23:43:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:09:44.128 23:43:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:44.128 23:43:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:44.128 true 00:09:44.128 23:43:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:44.128 23:43:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:09:44.128 23:43:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:44.128 23:43:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:44.128 [2024-11-26 23:43:32.197977] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:09:44.128 [2024-11-26 23:43:32.198029] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:44.128 [2024-11-26 23:43:32.198049] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006980 00:09:44.128 [2024-11-26 23:43:32.198064] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:44.128 [2024-11-26 23:43:32.200162] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:44.128 [2024-11-26 23:43:32.200200] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:09:44.128 BaseBdev1 00:09:44.128 23:43:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:44.128 23:43:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:44.128 23:43:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:09:44.128 23:43:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:44.128 23:43:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:44.128 BaseBdev2_malloc 00:09:44.128 23:43:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:44.128 23:43:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:09:44.128 23:43:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:44.128 23:43:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:44.128 true 00:09:44.128 23:43:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:44.128 23:43:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:09:44.128 23:43:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:44.128 23:43:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:44.128 [2024-11-26 23:43:32.238455] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:09:44.128 [2024-11-26 23:43:32.238503] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:44.128 [2024-11-26 23:43:32.238520] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:09:44.128 [2024-11-26 23:43:32.238536] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:44.128 [2024-11-26 23:43:32.240646] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:44.128 [2024-11-26 23:43:32.240681] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:09:44.128 BaseBdev2 00:09:44.128 23:43:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:44.128 23:43:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:44.128 23:43:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:09:44.128 23:43:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:44.128 23:43:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:44.392 BaseBdev3_malloc 00:09:44.392 23:43:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:44.392 23:43:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:09:44.392 23:43:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:44.392 23:43:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:44.392 true 00:09:44.392 23:43:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:44.392 23:43:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:09:44.392 23:43:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:44.392 23:43:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:44.392 [2024-11-26 23:43:32.279054] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:09:44.392 [2024-11-26 23:43:32.279146] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:44.392 [2024-11-26 23:43:32.279170] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008780 00:09:44.392 [2024-11-26 23:43:32.279179] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:44.392 [2024-11-26 23:43:32.281239] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:44.392 [2024-11-26 23:43:32.281275] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:09:44.392 BaseBdev3 00:09:44.392 23:43:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:44.392 23:43:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:44.392 23:43:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:09:44.392 23:43:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:44.392 23:43:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:44.392 BaseBdev4_malloc 00:09:44.392 23:43:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:44.392 23:43:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev4_malloc 00:09:44.392 23:43:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:44.392 23:43:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:44.392 true 00:09:44.392 23:43:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:44.393 23:43:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev4_malloc -p BaseBdev4 00:09:44.393 23:43:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:44.393 23:43:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:44.393 [2024-11-26 23:43:32.335260] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev4_malloc 00:09:44.393 [2024-11-26 23:43:32.335358] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:44.393 [2024-11-26 23:43:32.335396] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:09:44.393 [2024-11-26 23:43:32.335428] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:44.393 [2024-11-26 23:43:32.337398] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:44.393 [2024-11-26 23:43:32.337461] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:09:44.393 BaseBdev4 00:09:44.393 23:43:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:44.393 23:43:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 -s 00:09:44.393 23:43:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:44.393 23:43:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:44.393 [2024-11-26 23:43:32.347277] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:44.393 [2024-11-26 23:43:32.349082] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:44.393 [2024-11-26 23:43:32.349194] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:44.393 [2024-11-26 23:43:32.349264] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:09:44.393 [2024-11-26 23:43:32.349476] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000002000 00:09:44.393 [2024-11-26 23:43:32.349495] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:09:44.393 [2024-11-26 23:43:32.349733] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002ef0 00:09:44.393 [2024-11-26 23:43:32.349865] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000002000 00:09:44.393 [2024-11-26 23:43:32.349877] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000002000 00:09:44.393 [2024-11-26 23:43:32.349984] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:44.393 23:43:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:44.393 23:43:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online concat 64 4 00:09:44.393 23:43:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:44.393 23:43:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:44.393 23:43:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:44.393 23:43:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:44.393 23:43:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:44.393 23:43:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:44.393 23:43:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:44.393 23:43:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:44.393 23:43:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:44.393 23:43:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:44.393 23:43:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:44.393 23:43:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:44.393 23:43:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:44.393 23:43:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:44.393 23:43:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:44.393 "name": "raid_bdev1", 00:09:44.393 "uuid": "f82ed3af-5b8c-42ad-8069-a6076c15f02f", 00:09:44.393 "strip_size_kb": 64, 00:09:44.393 "state": "online", 00:09:44.393 "raid_level": "concat", 00:09:44.393 "superblock": true, 00:09:44.393 "num_base_bdevs": 4, 00:09:44.393 "num_base_bdevs_discovered": 4, 00:09:44.393 "num_base_bdevs_operational": 4, 00:09:44.393 "base_bdevs_list": [ 00:09:44.393 { 00:09:44.393 "name": "BaseBdev1", 00:09:44.393 "uuid": "cd8d3e0e-3765-58dc-94bb-a72438c3d90b", 00:09:44.393 "is_configured": true, 00:09:44.393 "data_offset": 2048, 00:09:44.393 "data_size": 63488 00:09:44.393 }, 00:09:44.393 { 00:09:44.393 "name": "BaseBdev2", 00:09:44.393 "uuid": "0f6a5c67-2b5d-5398-b0a6-b6f39abb8187", 00:09:44.393 "is_configured": true, 00:09:44.393 "data_offset": 2048, 00:09:44.393 "data_size": 63488 00:09:44.393 }, 00:09:44.393 { 00:09:44.393 "name": "BaseBdev3", 00:09:44.393 "uuid": "d6590c68-65e0-515b-a566-634f34c9045b", 00:09:44.393 "is_configured": true, 00:09:44.393 "data_offset": 2048, 00:09:44.393 "data_size": 63488 00:09:44.393 }, 00:09:44.393 { 00:09:44.393 "name": "BaseBdev4", 00:09:44.393 "uuid": "b1810edb-2927-5520-9c5e-28dce60a1e5d", 00:09:44.393 "is_configured": true, 00:09:44.393 "data_offset": 2048, 00:09:44.393 "data_size": 63488 00:09:44.393 } 00:09:44.393 ] 00:09:44.393 }' 00:09:44.393 23:43:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:44.393 23:43:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:44.979 23:43:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:09:44.979 23:43:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:09:44.979 [2024-11-26 23:43:32.914767] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000003090 00:09:45.914 23:43:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:09:45.914 23:43:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:45.914 23:43:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:45.914 23:43:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:45.914 23:43:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:09:45.914 23:43:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ concat = \r\a\i\d\1 ]] 00:09:45.914 23:43:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=4 00:09:45.914 23:43:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online concat 64 4 00:09:45.914 23:43:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:45.914 23:43:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:45.914 23:43:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:45.914 23:43:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:45.914 23:43:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:45.914 23:43:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:45.914 23:43:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:45.914 23:43:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:45.914 23:43:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:45.914 23:43:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:45.914 23:43:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:45.914 23:43:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:45.914 23:43:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:45.914 23:43:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:45.914 23:43:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:45.914 "name": "raid_bdev1", 00:09:45.914 "uuid": "f82ed3af-5b8c-42ad-8069-a6076c15f02f", 00:09:45.914 "strip_size_kb": 64, 00:09:45.914 "state": "online", 00:09:45.914 "raid_level": "concat", 00:09:45.914 "superblock": true, 00:09:45.914 "num_base_bdevs": 4, 00:09:45.914 "num_base_bdevs_discovered": 4, 00:09:45.914 "num_base_bdevs_operational": 4, 00:09:45.914 "base_bdevs_list": [ 00:09:45.914 { 00:09:45.914 "name": "BaseBdev1", 00:09:45.914 "uuid": "cd8d3e0e-3765-58dc-94bb-a72438c3d90b", 00:09:45.914 "is_configured": true, 00:09:45.914 "data_offset": 2048, 00:09:45.914 "data_size": 63488 00:09:45.914 }, 00:09:45.914 { 00:09:45.914 "name": "BaseBdev2", 00:09:45.914 "uuid": "0f6a5c67-2b5d-5398-b0a6-b6f39abb8187", 00:09:45.914 "is_configured": true, 00:09:45.914 "data_offset": 2048, 00:09:45.914 "data_size": 63488 00:09:45.914 }, 00:09:45.914 { 00:09:45.914 "name": "BaseBdev3", 00:09:45.914 "uuid": "d6590c68-65e0-515b-a566-634f34c9045b", 00:09:45.914 "is_configured": true, 00:09:45.914 "data_offset": 2048, 00:09:45.914 "data_size": 63488 00:09:45.914 }, 00:09:45.914 { 00:09:45.914 "name": "BaseBdev4", 00:09:45.914 "uuid": "b1810edb-2927-5520-9c5e-28dce60a1e5d", 00:09:45.914 "is_configured": true, 00:09:45.914 "data_offset": 2048, 00:09:45.914 "data_size": 63488 00:09:45.914 } 00:09:45.914 ] 00:09:45.914 }' 00:09:45.914 23:43:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:45.914 23:43:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:46.172 23:43:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:09:46.172 23:43:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:46.172 23:43:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:46.172 [2024-11-26 23:43:34.282475] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:46.172 [2024-11-26 23:43:34.282548] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:46.172 [2024-11-26 23:43:34.285111] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:46.172 [2024-11-26 23:43:34.285205] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:46.172 [2024-11-26 23:43:34.285270] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:46.172 [2024-11-26 23:43:34.285333] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002000 name raid_bdev1, state offline 00:09:46.172 23:43:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:46.172 23:43:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 83563 00:09:46.172 23:43:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # '[' -z 83563 ']' 00:09:46.172 23:43:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@958 -- # kill -0 83563 00:09:46.172 { 00:09:46.172 "results": [ 00:09:46.172 { 00:09:46.172 "job": "raid_bdev1", 00:09:46.172 "core_mask": "0x1", 00:09:46.172 "workload": "randrw", 00:09:46.172 "percentage": 50, 00:09:46.172 "status": "finished", 00:09:46.172 "queue_depth": 1, 00:09:46.172 "io_size": 131072, 00:09:46.172 "runtime": 1.368589, 00:09:46.172 "iops": 16500.205686294423, 00:09:46.172 "mibps": 2062.525710786803, 00:09:46.172 "io_failed": 1, 00:09:46.172 "io_timeout": 0, 00:09:46.172 "avg_latency_us": 83.61078873140846, 00:09:46.172 "min_latency_us": 25.041048034934498, 00:09:46.172 "max_latency_us": 1373.6803493449781 00:09:46.172 } 00:09:46.172 ], 00:09:46.172 "core_count": 1 00:09:46.172 } 00:09:46.172 23:43:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # uname 00:09:46.172 23:43:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:09:46.172 23:43:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 83563 00:09:46.430 23:43:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:09:46.430 23:43:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:09:46.430 23:43:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 83563' 00:09:46.430 killing process with pid 83563 00:09:46.430 23:43:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@973 -- # kill 83563 00:09:46.430 [2024-11-26 23:43:34.329742] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:09:46.430 23:43:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@978 -- # wait 83563 00:09:46.430 [2024-11-26 23:43:34.364949] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:09:46.689 23:43:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.0gt49KbknK 00:09:46.689 23:43:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:09:46.689 23:43:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:09:46.689 23:43:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.73 00:09:46.689 23:43:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy concat 00:09:46.689 ************************************ 00:09:46.689 END TEST raid_write_error_test 00:09:46.689 ************************************ 00:09:46.689 23:43:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:09:46.689 23:43:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:09:46.689 23:43:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.73 != \0\.\0\0 ]] 00:09:46.689 00:09:46.689 real 0m3.341s 00:09:46.689 user 0m4.236s 00:09:46.689 sys 0m0.538s 00:09:46.689 23:43:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:09:46.689 23:43:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:46.689 23:43:34 bdev_raid -- bdev/bdev_raid.sh@967 -- # for level in raid0 concat raid1 00:09:46.689 23:43:34 bdev_raid -- bdev/bdev_raid.sh@968 -- # run_test raid_state_function_test raid_state_function_test raid1 4 false 00:09:46.689 23:43:34 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:09:46.689 23:43:34 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:09:46.689 23:43:34 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:09:46.689 ************************************ 00:09:46.689 START TEST raid_state_function_test 00:09:46.689 ************************************ 00:09:46.689 23:43:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1129 -- # raid_state_function_test raid1 4 false 00:09:46.689 23:43:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=raid1 00:09:46.689 23:43:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=4 00:09:46.689 23:43:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:09:46.689 23:43:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:09:46.689 23:43:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:09:46.689 23:43:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:46.689 23:43:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:09:46.689 23:43:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:46.689 23:43:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:46.689 23:43:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:09:46.689 23:43:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:46.689 23:43:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:46.689 23:43:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:09:46.689 23:43:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:46.689 23:43:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:46.689 23:43:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev4 00:09:46.690 23:43:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:46.690 23:43:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:46.690 23:43:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:09:46.690 23:43:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:09:46.690 23:43:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:09:46.690 23:43:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:09:46.690 23:43:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:09:46.690 23:43:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:09:46.690 23:43:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' raid1 '!=' raid1 ']' 00:09:46.690 23:43:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@219 -- # strip_size=0 00:09:46.690 23:43:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:09:46.690 23:43:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:09:46.690 Process raid pid: 83690 00:09:46.690 23:43:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=83690 00:09:46.690 23:43:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:09:46.690 23:43:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 83690' 00:09:46.690 23:43:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 83690 00:09:46.690 23:43:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@835 -- # '[' -z 83690 ']' 00:09:46.690 23:43:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:46.690 23:43:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:09:46.690 23:43:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:46.690 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:46.690 23:43:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:09:46.690 23:43:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:46.690 [2024-11-26 23:43:34.742242] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:09:46.690 [2024-11-26 23:43:34.742456] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:09:46.948 [2024-11-26 23:43:34.896188] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:46.948 [2024-11-26 23:43:34.920685] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:09:46.948 [2024-11-26 23:43:34.962833] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:46.948 [2024-11-26 23:43:34.962954] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:47.514 23:43:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:09:47.514 23:43:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@868 -- # return 0 00:09:47.514 23:43:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:09:47.514 23:43:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:47.514 23:43:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:47.514 [2024-11-26 23:43:35.569354] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:47.514 [2024-11-26 23:43:35.569460] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:47.514 [2024-11-26 23:43:35.569499] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:47.514 [2024-11-26 23:43:35.569526] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:47.514 [2024-11-26 23:43:35.569579] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:47.514 [2024-11-26 23:43:35.569605] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:47.514 [2024-11-26 23:43:35.569655] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:09:47.514 [2024-11-26 23:43:35.569678] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:09:47.514 23:43:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:47.514 23:43:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:09:47.514 23:43:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:47.514 23:43:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:47.514 23:43:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:47.514 23:43:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:47.514 23:43:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:47.514 23:43:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:47.514 23:43:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:47.514 23:43:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:47.514 23:43:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:47.514 23:43:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:47.514 23:43:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:47.514 23:43:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:47.514 23:43:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:47.514 23:43:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:47.514 23:43:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:47.514 "name": "Existed_Raid", 00:09:47.514 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:47.514 "strip_size_kb": 0, 00:09:47.514 "state": "configuring", 00:09:47.514 "raid_level": "raid1", 00:09:47.514 "superblock": false, 00:09:47.514 "num_base_bdevs": 4, 00:09:47.514 "num_base_bdevs_discovered": 0, 00:09:47.514 "num_base_bdevs_operational": 4, 00:09:47.514 "base_bdevs_list": [ 00:09:47.514 { 00:09:47.514 "name": "BaseBdev1", 00:09:47.514 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:47.514 "is_configured": false, 00:09:47.514 "data_offset": 0, 00:09:47.514 "data_size": 0 00:09:47.514 }, 00:09:47.514 { 00:09:47.514 "name": "BaseBdev2", 00:09:47.514 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:47.514 "is_configured": false, 00:09:47.514 "data_offset": 0, 00:09:47.514 "data_size": 0 00:09:47.514 }, 00:09:47.514 { 00:09:47.514 "name": "BaseBdev3", 00:09:47.515 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:47.515 "is_configured": false, 00:09:47.515 "data_offset": 0, 00:09:47.515 "data_size": 0 00:09:47.515 }, 00:09:47.515 { 00:09:47.515 "name": "BaseBdev4", 00:09:47.515 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:47.515 "is_configured": false, 00:09:47.515 "data_offset": 0, 00:09:47.515 "data_size": 0 00:09:47.515 } 00:09:47.515 ] 00:09:47.515 }' 00:09:47.515 23:43:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:47.515 23:43:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.083 23:43:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:48.083 23:43:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:48.083 23:43:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.083 [2024-11-26 23:43:36.016390] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:48.083 [2024-11-26 23:43:36.016469] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:09:48.083 23:43:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:48.083 23:43:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:09:48.083 23:43:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:48.083 23:43:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.083 [2024-11-26 23:43:36.024397] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:48.083 [2024-11-26 23:43:36.024471] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:48.083 [2024-11-26 23:43:36.024497] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:48.083 [2024-11-26 23:43:36.024519] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:48.083 [2024-11-26 23:43:36.024537] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:48.083 [2024-11-26 23:43:36.024557] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:48.083 [2024-11-26 23:43:36.024574] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:09:48.083 [2024-11-26 23:43:36.024595] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:09:48.083 23:43:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:48.083 23:43:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:09:48.083 23:43:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:48.083 23:43:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.083 [2024-11-26 23:43:36.041194] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:48.083 BaseBdev1 00:09:48.083 23:43:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:48.083 23:43:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:09:48.083 23:43:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:09:48.083 23:43:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:48.083 23:43:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:09:48.083 23:43:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:48.083 23:43:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:48.083 23:43:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:48.083 23:43:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:48.083 23:43:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.083 23:43:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:48.083 23:43:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:09:48.083 23:43:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:48.083 23:43:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.083 [ 00:09:48.083 { 00:09:48.083 "name": "BaseBdev1", 00:09:48.083 "aliases": [ 00:09:48.083 "59f79837-d536-4dde-9f21-cf802890ba50" 00:09:48.083 ], 00:09:48.083 "product_name": "Malloc disk", 00:09:48.083 "block_size": 512, 00:09:48.083 "num_blocks": 65536, 00:09:48.083 "uuid": "59f79837-d536-4dde-9f21-cf802890ba50", 00:09:48.083 "assigned_rate_limits": { 00:09:48.083 "rw_ios_per_sec": 0, 00:09:48.083 "rw_mbytes_per_sec": 0, 00:09:48.083 "r_mbytes_per_sec": 0, 00:09:48.083 "w_mbytes_per_sec": 0 00:09:48.083 }, 00:09:48.083 "claimed": true, 00:09:48.083 "claim_type": "exclusive_write", 00:09:48.083 "zoned": false, 00:09:48.083 "supported_io_types": { 00:09:48.083 "read": true, 00:09:48.083 "write": true, 00:09:48.083 "unmap": true, 00:09:48.083 "flush": true, 00:09:48.083 "reset": true, 00:09:48.083 "nvme_admin": false, 00:09:48.083 "nvme_io": false, 00:09:48.083 "nvme_io_md": false, 00:09:48.083 "write_zeroes": true, 00:09:48.083 "zcopy": true, 00:09:48.084 "get_zone_info": false, 00:09:48.084 "zone_management": false, 00:09:48.084 "zone_append": false, 00:09:48.084 "compare": false, 00:09:48.084 "compare_and_write": false, 00:09:48.084 "abort": true, 00:09:48.084 "seek_hole": false, 00:09:48.084 "seek_data": false, 00:09:48.084 "copy": true, 00:09:48.084 "nvme_iov_md": false 00:09:48.084 }, 00:09:48.084 "memory_domains": [ 00:09:48.084 { 00:09:48.084 "dma_device_id": "system", 00:09:48.084 "dma_device_type": 1 00:09:48.084 }, 00:09:48.084 { 00:09:48.084 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:48.084 "dma_device_type": 2 00:09:48.084 } 00:09:48.084 ], 00:09:48.084 "driver_specific": {} 00:09:48.084 } 00:09:48.084 ] 00:09:48.084 23:43:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:48.084 23:43:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:09:48.084 23:43:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:09:48.084 23:43:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:48.084 23:43:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:48.084 23:43:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:48.084 23:43:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:48.084 23:43:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:48.084 23:43:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:48.084 23:43:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:48.084 23:43:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:48.084 23:43:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:48.084 23:43:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:48.084 23:43:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:48.084 23:43:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:48.084 23:43:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.084 23:43:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:48.084 23:43:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:48.084 "name": "Existed_Raid", 00:09:48.084 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:48.084 "strip_size_kb": 0, 00:09:48.084 "state": "configuring", 00:09:48.084 "raid_level": "raid1", 00:09:48.084 "superblock": false, 00:09:48.084 "num_base_bdevs": 4, 00:09:48.084 "num_base_bdevs_discovered": 1, 00:09:48.084 "num_base_bdevs_operational": 4, 00:09:48.084 "base_bdevs_list": [ 00:09:48.084 { 00:09:48.084 "name": "BaseBdev1", 00:09:48.084 "uuid": "59f79837-d536-4dde-9f21-cf802890ba50", 00:09:48.084 "is_configured": true, 00:09:48.084 "data_offset": 0, 00:09:48.084 "data_size": 65536 00:09:48.084 }, 00:09:48.084 { 00:09:48.084 "name": "BaseBdev2", 00:09:48.084 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:48.084 "is_configured": false, 00:09:48.084 "data_offset": 0, 00:09:48.084 "data_size": 0 00:09:48.084 }, 00:09:48.084 { 00:09:48.084 "name": "BaseBdev3", 00:09:48.084 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:48.084 "is_configured": false, 00:09:48.084 "data_offset": 0, 00:09:48.084 "data_size": 0 00:09:48.084 }, 00:09:48.084 { 00:09:48.084 "name": "BaseBdev4", 00:09:48.084 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:48.084 "is_configured": false, 00:09:48.084 "data_offset": 0, 00:09:48.084 "data_size": 0 00:09:48.084 } 00:09:48.084 ] 00:09:48.084 }' 00:09:48.084 23:43:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:48.084 23:43:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.344 23:43:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:48.344 23:43:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:48.344 23:43:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.344 [2024-11-26 23:43:36.464526] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:48.344 [2024-11-26 23:43:36.464634] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:09:48.344 23:43:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:48.344 23:43:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:09:48.344 23:43:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:48.344 23:43:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.603 [2024-11-26 23:43:36.476553] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:48.603 [2024-11-26 23:43:36.478437] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:48.603 [2024-11-26 23:43:36.478478] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:48.603 [2024-11-26 23:43:36.478488] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:48.603 [2024-11-26 23:43:36.478496] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:48.603 [2024-11-26 23:43:36.478503] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:09:48.603 [2024-11-26 23:43:36.478512] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:09:48.603 23:43:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:48.603 23:43:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:09:48.603 23:43:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:48.603 23:43:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:09:48.603 23:43:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:48.603 23:43:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:48.603 23:43:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:48.603 23:43:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:48.603 23:43:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:48.603 23:43:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:48.603 23:43:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:48.603 23:43:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:48.603 23:43:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:48.603 23:43:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:48.603 23:43:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:48.603 23:43:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:48.603 23:43:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.603 23:43:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:48.603 23:43:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:48.603 "name": "Existed_Raid", 00:09:48.603 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:48.603 "strip_size_kb": 0, 00:09:48.603 "state": "configuring", 00:09:48.603 "raid_level": "raid1", 00:09:48.603 "superblock": false, 00:09:48.603 "num_base_bdevs": 4, 00:09:48.603 "num_base_bdevs_discovered": 1, 00:09:48.603 "num_base_bdevs_operational": 4, 00:09:48.603 "base_bdevs_list": [ 00:09:48.603 { 00:09:48.603 "name": "BaseBdev1", 00:09:48.603 "uuid": "59f79837-d536-4dde-9f21-cf802890ba50", 00:09:48.603 "is_configured": true, 00:09:48.603 "data_offset": 0, 00:09:48.603 "data_size": 65536 00:09:48.603 }, 00:09:48.603 { 00:09:48.603 "name": "BaseBdev2", 00:09:48.603 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:48.603 "is_configured": false, 00:09:48.603 "data_offset": 0, 00:09:48.603 "data_size": 0 00:09:48.603 }, 00:09:48.603 { 00:09:48.603 "name": "BaseBdev3", 00:09:48.603 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:48.603 "is_configured": false, 00:09:48.603 "data_offset": 0, 00:09:48.603 "data_size": 0 00:09:48.603 }, 00:09:48.603 { 00:09:48.603 "name": "BaseBdev4", 00:09:48.603 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:48.603 "is_configured": false, 00:09:48.603 "data_offset": 0, 00:09:48.603 "data_size": 0 00:09:48.603 } 00:09:48.603 ] 00:09:48.603 }' 00:09:48.603 23:43:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:48.603 23:43:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.862 23:43:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:09:48.862 23:43:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:48.862 23:43:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.862 [2024-11-26 23:43:36.946684] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:48.862 BaseBdev2 00:09:48.862 23:43:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:48.862 23:43:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:09:48.862 23:43:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:09:48.862 23:43:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:48.862 23:43:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:09:48.862 23:43:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:48.862 23:43:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:48.862 23:43:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:48.862 23:43:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:48.862 23:43:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.862 23:43:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:48.862 23:43:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:09:48.862 23:43:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:48.862 23:43:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.862 [ 00:09:48.862 { 00:09:48.862 "name": "BaseBdev2", 00:09:48.862 "aliases": [ 00:09:48.862 "aa2818d2-95f4-4950-9d6b-38643c5b3939" 00:09:48.862 ], 00:09:48.862 "product_name": "Malloc disk", 00:09:48.862 "block_size": 512, 00:09:48.862 "num_blocks": 65536, 00:09:48.862 "uuid": "aa2818d2-95f4-4950-9d6b-38643c5b3939", 00:09:48.862 "assigned_rate_limits": { 00:09:48.862 "rw_ios_per_sec": 0, 00:09:48.862 "rw_mbytes_per_sec": 0, 00:09:48.862 "r_mbytes_per_sec": 0, 00:09:48.862 "w_mbytes_per_sec": 0 00:09:48.862 }, 00:09:48.862 "claimed": true, 00:09:48.862 "claim_type": "exclusive_write", 00:09:48.862 "zoned": false, 00:09:48.862 "supported_io_types": { 00:09:48.862 "read": true, 00:09:48.862 "write": true, 00:09:48.862 "unmap": true, 00:09:48.862 "flush": true, 00:09:48.862 "reset": true, 00:09:48.862 "nvme_admin": false, 00:09:48.862 "nvme_io": false, 00:09:48.862 "nvme_io_md": false, 00:09:48.862 "write_zeroes": true, 00:09:48.862 "zcopy": true, 00:09:48.862 "get_zone_info": false, 00:09:48.862 "zone_management": false, 00:09:48.862 "zone_append": false, 00:09:48.862 "compare": false, 00:09:48.862 "compare_and_write": false, 00:09:48.862 "abort": true, 00:09:48.862 "seek_hole": false, 00:09:48.862 "seek_data": false, 00:09:48.862 "copy": true, 00:09:48.862 "nvme_iov_md": false 00:09:48.862 }, 00:09:48.862 "memory_domains": [ 00:09:48.862 { 00:09:48.862 "dma_device_id": "system", 00:09:48.862 "dma_device_type": 1 00:09:48.862 }, 00:09:48.862 { 00:09:48.862 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:48.862 "dma_device_type": 2 00:09:48.862 } 00:09:48.862 ], 00:09:48.862 "driver_specific": {} 00:09:48.862 } 00:09:48.862 ] 00:09:48.862 23:43:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:48.862 23:43:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:09:48.862 23:43:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:09:48.862 23:43:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:48.862 23:43:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:09:48.862 23:43:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:48.863 23:43:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:48.863 23:43:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:48.863 23:43:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:48.863 23:43:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:48.863 23:43:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:48.863 23:43:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:48.863 23:43:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:48.863 23:43:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:48.863 23:43:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:48.863 23:43:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:48.863 23:43:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.863 23:43:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:49.121 23:43:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:49.121 23:43:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:49.121 "name": "Existed_Raid", 00:09:49.121 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:49.121 "strip_size_kb": 0, 00:09:49.121 "state": "configuring", 00:09:49.121 "raid_level": "raid1", 00:09:49.121 "superblock": false, 00:09:49.121 "num_base_bdevs": 4, 00:09:49.121 "num_base_bdevs_discovered": 2, 00:09:49.121 "num_base_bdevs_operational": 4, 00:09:49.121 "base_bdevs_list": [ 00:09:49.121 { 00:09:49.121 "name": "BaseBdev1", 00:09:49.121 "uuid": "59f79837-d536-4dde-9f21-cf802890ba50", 00:09:49.121 "is_configured": true, 00:09:49.121 "data_offset": 0, 00:09:49.121 "data_size": 65536 00:09:49.121 }, 00:09:49.121 { 00:09:49.121 "name": "BaseBdev2", 00:09:49.121 "uuid": "aa2818d2-95f4-4950-9d6b-38643c5b3939", 00:09:49.121 "is_configured": true, 00:09:49.121 "data_offset": 0, 00:09:49.121 "data_size": 65536 00:09:49.121 }, 00:09:49.121 { 00:09:49.121 "name": "BaseBdev3", 00:09:49.121 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:49.121 "is_configured": false, 00:09:49.121 "data_offset": 0, 00:09:49.121 "data_size": 0 00:09:49.121 }, 00:09:49.121 { 00:09:49.121 "name": "BaseBdev4", 00:09:49.121 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:49.121 "is_configured": false, 00:09:49.121 "data_offset": 0, 00:09:49.121 "data_size": 0 00:09:49.121 } 00:09:49.122 ] 00:09:49.122 }' 00:09:49.122 23:43:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:49.122 23:43:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:49.381 23:43:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:09:49.381 23:43:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:49.381 23:43:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:49.381 [2024-11-26 23:43:37.393953] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:49.381 BaseBdev3 00:09:49.381 23:43:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:49.381 23:43:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:09:49.381 23:43:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:09:49.381 23:43:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:49.381 23:43:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:09:49.381 23:43:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:49.381 23:43:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:49.381 23:43:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:49.381 23:43:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:49.381 23:43:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:49.381 23:43:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:49.381 23:43:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:09:49.381 23:43:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:49.381 23:43:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:49.381 [ 00:09:49.381 { 00:09:49.381 "name": "BaseBdev3", 00:09:49.381 "aliases": [ 00:09:49.381 "9d3e1886-3d0e-478a-beef-d5d65a4cd415" 00:09:49.381 ], 00:09:49.381 "product_name": "Malloc disk", 00:09:49.381 "block_size": 512, 00:09:49.381 "num_blocks": 65536, 00:09:49.381 "uuid": "9d3e1886-3d0e-478a-beef-d5d65a4cd415", 00:09:49.381 "assigned_rate_limits": { 00:09:49.381 "rw_ios_per_sec": 0, 00:09:49.381 "rw_mbytes_per_sec": 0, 00:09:49.381 "r_mbytes_per_sec": 0, 00:09:49.381 "w_mbytes_per_sec": 0 00:09:49.381 }, 00:09:49.381 "claimed": true, 00:09:49.381 "claim_type": "exclusive_write", 00:09:49.381 "zoned": false, 00:09:49.381 "supported_io_types": { 00:09:49.381 "read": true, 00:09:49.381 "write": true, 00:09:49.381 "unmap": true, 00:09:49.381 "flush": true, 00:09:49.381 "reset": true, 00:09:49.381 "nvme_admin": false, 00:09:49.381 "nvme_io": false, 00:09:49.381 "nvme_io_md": false, 00:09:49.381 "write_zeroes": true, 00:09:49.381 "zcopy": true, 00:09:49.381 "get_zone_info": false, 00:09:49.381 "zone_management": false, 00:09:49.381 "zone_append": false, 00:09:49.381 "compare": false, 00:09:49.381 "compare_and_write": false, 00:09:49.381 "abort": true, 00:09:49.381 "seek_hole": false, 00:09:49.381 "seek_data": false, 00:09:49.381 "copy": true, 00:09:49.381 "nvme_iov_md": false 00:09:49.381 }, 00:09:49.381 "memory_domains": [ 00:09:49.381 { 00:09:49.381 "dma_device_id": "system", 00:09:49.381 "dma_device_type": 1 00:09:49.381 }, 00:09:49.381 { 00:09:49.381 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:49.381 "dma_device_type": 2 00:09:49.381 } 00:09:49.381 ], 00:09:49.381 "driver_specific": {} 00:09:49.381 } 00:09:49.381 ] 00:09:49.381 23:43:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:49.381 23:43:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:09:49.381 23:43:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:09:49.381 23:43:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:49.381 23:43:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:09:49.381 23:43:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:49.381 23:43:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:49.381 23:43:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:49.382 23:43:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:49.382 23:43:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:49.382 23:43:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:49.382 23:43:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:49.382 23:43:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:49.382 23:43:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:49.382 23:43:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:49.382 23:43:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:49.382 23:43:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:49.382 23:43:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:49.382 23:43:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:49.382 23:43:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:49.382 "name": "Existed_Raid", 00:09:49.382 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:49.382 "strip_size_kb": 0, 00:09:49.382 "state": "configuring", 00:09:49.382 "raid_level": "raid1", 00:09:49.382 "superblock": false, 00:09:49.382 "num_base_bdevs": 4, 00:09:49.382 "num_base_bdevs_discovered": 3, 00:09:49.382 "num_base_bdevs_operational": 4, 00:09:49.382 "base_bdevs_list": [ 00:09:49.382 { 00:09:49.382 "name": "BaseBdev1", 00:09:49.382 "uuid": "59f79837-d536-4dde-9f21-cf802890ba50", 00:09:49.382 "is_configured": true, 00:09:49.382 "data_offset": 0, 00:09:49.382 "data_size": 65536 00:09:49.382 }, 00:09:49.382 { 00:09:49.382 "name": "BaseBdev2", 00:09:49.382 "uuid": "aa2818d2-95f4-4950-9d6b-38643c5b3939", 00:09:49.382 "is_configured": true, 00:09:49.382 "data_offset": 0, 00:09:49.382 "data_size": 65536 00:09:49.382 }, 00:09:49.382 { 00:09:49.382 "name": "BaseBdev3", 00:09:49.382 "uuid": "9d3e1886-3d0e-478a-beef-d5d65a4cd415", 00:09:49.382 "is_configured": true, 00:09:49.382 "data_offset": 0, 00:09:49.382 "data_size": 65536 00:09:49.382 }, 00:09:49.382 { 00:09:49.382 "name": "BaseBdev4", 00:09:49.382 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:49.382 "is_configured": false, 00:09:49.382 "data_offset": 0, 00:09:49.382 "data_size": 0 00:09:49.382 } 00:09:49.382 ] 00:09:49.382 }' 00:09:49.382 23:43:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:49.382 23:43:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:49.949 23:43:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:09:49.950 23:43:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:49.950 23:43:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:49.950 [2024-11-26 23:43:37.900114] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:09:49.950 [2024-11-26 23:43:37.900242] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:09:49.950 [2024-11-26 23:43:37.900268] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:09:49.950 [2024-11-26 23:43:37.900603] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:09:49.950 [2024-11-26 23:43:37.900796] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:09:49.950 [2024-11-26 23:43:37.900842] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:09:49.950 [2024-11-26 23:43:37.901089] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:49.950 BaseBdev4 00:09:49.950 23:43:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:49.950 23:43:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev4 00:09:49.950 23:43:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:09:49.950 23:43:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:49.950 23:43:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:09:49.950 23:43:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:49.950 23:43:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:49.950 23:43:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:49.950 23:43:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:49.950 23:43:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:49.950 23:43:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:49.950 23:43:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:09:49.950 23:43:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:49.950 23:43:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:49.950 [ 00:09:49.950 { 00:09:49.950 "name": "BaseBdev4", 00:09:49.950 "aliases": [ 00:09:49.950 "db4c0e7a-cfcf-4fa5-9dfe-362761b70f5b" 00:09:49.950 ], 00:09:49.950 "product_name": "Malloc disk", 00:09:49.950 "block_size": 512, 00:09:49.950 "num_blocks": 65536, 00:09:49.950 "uuid": "db4c0e7a-cfcf-4fa5-9dfe-362761b70f5b", 00:09:49.950 "assigned_rate_limits": { 00:09:49.950 "rw_ios_per_sec": 0, 00:09:49.950 "rw_mbytes_per_sec": 0, 00:09:49.950 "r_mbytes_per_sec": 0, 00:09:49.950 "w_mbytes_per_sec": 0 00:09:49.950 }, 00:09:49.950 "claimed": true, 00:09:49.950 "claim_type": "exclusive_write", 00:09:49.950 "zoned": false, 00:09:49.950 "supported_io_types": { 00:09:49.950 "read": true, 00:09:49.950 "write": true, 00:09:49.950 "unmap": true, 00:09:49.950 "flush": true, 00:09:49.950 "reset": true, 00:09:49.950 "nvme_admin": false, 00:09:49.950 "nvme_io": false, 00:09:49.950 "nvme_io_md": false, 00:09:49.950 "write_zeroes": true, 00:09:49.950 "zcopy": true, 00:09:49.950 "get_zone_info": false, 00:09:49.950 "zone_management": false, 00:09:49.950 "zone_append": false, 00:09:49.950 "compare": false, 00:09:49.950 "compare_and_write": false, 00:09:49.950 "abort": true, 00:09:49.950 "seek_hole": false, 00:09:49.950 "seek_data": false, 00:09:49.950 "copy": true, 00:09:49.950 "nvme_iov_md": false 00:09:49.950 }, 00:09:49.950 "memory_domains": [ 00:09:49.950 { 00:09:49.950 "dma_device_id": "system", 00:09:49.950 "dma_device_type": 1 00:09:49.950 }, 00:09:49.950 { 00:09:49.950 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:49.950 "dma_device_type": 2 00:09:49.950 } 00:09:49.950 ], 00:09:49.950 "driver_specific": {} 00:09:49.950 } 00:09:49.950 ] 00:09:49.950 23:43:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:49.950 23:43:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:09:49.950 23:43:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:09:49.950 23:43:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:49.950 23:43:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid1 0 4 00:09:49.950 23:43:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:49.950 23:43:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:49.950 23:43:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:49.950 23:43:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:49.950 23:43:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:49.950 23:43:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:49.950 23:43:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:49.950 23:43:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:49.950 23:43:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:49.950 23:43:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:49.950 23:43:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:49.950 23:43:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:49.950 23:43:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:49.950 23:43:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:49.950 23:43:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:49.950 "name": "Existed_Raid", 00:09:49.950 "uuid": "04bf6861-09f3-466b-bbb7-16c4b7f13ff1", 00:09:49.950 "strip_size_kb": 0, 00:09:49.950 "state": "online", 00:09:49.950 "raid_level": "raid1", 00:09:49.950 "superblock": false, 00:09:49.950 "num_base_bdevs": 4, 00:09:49.950 "num_base_bdevs_discovered": 4, 00:09:49.950 "num_base_bdevs_operational": 4, 00:09:49.950 "base_bdevs_list": [ 00:09:49.950 { 00:09:49.950 "name": "BaseBdev1", 00:09:49.950 "uuid": "59f79837-d536-4dde-9f21-cf802890ba50", 00:09:49.950 "is_configured": true, 00:09:49.950 "data_offset": 0, 00:09:49.950 "data_size": 65536 00:09:49.950 }, 00:09:49.950 { 00:09:49.950 "name": "BaseBdev2", 00:09:49.950 "uuid": "aa2818d2-95f4-4950-9d6b-38643c5b3939", 00:09:49.950 "is_configured": true, 00:09:49.950 "data_offset": 0, 00:09:49.950 "data_size": 65536 00:09:49.950 }, 00:09:49.950 { 00:09:49.950 "name": "BaseBdev3", 00:09:49.950 "uuid": "9d3e1886-3d0e-478a-beef-d5d65a4cd415", 00:09:49.950 "is_configured": true, 00:09:49.950 "data_offset": 0, 00:09:49.950 "data_size": 65536 00:09:49.950 }, 00:09:49.950 { 00:09:49.950 "name": "BaseBdev4", 00:09:49.950 "uuid": "db4c0e7a-cfcf-4fa5-9dfe-362761b70f5b", 00:09:49.950 "is_configured": true, 00:09:49.950 "data_offset": 0, 00:09:49.950 "data_size": 65536 00:09:49.950 } 00:09:49.950 ] 00:09:49.950 }' 00:09:49.950 23:43:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:49.950 23:43:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:50.209 23:43:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:09:50.209 23:43:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:09:50.209 23:43:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:50.209 23:43:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:50.209 23:43:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:09:50.209 23:43:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:50.467 23:43:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:09:50.467 23:43:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:50.467 23:43:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:50.467 23:43:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:50.467 [2024-11-26 23:43:38.343811] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:50.467 23:43:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:50.467 23:43:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:50.467 "name": "Existed_Raid", 00:09:50.467 "aliases": [ 00:09:50.467 "04bf6861-09f3-466b-bbb7-16c4b7f13ff1" 00:09:50.467 ], 00:09:50.467 "product_name": "Raid Volume", 00:09:50.467 "block_size": 512, 00:09:50.467 "num_blocks": 65536, 00:09:50.467 "uuid": "04bf6861-09f3-466b-bbb7-16c4b7f13ff1", 00:09:50.467 "assigned_rate_limits": { 00:09:50.467 "rw_ios_per_sec": 0, 00:09:50.467 "rw_mbytes_per_sec": 0, 00:09:50.467 "r_mbytes_per_sec": 0, 00:09:50.467 "w_mbytes_per_sec": 0 00:09:50.467 }, 00:09:50.467 "claimed": false, 00:09:50.467 "zoned": false, 00:09:50.467 "supported_io_types": { 00:09:50.467 "read": true, 00:09:50.467 "write": true, 00:09:50.467 "unmap": false, 00:09:50.467 "flush": false, 00:09:50.467 "reset": true, 00:09:50.467 "nvme_admin": false, 00:09:50.467 "nvme_io": false, 00:09:50.467 "nvme_io_md": false, 00:09:50.467 "write_zeroes": true, 00:09:50.467 "zcopy": false, 00:09:50.467 "get_zone_info": false, 00:09:50.467 "zone_management": false, 00:09:50.467 "zone_append": false, 00:09:50.467 "compare": false, 00:09:50.467 "compare_and_write": false, 00:09:50.467 "abort": false, 00:09:50.467 "seek_hole": false, 00:09:50.467 "seek_data": false, 00:09:50.467 "copy": false, 00:09:50.467 "nvme_iov_md": false 00:09:50.467 }, 00:09:50.467 "memory_domains": [ 00:09:50.467 { 00:09:50.467 "dma_device_id": "system", 00:09:50.467 "dma_device_type": 1 00:09:50.467 }, 00:09:50.467 { 00:09:50.467 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:50.467 "dma_device_type": 2 00:09:50.467 }, 00:09:50.467 { 00:09:50.467 "dma_device_id": "system", 00:09:50.467 "dma_device_type": 1 00:09:50.467 }, 00:09:50.467 { 00:09:50.467 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:50.467 "dma_device_type": 2 00:09:50.468 }, 00:09:50.468 { 00:09:50.468 "dma_device_id": "system", 00:09:50.468 "dma_device_type": 1 00:09:50.468 }, 00:09:50.468 { 00:09:50.468 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:50.468 "dma_device_type": 2 00:09:50.468 }, 00:09:50.468 { 00:09:50.468 "dma_device_id": "system", 00:09:50.468 "dma_device_type": 1 00:09:50.468 }, 00:09:50.468 { 00:09:50.468 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:50.468 "dma_device_type": 2 00:09:50.468 } 00:09:50.468 ], 00:09:50.468 "driver_specific": { 00:09:50.468 "raid": { 00:09:50.468 "uuid": "04bf6861-09f3-466b-bbb7-16c4b7f13ff1", 00:09:50.468 "strip_size_kb": 0, 00:09:50.468 "state": "online", 00:09:50.468 "raid_level": "raid1", 00:09:50.468 "superblock": false, 00:09:50.468 "num_base_bdevs": 4, 00:09:50.468 "num_base_bdevs_discovered": 4, 00:09:50.468 "num_base_bdevs_operational": 4, 00:09:50.468 "base_bdevs_list": [ 00:09:50.468 { 00:09:50.468 "name": "BaseBdev1", 00:09:50.468 "uuid": "59f79837-d536-4dde-9f21-cf802890ba50", 00:09:50.468 "is_configured": true, 00:09:50.468 "data_offset": 0, 00:09:50.468 "data_size": 65536 00:09:50.468 }, 00:09:50.468 { 00:09:50.468 "name": "BaseBdev2", 00:09:50.468 "uuid": "aa2818d2-95f4-4950-9d6b-38643c5b3939", 00:09:50.468 "is_configured": true, 00:09:50.468 "data_offset": 0, 00:09:50.468 "data_size": 65536 00:09:50.468 }, 00:09:50.468 { 00:09:50.468 "name": "BaseBdev3", 00:09:50.468 "uuid": "9d3e1886-3d0e-478a-beef-d5d65a4cd415", 00:09:50.468 "is_configured": true, 00:09:50.468 "data_offset": 0, 00:09:50.468 "data_size": 65536 00:09:50.468 }, 00:09:50.468 { 00:09:50.468 "name": "BaseBdev4", 00:09:50.468 "uuid": "db4c0e7a-cfcf-4fa5-9dfe-362761b70f5b", 00:09:50.468 "is_configured": true, 00:09:50.468 "data_offset": 0, 00:09:50.468 "data_size": 65536 00:09:50.468 } 00:09:50.468 ] 00:09:50.468 } 00:09:50.468 } 00:09:50.468 }' 00:09:50.468 23:43:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:50.468 23:43:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:09:50.468 BaseBdev2 00:09:50.468 BaseBdev3 00:09:50.468 BaseBdev4' 00:09:50.468 23:43:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:50.468 23:43:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:50.468 23:43:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:50.468 23:43:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:50.468 23:43:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:09:50.468 23:43:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:50.468 23:43:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:50.468 23:43:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:50.468 23:43:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:50.468 23:43:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:50.468 23:43:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:50.468 23:43:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:09:50.468 23:43:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:50.468 23:43:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:50.468 23:43:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:50.468 23:43:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:50.468 23:43:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:50.468 23:43:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:50.468 23:43:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:50.468 23:43:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:09:50.468 23:43:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:50.468 23:43:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:50.468 23:43:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:50.468 23:43:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:50.468 23:43:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:50.468 23:43:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:50.468 23:43:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:50.468 23:43:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:09:50.468 23:43:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:50.468 23:43:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:50.468 23:43:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:50.727 23:43:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:50.727 23:43:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:50.727 23:43:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:50.727 23:43:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:09:50.727 23:43:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:50.727 23:43:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:50.727 [2024-11-26 23:43:38.623079] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:09:50.727 23:43:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:50.727 23:43:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:09:50.727 23:43:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy raid1 00:09:50.727 23:43:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:09:50.727 23:43:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@199 -- # return 0 00:09:50.727 23:43:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:09:50.727 23:43:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid1 0 3 00:09:50.727 23:43:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:50.727 23:43:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:50.727 23:43:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:50.727 23:43:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:50.727 23:43:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:50.727 23:43:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:50.727 23:43:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:50.727 23:43:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:50.727 23:43:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:50.727 23:43:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:50.727 23:43:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:50.727 23:43:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:50.727 23:43:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:50.727 23:43:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:50.727 23:43:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:50.727 "name": "Existed_Raid", 00:09:50.727 "uuid": "04bf6861-09f3-466b-bbb7-16c4b7f13ff1", 00:09:50.727 "strip_size_kb": 0, 00:09:50.727 "state": "online", 00:09:50.727 "raid_level": "raid1", 00:09:50.727 "superblock": false, 00:09:50.727 "num_base_bdevs": 4, 00:09:50.727 "num_base_bdevs_discovered": 3, 00:09:50.727 "num_base_bdevs_operational": 3, 00:09:50.727 "base_bdevs_list": [ 00:09:50.727 { 00:09:50.727 "name": null, 00:09:50.727 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:50.727 "is_configured": false, 00:09:50.727 "data_offset": 0, 00:09:50.727 "data_size": 65536 00:09:50.727 }, 00:09:50.727 { 00:09:50.727 "name": "BaseBdev2", 00:09:50.727 "uuid": "aa2818d2-95f4-4950-9d6b-38643c5b3939", 00:09:50.727 "is_configured": true, 00:09:50.727 "data_offset": 0, 00:09:50.727 "data_size": 65536 00:09:50.727 }, 00:09:50.727 { 00:09:50.727 "name": "BaseBdev3", 00:09:50.727 "uuid": "9d3e1886-3d0e-478a-beef-d5d65a4cd415", 00:09:50.727 "is_configured": true, 00:09:50.727 "data_offset": 0, 00:09:50.727 "data_size": 65536 00:09:50.727 }, 00:09:50.727 { 00:09:50.727 "name": "BaseBdev4", 00:09:50.727 "uuid": "db4c0e7a-cfcf-4fa5-9dfe-362761b70f5b", 00:09:50.727 "is_configured": true, 00:09:50.727 "data_offset": 0, 00:09:50.727 "data_size": 65536 00:09:50.727 } 00:09:50.728 ] 00:09:50.728 }' 00:09:50.728 23:43:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:50.728 23:43:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:50.987 23:43:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:09:50.987 23:43:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:50.987 23:43:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:50.987 23:43:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:50.987 23:43:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:50.987 23:43:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:09:50.987 23:43:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:51.246 23:43:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:09:51.246 23:43:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:09:51.246 23:43:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:09:51.246 23:43:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:51.246 23:43:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:51.246 [2024-11-26 23:43:39.137802] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:09:51.246 23:43:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:51.246 23:43:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:09:51.246 23:43:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:51.246 23:43:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:09:51.246 23:43:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:51.246 23:43:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:51.246 23:43:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:51.246 23:43:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:51.246 23:43:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:09:51.246 23:43:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:09:51.246 23:43:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:09:51.246 23:43:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:51.246 23:43:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:51.246 [2024-11-26 23:43:39.192915] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:09:51.246 23:43:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:51.246 23:43:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:09:51.246 23:43:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:51.246 23:43:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:09:51.246 23:43:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:51.246 23:43:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:51.246 23:43:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:51.246 23:43:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:51.246 23:43:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:09:51.246 23:43:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:09:51.246 23:43:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev4 00:09:51.246 23:43:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:51.246 23:43:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:51.246 [2024-11-26 23:43:39.256073] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev4 00:09:51.246 [2024-11-26 23:43:39.256163] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:51.246 [2024-11-26 23:43:39.267593] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:51.246 [2024-11-26 23:43:39.267639] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:51.246 [2024-11-26 23:43:39.267651] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:09:51.246 23:43:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:51.246 23:43:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:09:51.246 23:43:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:51.246 23:43:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:51.246 23:43:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:09:51.246 23:43:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:51.246 23:43:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:51.246 23:43:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:51.246 23:43:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:09:51.246 23:43:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:09:51.246 23:43:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 4 -gt 2 ']' 00:09:51.246 23:43:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:09:51.246 23:43:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:51.246 23:43:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:09:51.246 23:43:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:51.246 23:43:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:51.246 BaseBdev2 00:09:51.246 23:43:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:51.246 23:43:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:09:51.246 23:43:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:09:51.246 23:43:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:51.246 23:43:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:09:51.246 23:43:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:51.246 23:43:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:51.246 23:43:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:51.246 23:43:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:51.246 23:43:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:51.246 23:43:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:51.246 23:43:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:09:51.246 23:43:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:51.246 23:43:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:51.246 [ 00:09:51.246 { 00:09:51.246 "name": "BaseBdev2", 00:09:51.246 "aliases": [ 00:09:51.246 "d8c630d6-cda4-4dea-b338-ada656de458c" 00:09:51.246 ], 00:09:51.246 "product_name": "Malloc disk", 00:09:51.246 "block_size": 512, 00:09:51.246 "num_blocks": 65536, 00:09:51.246 "uuid": "d8c630d6-cda4-4dea-b338-ada656de458c", 00:09:51.246 "assigned_rate_limits": { 00:09:51.246 "rw_ios_per_sec": 0, 00:09:51.246 "rw_mbytes_per_sec": 0, 00:09:51.246 "r_mbytes_per_sec": 0, 00:09:51.246 "w_mbytes_per_sec": 0 00:09:51.246 }, 00:09:51.246 "claimed": false, 00:09:51.246 "zoned": false, 00:09:51.246 "supported_io_types": { 00:09:51.246 "read": true, 00:09:51.246 "write": true, 00:09:51.246 "unmap": true, 00:09:51.246 "flush": true, 00:09:51.246 "reset": true, 00:09:51.246 "nvme_admin": false, 00:09:51.246 "nvme_io": false, 00:09:51.246 "nvme_io_md": false, 00:09:51.246 "write_zeroes": true, 00:09:51.246 "zcopy": true, 00:09:51.246 "get_zone_info": false, 00:09:51.246 "zone_management": false, 00:09:51.246 "zone_append": false, 00:09:51.246 "compare": false, 00:09:51.246 "compare_and_write": false, 00:09:51.246 "abort": true, 00:09:51.246 "seek_hole": false, 00:09:51.246 "seek_data": false, 00:09:51.246 "copy": true, 00:09:51.246 "nvme_iov_md": false 00:09:51.246 }, 00:09:51.246 "memory_domains": [ 00:09:51.246 { 00:09:51.246 "dma_device_id": "system", 00:09:51.246 "dma_device_type": 1 00:09:51.246 }, 00:09:51.246 { 00:09:51.246 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:51.246 "dma_device_type": 2 00:09:51.246 } 00:09:51.246 ], 00:09:51.246 "driver_specific": {} 00:09:51.246 } 00:09:51.246 ] 00:09:51.246 23:43:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:51.246 23:43:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:09:51.246 23:43:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:09:51.246 23:43:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:51.246 23:43:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:09:51.246 23:43:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:51.246 23:43:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:51.506 BaseBdev3 00:09:51.506 23:43:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:51.506 23:43:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:09:51.506 23:43:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:09:51.507 23:43:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:51.507 23:43:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:09:51.507 23:43:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:51.507 23:43:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:51.507 23:43:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:51.507 23:43:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:51.507 23:43:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:51.507 23:43:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:51.507 23:43:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:09:51.507 23:43:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:51.507 23:43:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:51.507 [ 00:09:51.507 { 00:09:51.507 "name": "BaseBdev3", 00:09:51.507 "aliases": [ 00:09:51.507 "23ddc50f-e8f5-4eed-ae5f-164b59ab9a0e" 00:09:51.507 ], 00:09:51.507 "product_name": "Malloc disk", 00:09:51.507 "block_size": 512, 00:09:51.507 "num_blocks": 65536, 00:09:51.507 "uuid": "23ddc50f-e8f5-4eed-ae5f-164b59ab9a0e", 00:09:51.507 "assigned_rate_limits": { 00:09:51.507 "rw_ios_per_sec": 0, 00:09:51.507 "rw_mbytes_per_sec": 0, 00:09:51.507 "r_mbytes_per_sec": 0, 00:09:51.507 "w_mbytes_per_sec": 0 00:09:51.507 }, 00:09:51.507 "claimed": false, 00:09:51.507 "zoned": false, 00:09:51.507 "supported_io_types": { 00:09:51.507 "read": true, 00:09:51.507 "write": true, 00:09:51.507 "unmap": true, 00:09:51.507 "flush": true, 00:09:51.507 "reset": true, 00:09:51.507 "nvme_admin": false, 00:09:51.507 "nvme_io": false, 00:09:51.507 "nvme_io_md": false, 00:09:51.507 "write_zeroes": true, 00:09:51.507 "zcopy": true, 00:09:51.507 "get_zone_info": false, 00:09:51.507 "zone_management": false, 00:09:51.507 "zone_append": false, 00:09:51.507 "compare": false, 00:09:51.507 "compare_and_write": false, 00:09:51.507 "abort": true, 00:09:51.507 "seek_hole": false, 00:09:51.507 "seek_data": false, 00:09:51.507 "copy": true, 00:09:51.507 "nvme_iov_md": false 00:09:51.507 }, 00:09:51.507 "memory_domains": [ 00:09:51.507 { 00:09:51.507 "dma_device_id": "system", 00:09:51.507 "dma_device_type": 1 00:09:51.507 }, 00:09:51.507 { 00:09:51.507 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:51.507 "dma_device_type": 2 00:09:51.507 } 00:09:51.507 ], 00:09:51.507 "driver_specific": {} 00:09:51.507 } 00:09:51.507 ] 00:09:51.507 23:43:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:51.507 23:43:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:09:51.507 23:43:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:09:51.507 23:43:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:51.507 23:43:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:09:51.507 23:43:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:51.507 23:43:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:51.507 BaseBdev4 00:09:51.507 23:43:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:51.507 23:43:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev4 00:09:51.507 23:43:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:09:51.507 23:43:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:51.507 23:43:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:09:51.507 23:43:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:51.507 23:43:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:51.507 23:43:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:51.507 23:43:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:51.507 23:43:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:51.507 23:43:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:51.507 23:43:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:09:51.507 23:43:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:51.507 23:43:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:51.507 [ 00:09:51.507 { 00:09:51.507 "name": "BaseBdev4", 00:09:51.507 "aliases": [ 00:09:51.507 "aa840108-0a37-4337-ac98-77500cd6ba01" 00:09:51.507 ], 00:09:51.507 "product_name": "Malloc disk", 00:09:51.507 "block_size": 512, 00:09:51.507 "num_blocks": 65536, 00:09:51.507 "uuid": "aa840108-0a37-4337-ac98-77500cd6ba01", 00:09:51.507 "assigned_rate_limits": { 00:09:51.507 "rw_ios_per_sec": 0, 00:09:51.507 "rw_mbytes_per_sec": 0, 00:09:51.507 "r_mbytes_per_sec": 0, 00:09:51.507 "w_mbytes_per_sec": 0 00:09:51.507 }, 00:09:51.507 "claimed": false, 00:09:51.507 "zoned": false, 00:09:51.507 "supported_io_types": { 00:09:51.507 "read": true, 00:09:51.507 "write": true, 00:09:51.507 "unmap": true, 00:09:51.507 "flush": true, 00:09:51.507 "reset": true, 00:09:51.507 "nvme_admin": false, 00:09:51.507 "nvme_io": false, 00:09:51.507 "nvme_io_md": false, 00:09:51.507 "write_zeroes": true, 00:09:51.507 "zcopy": true, 00:09:51.507 "get_zone_info": false, 00:09:51.507 "zone_management": false, 00:09:51.507 "zone_append": false, 00:09:51.507 "compare": false, 00:09:51.507 "compare_and_write": false, 00:09:51.507 "abort": true, 00:09:51.507 "seek_hole": false, 00:09:51.507 "seek_data": false, 00:09:51.507 "copy": true, 00:09:51.507 "nvme_iov_md": false 00:09:51.507 }, 00:09:51.507 "memory_domains": [ 00:09:51.507 { 00:09:51.507 "dma_device_id": "system", 00:09:51.507 "dma_device_type": 1 00:09:51.507 }, 00:09:51.507 { 00:09:51.507 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:51.507 "dma_device_type": 2 00:09:51.507 } 00:09:51.507 ], 00:09:51.507 "driver_specific": {} 00:09:51.507 } 00:09:51.507 ] 00:09:51.507 23:43:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:51.507 23:43:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:09:51.507 23:43:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:09:51.507 23:43:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:51.507 23:43:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:09:51.507 23:43:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:51.507 23:43:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:51.507 [2024-11-26 23:43:39.484378] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:51.507 [2024-11-26 23:43:39.484460] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:51.507 [2024-11-26 23:43:39.484513] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:51.507 [2024-11-26 23:43:39.486300] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:51.507 [2024-11-26 23:43:39.486410] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:09:51.507 23:43:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:51.507 23:43:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:09:51.507 23:43:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:51.507 23:43:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:51.507 23:43:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:51.507 23:43:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:51.507 23:43:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:51.507 23:43:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:51.507 23:43:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:51.507 23:43:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:51.507 23:43:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:51.507 23:43:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:51.507 23:43:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:51.507 23:43:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:51.507 23:43:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:51.507 23:43:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:51.507 23:43:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:51.507 "name": "Existed_Raid", 00:09:51.507 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:51.507 "strip_size_kb": 0, 00:09:51.507 "state": "configuring", 00:09:51.507 "raid_level": "raid1", 00:09:51.507 "superblock": false, 00:09:51.507 "num_base_bdevs": 4, 00:09:51.507 "num_base_bdevs_discovered": 3, 00:09:51.507 "num_base_bdevs_operational": 4, 00:09:51.507 "base_bdevs_list": [ 00:09:51.507 { 00:09:51.507 "name": "BaseBdev1", 00:09:51.507 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:51.507 "is_configured": false, 00:09:51.508 "data_offset": 0, 00:09:51.508 "data_size": 0 00:09:51.508 }, 00:09:51.508 { 00:09:51.508 "name": "BaseBdev2", 00:09:51.508 "uuid": "d8c630d6-cda4-4dea-b338-ada656de458c", 00:09:51.508 "is_configured": true, 00:09:51.508 "data_offset": 0, 00:09:51.508 "data_size": 65536 00:09:51.508 }, 00:09:51.508 { 00:09:51.508 "name": "BaseBdev3", 00:09:51.508 "uuid": "23ddc50f-e8f5-4eed-ae5f-164b59ab9a0e", 00:09:51.508 "is_configured": true, 00:09:51.508 "data_offset": 0, 00:09:51.508 "data_size": 65536 00:09:51.508 }, 00:09:51.508 { 00:09:51.508 "name": "BaseBdev4", 00:09:51.508 "uuid": "aa840108-0a37-4337-ac98-77500cd6ba01", 00:09:51.508 "is_configured": true, 00:09:51.508 "data_offset": 0, 00:09:51.508 "data_size": 65536 00:09:51.508 } 00:09:51.508 ] 00:09:51.508 }' 00:09:51.508 23:43:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:51.508 23:43:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:52.075 23:43:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:09:52.075 23:43:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:52.075 23:43:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:52.075 [2024-11-26 23:43:39.943594] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:09:52.075 23:43:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:52.075 23:43:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:09:52.075 23:43:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:52.075 23:43:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:52.075 23:43:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:52.075 23:43:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:52.075 23:43:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:52.075 23:43:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:52.075 23:43:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:52.075 23:43:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:52.075 23:43:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:52.075 23:43:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:52.076 23:43:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:52.076 23:43:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:52.076 23:43:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:52.076 23:43:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:52.076 23:43:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:52.076 "name": "Existed_Raid", 00:09:52.076 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:52.076 "strip_size_kb": 0, 00:09:52.076 "state": "configuring", 00:09:52.076 "raid_level": "raid1", 00:09:52.076 "superblock": false, 00:09:52.076 "num_base_bdevs": 4, 00:09:52.076 "num_base_bdevs_discovered": 2, 00:09:52.076 "num_base_bdevs_operational": 4, 00:09:52.076 "base_bdevs_list": [ 00:09:52.076 { 00:09:52.076 "name": "BaseBdev1", 00:09:52.076 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:52.076 "is_configured": false, 00:09:52.076 "data_offset": 0, 00:09:52.076 "data_size": 0 00:09:52.076 }, 00:09:52.076 { 00:09:52.076 "name": null, 00:09:52.076 "uuid": "d8c630d6-cda4-4dea-b338-ada656de458c", 00:09:52.076 "is_configured": false, 00:09:52.076 "data_offset": 0, 00:09:52.076 "data_size": 65536 00:09:52.076 }, 00:09:52.076 { 00:09:52.076 "name": "BaseBdev3", 00:09:52.076 "uuid": "23ddc50f-e8f5-4eed-ae5f-164b59ab9a0e", 00:09:52.076 "is_configured": true, 00:09:52.076 "data_offset": 0, 00:09:52.076 "data_size": 65536 00:09:52.076 }, 00:09:52.076 { 00:09:52.076 "name": "BaseBdev4", 00:09:52.076 "uuid": "aa840108-0a37-4337-ac98-77500cd6ba01", 00:09:52.076 "is_configured": true, 00:09:52.076 "data_offset": 0, 00:09:52.076 "data_size": 65536 00:09:52.076 } 00:09:52.076 ] 00:09:52.076 }' 00:09:52.076 23:43:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:52.076 23:43:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:52.334 23:43:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:09:52.334 23:43:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:52.334 23:43:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:52.334 23:43:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:52.334 23:43:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:52.334 23:43:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:09:52.334 23:43:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:09:52.334 23:43:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:52.334 23:43:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:52.334 [2024-11-26 23:43:40.393701] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:52.334 BaseBdev1 00:09:52.334 23:43:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:52.334 23:43:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:09:52.334 23:43:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:09:52.334 23:43:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:52.334 23:43:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:09:52.334 23:43:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:52.334 23:43:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:52.334 23:43:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:52.334 23:43:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:52.334 23:43:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:52.334 23:43:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:52.334 23:43:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:09:52.334 23:43:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:52.334 23:43:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:52.334 [ 00:09:52.334 { 00:09:52.334 "name": "BaseBdev1", 00:09:52.334 "aliases": [ 00:09:52.334 "9d2660e9-c8e2-42e2-8820-cb3af5b7e5ec" 00:09:52.334 ], 00:09:52.334 "product_name": "Malloc disk", 00:09:52.334 "block_size": 512, 00:09:52.334 "num_blocks": 65536, 00:09:52.334 "uuid": "9d2660e9-c8e2-42e2-8820-cb3af5b7e5ec", 00:09:52.334 "assigned_rate_limits": { 00:09:52.334 "rw_ios_per_sec": 0, 00:09:52.334 "rw_mbytes_per_sec": 0, 00:09:52.334 "r_mbytes_per_sec": 0, 00:09:52.334 "w_mbytes_per_sec": 0 00:09:52.334 }, 00:09:52.334 "claimed": true, 00:09:52.334 "claim_type": "exclusive_write", 00:09:52.334 "zoned": false, 00:09:52.334 "supported_io_types": { 00:09:52.334 "read": true, 00:09:52.334 "write": true, 00:09:52.334 "unmap": true, 00:09:52.334 "flush": true, 00:09:52.334 "reset": true, 00:09:52.334 "nvme_admin": false, 00:09:52.334 "nvme_io": false, 00:09:52.334 "nvme_io_md": false, 00:09:52.334 "write_zeroes": true, 00:09:52.334 "zcopy": true, 00:09:52.334 "get_zone_info": false, 00:09:52.334 "zone_management": false, 00:09:52.334 "zone_append": false, 00:09:52.334 "compare": false, 00:09:52.334 "compare_and_write": false, 00:09:52.334 "abort": true, 00:09:52.334 "seek_hole": false, 00:09:52.334 "seek_data": false, 00:09:52.334 "copy": true, 00:09:52.334 "nvme_iov_md": false 00:09:52.334 }, 00:09:52.334 "memory_domains": [ 00:09:52.334 { 00:09:52.334 "dma_device_id": "system", 00:09:52.334 "dma_device_type": 1 00:09:52.334 }, 00:09:52.334 { 00:09:52.334 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:52.334 "dma_device_type": 2 00:09:52.334 } 00:09:52.334 ], 00:09:52.334 "driver_specific": {} 00:09:52.334 } 00:09:52.334 ] 00:09:52.334 23:43:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:52.334 23:43:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:09:52.334 23:43:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:09:52.334 23:43:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:52.335 23:43:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:52.335 23:43:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:52.335 23:43:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:52.335 23:43:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:52.335 23:43:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:52.335 23:43:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:52.335 23:43:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:52.335 23:43:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:52.335 23:43:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:52.335 23:43:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:52.335 23:43:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:52.335 23:43:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:52.335 23:43:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:52.593 23:43:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:52.593 "name": "Existed_Raid", 00:09:52.593 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:52.593 "strip_size_kb": 0, 00:09:52.593 "state": "configuring", 00:09:52.593 "raid_level": "raid1", 00:09:52.593 "superblock": false, 00:09:52.593 "num_base_bdevs": 4, 00:09:52.593 "num_base_bdevs_discovered": 3, 00:09:52.593 "num_base_bdevs_operational": 4, 00:09:52.593 "base_bdevs_list": [ 00:09:52.593 { 00:09:52.593 "name": "BaseBdev1", 00:09:52.593 "uuid": "9d2660e9-c8e2-42e2-8820-cb3af5b7e5ec", 00:09:52.593 "is_configured": true, 00:09:52.593 "data_offset": 0, 00:09:52.593 "data_size": 65536 00:09:52.593 }, 00:09:52.593 { 00:09:52.593 "name": null, 00:09:52.593 "uuid": "d8c630d6-cda4-4dea-b338-ada656de458c", 00:09:52.593 "is_configured": false, 00:09:52.593 "data_offset": 0, 00:09:52.593 "data_size": 65536 00:09:52.593 }, 00:09:52.593 { 00:09:52.593 "name": "BaseBdev3", 00:09:52.593 "uuid": "23ddc50f-e8f5-4eed-ae5f-164b59ab9a0e", 00:09:52.593 "is_configured": true, 00:09:52.593 "data_offset": 0, 00:09:52.593 "data_size": 65536 00:09:52.593 }, 00:09:52.593 { 00:09:52.593 "name": "BaseBdev4", 00:09:52.593 "uuid": "aa840108-0a37-4337-ac98-77500cd6ba01", 00:09:52.593 "is_configured": true, 00:09:52.593 "data_offset": 0, 00:09:52.593 "data_size": 65536 00:09:52.593 } 00:09:52.593 ] 00:09:52.593 }' 00:09:52.593 23:43:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:52.593 23:43:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:52.851 23:43:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:52.851 23:43:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:09:52.851 23:43:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:52.851 23:43:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:52.851 23:43:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:52.851 23:43:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:09:52.851 23:43:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:09:52.851 23:43:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:52.851 23:43:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:52.851 [2024-11-26 23:43:40.852957] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:09:52.851 23:43:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:52.851 23:43:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:09:52.851 23:43:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:52.851 23:43:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:52.851 23:43:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:52.851 23:43:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:52.851 23:43:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:52.851 23:43:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:52.851 23:43:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:52.851 23:43:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:52.851 23:43:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:52.851 23:43:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:52.851 23:43:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:52.851 23:43:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:52.851 23:43:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:52.851 23:43:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:52.851 23:43:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:52.851 "name": "Existed_Raid", 00:09:52.851 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:52.851 "strip_size_kb": 0, 00:09:52.851 "state": "configuring", 00:09:52.851 "raid_level": "raid1", 00:09:52.851 "superblock": false, 00:09:52.851 "num_base_bdevs": 4, 00:09:52.851 "num_base_bdevs_discovered": 2, 00:09:52.851 "num_base_bdevs_operational": 4, 00:09:52.851 "base_bdevs_list": [ 00:09:52.851 { 00:09:52.851 "name": "BaseBdev1", 00:09:52.851 "uuid": "9d2660e9-c8e2-42e2-8820-cb3af5b7e5ec", 00:09:52.851 "is_configured": true, 00:09:52.851 "data_offset": 0, 00:09:52.851 "data_size": 65536 00:09:52.851 }, 00:09:52.851 { 00:09:52.851 "name": null, 00:09:52.851 "uuid": "d8c630d6-cda4-4dea-b338-ada656de458c", 00:09:52.851 "is_configured": false, 00:09:52.851 "data_offset": 0, 00:09:52.851 "data_size": 65536 00:09:52.851 }, 00:09:52.851 { 00:09:52.851 "name": null, 00:09:52.851 "uuid": "23ddc50f-e8f5-4eed-ae5f-164b59ab9a0e", 00:09:52.851 "is_configured": false, 00:09:52.851 "data_offset": 0, 00:09:52.851 "data_size": 65536 00:09:52.851 }, 00:09:52.851 { 00:09:52.851 "name": "BaseBdev4", 00:09:52.851 "uuid": "aa840108-0a37-4337-ac98-77500cd6ba01", 00:09:52.851 "is_configured": true, 00:09:52.851 "data_offset": 0, 00:09:52.851 "data_size": 65536 00:09:52.851 } 00:09:52.851 ] 00:09:52.851 }' 00:09:52.852 23:43:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:52.852 23:43:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:53.432 23:43:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:53.432 23:43:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:09:53.432 23:43:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:53.432 23:43:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:53.432 23:43:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:53.432 23:43:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:09:53.432 23:43:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:09:53.432 23:43:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:53.432 23:43:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:53.432 [2024-11-26 23:43:41.356133] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:53.432 23:43:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:53.432 23:43:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:09:53.432 23:43:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:53.432 23:43:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:53.432 23:43:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:53.432 23:43:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:53.432 23:43:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:53.432 23:43:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:53.432 23:43:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:53.432 23:43:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:53.432 23:43:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:53.432 23:43:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:53.432 23:43:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:53.432 23:43:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:53.432 23:43:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:53.432 23:43:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:53.432 23:43:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:53.432 "name": "Existed_Raid", 00:09:53.432 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:53.432 "strip_size_kb": 0, 00:09:53.432 "state": "configuring", 00:09:53.432 "raid_level": "raid1", 00:09:53.432 "superblock": false, 00:09:53.432 "num_base_bdevs": 4, 00:09:53.432 "num_base_bdevs_discovered": 3, 00:09:53.432 "num_base_bdevs_operational": 4, 00:09:53.432 "base_bdevs_list": [ 00:09:53.432 { 00:09:53.432 "name": "BaseBdev1", 00:09:53.432 "uuid": "9d2660e9-c8e2-42e2-8820-cb3af5b7e5ec", 00:09:53.432 "is_configured": true, 00:09:53.432 "data_offset": 0, 00:09:53.432 "data_size": 65536 00:09:53.432 }, 00:09:53.432 { 00:09:53.432 "name": null, 00:09:53.432 "uuid": "d8c630d6-cda4-4dea-b338-ada656de458c", 00:09:53.432 "is_configured": false, 00:09:53.432 "data_offset": 0, 00:09:53.432 "data_size": 65536 00:09:53.432 }, 00:09:53.432 { 00:09:53.432 "name": "BaseBdev3", 00:09:53.432 "uuid": "23ddc50f-e8f5-4eed-ae5f-164b59ab9a0e", 00:09:53.432 "is_configured": true, 00:09:53.432 "data_offset": 0, 00:09:53.432 "data_size": 65536 00:09:53.432 }, 00:09:53.432 { 00:09:53.432 "name": "BaseBdev4", 00:09:53.432 "uuid": "aa840108-0a37-4337-ac98-77500cd6ba01", 00:09:53.432 "is_configured": true, 00:09:53.432 "data_offset": 0, 00:09:53.432 "data_size": 65536 00:09:53.432 } 00:09:53.432 ] 00:09:53.432 }' 00:09:53.432 23:43:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:53.432 23:43:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:53.690 23:43:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:09:53.690 23:43:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:53.690 23:43:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:53.690 23:43:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:53.690 23:43:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:53.690 23:43:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:09:53.690 23:43:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:09:53.690 23:43:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:53.690 23:43:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:53.690 [2024-11-26 23:43:41.815379] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:09:53.947 23:43:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:53.947 23:43:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:09:53.947 23:43:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:53.947 23:43:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:53.947 23:43:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:53.947 23:43:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:53.947 23:43:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:53.947 23:43:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:53.947 23:43:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:53.947 23:43:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:53.947 23:43:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:53.947 23:43:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:53.947 23:43:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:53.947 23:43:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:53.948 23:43:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:53.948 23:43:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:53.948 23:43:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:53.948 "name": "Existed_Raid", 00:09:53.948 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:53.948 "strip_size_kb": 0, 00:09:53.948 "state": "configuring", 00:09:53.948 "raid_level": "raid1", 00:09:53.948 "superblock": false, 00:09:53.948 "num_base_bdevs": 4, 00:09:53.948 "num_base_bdevs_discovered": 2, 00:09:53.948 "num_base_bdevs_operational": 4, 00:09:53.948 "base_bdevs_list": [ 00:09:53.948 { 00:09:53.948 "name": null, 00:09:53.948 "uuid": "9d2660e9-c8e2-42e2-8820-cb3af5b7e5ec", 00:09:53.948 "is_configured": false, 00:09:53.948 "data_offset": 0, 00:09:53.948 "data_size": 65536 00:09:53.948 }, 00:09:53.948 { 00:09:53.948 "name": null, 00:09:53.948 "uuid": "d8c630d6-cda4-4dea-b338-ada656de458c", 00:09:53.948 "is_configured": false, 00:09:53.948 "data_offset": 0, 00:09:53.948 "data_size": 65536 00:09:53.948 }, 00:09:53.948 { 00:09:53.948 "name": "BaseBdev3", 00:09:53.948 "uuid": "23ddc50f-e8f5-4eed-ae5f-164b59ab9a0e", 00:09:53.948 "is_configured": true, 00:09:53.948 "data_offset": 0, 00:09:53.948 "data_size": 65536 00:09:53.948 }, 00:09:53.948 { 00:09:53.948 "name": "BaseBdev4", 00:09:53.948 "uuid": "aa840108-0a37-4337-ac98-77500cd6ba01", 00:09:53.948 "is_configured": true, 00:09:53.948 "data_offset": 0, 00:09:53.948 "data_size": 65536 00:09:53.948 } 00:09:53.948 ] 00:09:53.948 }' 00:09:53.948 23:43:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:53.948 23:43:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:54.206 23:43:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:54.206 23:43:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:54.206 23:43:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:54.206 23:43:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:09:54.206 23:43:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:54.206 23:43:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:09:54.206 23:43:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:09:54.206 23:43:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:54.206 23:43:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:54.206 [2024-11-26 23:43:42.300974] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:54.206 23:43:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:54.206 23:43:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:09:54.206 23:43:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:54.206 23:43:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:54.206 23:43:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:54.206 23:43:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:54.207 23:43:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:54.207 23:43:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:54.207 23:43:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:54.207 23:43:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:54.207 23:43:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:54.207 23:43:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:54.207 23:43:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:54.207 23:43:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:54.207 23:43:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:54.207 23:43:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:54.465 23:43:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:54.465 "name": "Existed_Raid", 00:09:54.465 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:54.465 "strip_size_kb": 0, 00:09:54.465 "state": "configuring", 00:09:54.465 "raid_level": "raid1", 00:09:54.466 "superblock": false, 00:09:54.466 "num_base_bdevs": 4, 00:09:54.466 "num_base_bdevs_discovered": 3, 00:09:54.466 "num_base_bdevs_operational": 4, 00:09:54.466 "base_bdevs_list": [ 00:09:54.466 { 00:09:54.466 "name": null, 00:09:54.466 "uuid": "9d2660e9-c8e2-42e2-8820-cb3af5b7e5ec", 00:09:54.466 "is_configured": false, 00:09:54.466 "data_offset": 0, 00:09:54.466 "data_size": 65536 00:09:54.466 }, 00:09:54.466 { 00:09:54.466 "name": "BaseBdev2", 00:09:54.466 "uuid": "d8c630d6-cda4-4dea-b338-ada656de458c", 00:09:54.466 "is_configured": true, 00:09:54.466 "data_offset": 0, 00:09:54.466 "data_size": 65536 00:09:54.466 }, 00:09:54.466 { 00:09:54.466 "name": "BaseBdev3", 00:09:54.466 "uuid": "23ddc50f-e8f5-4eed-ae5f-164b59ab9a0e", 00:09:54.466 "is_configured": true, 00:09:54.466 "data_offset": 0, 00:09:54.466 "data_size": 65536 00:09:54.466 }, 00:09:54.466 { 00:09:54.466 "name": "BaseBdev4", 00:09:54.466 "uuid": "aa840108-0a37-4337-ac98-77500cd6ba01", 00:09:54.466 "is_configured": true, 00:09:54.466 "data_offset": 0, 00:09:54.466 "data_size": 65536 00:09:54.466 } 00:09:54.466 ] 00:09:54.466 }' 00:09:54.466 23:43:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:54.466 23:43:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:54.726 23:43:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:54.726 23:43:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:09:54.726 23:43:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:54.726 23:43:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:54.726 23:43:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:54.726 23:43:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:09:54.726 23:43:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:54.726 23:43:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:54.726 23:43:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:09:54.726 23:43:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:54.726 23:43:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:54.726 23:43:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 9d2660e9-c8e2-42e2-8820-cb3af5b7e5ec 00:09:54.726 23:43:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:54.726 23:43:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:54.726 [2024-11-26 23:43:42.810974] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:09:54.726 [2024-11-26 23:43:42.811020] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:09:54.726 [2024-11-26 23:43:42.811031] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:09:54.726 [2024-11-26 23:43:42.811274] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002a10 00:09:54.726 [2024-11-26 23:43:42.811426] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:09:54.726 [2024-11-26 23:43:42.811436] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001c80 00:09:54.726 [2024-11-26 23:43:42.811613] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:54.726 NewBaseBdev 00:09:54.726 23:43:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:54.726 23:43:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:09:54.726 23:43:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:09:54.726 23:43:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:54.726 23:43:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:09:54.726 23:43:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:54.726 23:43:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:54.726 23:43:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:54.726 23:43:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:54.726 23:43:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:54.726 23:43:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:54.726 23:43:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:09:54.726 23:43:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:54.726 23:43:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:54.726 [ 00:09:54.726 { 00:09:54.726 "name": "NewBaseBdev", 00:09:54.726 "aliases": [ 00:09:54.726 "9d2660e9-c8e2-42e2-8820-cb3af5b7e5ec" 00:09:54.726 ], 00:09:54.726 "product_name": "Malloc disk", 00:09:54.726 "block_size": 512, 00:09:54.726 "num_blocks": 65536, 00:09:54.726 "uuid": "9d2660e9-c8e2-42e2-8820-cb3af5b7e5ec", 00:09:54.726 "assigned_rate_limits": { 00:09:54.726 "rw_ios_per_sec": 0, 00:09:54.726 "rw_mbytes_per_sec": 0, 00:09:54.726 "r_mbytes_per_sec": 0, 00:09:54.726 "w_mbytes_per_sec": 0 00:09:54.726 }, 00:09:54.726 "claimed": true, 00:09:54.726 "claim_type": "exclusive_write", 00:09:54.726 "zoned": false, 00:09:54.726 "supported_io_types": { 00:09:54.726 "read": true, 00:09:54.726 "write": true, 00:09:54.726 "unmap": true, 00:09:54.726 "flush": true, 00:09:54.726 "reset": true, 00:09:54.726 "nvme_admin": false, 00:09:54.726 "nvme_io": false, 00:09:54.726 "nvme_io_md": false, 00:09:54.726 "write_zeroes": true, 00:09:54.726 "zcopy": true, 00:09:54.726 "get_zone_info": false, 00:09:54.726 "zone_management": false, 00:09:54.726 "zone_append": false, 00:09:54.726 "compare": false, 00:09:54.726 "compare_and_write": false, 00:09:54.726 "abort": true, 00:09:54.726 "seek_hole": false, 00:09:54.726 "seek_data": false, 00:09:54.726 "copy": true, 00:09:54.726 "nvme_iov_md": false 00:09:54.726 }, 00:09:54.726 "memory_domains": [ 00:09:54.726 { 00:09:54.726 "dma_device_id": "system", 00:09:54.726 "dma_device_type": 1 00:09:54.726 }, 00:09:54.726 { 00:09:54.726 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:54.726 "dma_device_type": 2 00:09:54.726 } 00:09:54.726 ], 00:09:54.726 "driver_specific": {} 00:09:54.726 } 00:09:54.726 ] 00:09:54.726 23:43:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:54.726 23:43:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:09:54.726 23:43:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid1 0 4 00:09:54.726 23:43:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:54.726 23:43:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:54.726 23:43:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:54.726 23:43:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:54.726 23:43:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:54.726 23:43:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:54.726 23:43:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:54.726 23:43:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:54.726 23:43:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:54.726 23:43:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:54.726 23:43:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:54.726 23:43:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:54.726 23:43:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:54.985 23:43:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:54.985 23:43:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:54.985 "name": "Existed_Raid", 00:09:54.985 "uuid": "5b8b3716-9bc2-4cfa-94a9-12299c229bdc", 00:09:54.985 "strip_size_kb": 0, 00:09:54.985 "state": "online", 00:09:54.985 "raid_level": "raid1", 00:09:54.985 "superblock": false, 00:09:54.985 "num_base_bdevs": 4, 00:09:54.985 "num_base_bdevs_discovered": 4, 00:09:54.985 "num_base_bdevs_operational": 4, 00:09:54.985 "base_bdevs_list": [ 00:09:54.985 { 00:09:54.985 "name": "NewBaseBdev", 00:09:54.985 "uuid": "9d2660e9-c8e2-42e2-8820-cb3af5b7e5ec", 00:09:54.985 "is_configured": true, 00:09:54.985 "data_offset": 0, 00:09:54.985 "data_size": 65536 00:09:54.985 }, 00:09:54.985 { 00:09:54.985 "name": "BaseBdev2", 00:09:54.985 "uuid": "d8c630d6-cda4-4dea-b338-ada656de458c", 00:09:54.985 "is_configured": true, 00:09:54.985 "data_offset": 0, 00:09:54.985 "data_size": 65536 00:09:54.985 }, 00:09:54.985 { 00:09:54.985 "name": "BaseBdev3", 00:09:54.985 "uuid": "23ddc50f-e8f5-4eed-ae5f-164b59ab9a0e", 00:09:54.985 "is_configured": true, 00:09:54.985 "data_offset": 0, 00:09:54.985 "data_size": 65536 00:09:54.985 }, 00:09:54.985 { 00:09:54.985 "name": "BaseBdev4", 00:09:54.985 "uuid": "aa840108-0a37-4337-ac98-77500cd6ba01", 00:09:54.985 "is_configured": true, 00:09:54.985 "data_offset": 0, 00:09:54.985 "data_size": 65536 00:09:54.985 } 00:09:54.985 ] 00:09:54.985 }' 00:09:54.985 23:43:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:54.985 23:43:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:55.243 23:43:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:09:55.243 23:43:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:09:55.243 23:43:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:55.243 23:43:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:55.243 23:43:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:09:55.243 23:43:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:55.243 23:43:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:09:55.243 23:43:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:55.243 23:43:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:55.243 23:43:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:55.243 [2024-11-26 23:43:43.282566] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:55.243 23:43:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:55.243 23:43:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:55.243 "name": "Existed_Raid", 00:09:55.243 "aliases": [ 00:09:55.243 "5b8b3716-9bc2-4cfa-94a9-12299c229bdc" 00:09:55.243 ], 00:09:55.243 "product_name": "Raid Volume", 00:09:55.243 "block_size": 512, 00:09:55.243 "num_blocks": 65536, 00:09:55.243 "uuid": "5b8b3716-9bc2-4cfa-94a9-12299c229bdc", 00:09:55.243 "assigned_rate_limits": { 00:09:55.243 "rw_ios_per_sec": 0, 00:09:55.243 "rw_mbytes_per_sec": 0, 00:09:55.243 "r_mbytes_per_sec": 0, 00:09:55.243 "w_mbytes_per_sec": 0 00:09:55.243 }, 00:09:55.243 "claimed": false, 00:09:55.243 "zoned": false, 00:09:55.243 "supported_io_types": { 00:09:55.243 "read": true, 00:09:55.243 "write": true, 00:09:55.244 "unmap": false, 00:09:55.244 "flush": false, 00:09:55.244 "reset": true, 00:09:55.244 "nvme_admin": false, 00:09:55.244 "nvme_io": false, 00:09:55.244 "nvme_io_md": false, 00:09:55.244 "write_zeroes": true, 00:09:55.244 "zcopy": false, 00:09:55.244 "get_zone_info": false, 00:09:55.244 "zone_management": false, 00:09:55.244 "zone_append": false, 00:09:55.244 "compare": false, 00:09:55.244 "compare_and_write": false, 00:09:55.244 "abort": false, 00:09:55.244 "seek_hole": false, 00:09:55.244 "seek_data": false, 00:09:55.244 "copy": false, 00:09:55.244 "nvme_iov_md": false 00:09:55.244 }, 00:09:55.244 "memory_domains": [ 00:09:55.244 { 00:09:55.244 "dma_device_id": "system", 00:09:55.244 "dma_device_type": 1 00:09:55.244 }, 00:09:55.244 { 00:09:55.244 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:55.244 "dma_device_type": 2 00:09:55.244 }, 00:09:55.244 { 00:09:55.244 "dma_device_id": "system", 00:09:55.244 "dma_device_type": 1 00:09:55.244 }, 00:09:55.244 { 00:09:55.244 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:55.244 "dma_device_type": 2 00:09:55.244 }, 00:09:55.244 { 00:09:55.244 "dma_device_id": "system", 00:09:55.244 "dma_device_type": 1 00:09:55.244 }, 00:09:55.244 { 00:09:55.244 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:55.244 "dma_device_type": 2 00:09:55.244 }, 00:09:55.244 { 00:09:55.244 "dma_device_id": "system", 00:09:55.244 "dma_device_type": 1 00:09:55.244 }, 00:09:55.244 { 00:09:55.244 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:55.244 "dma_device_type": 2 00:09:55.244 } 00:09:55.244 ], 00:09:55.244 "driver_specific": { 00:09:55.244 "raid": { 00:09:55.244 "uuid": "5b8b3716-9bc2-4cfa-94a9-12299c229bdc", 00:09:55.244 "strip_size_kb": 0, 00:09:55.244 "state": "online", 00:09:55.244 "raid_level": "raid1", 00:09:55.244 "superblock": false, 00:09:55.244 "num_base_bdevs": 4, 00:09:55.244 "num_base_bdevs_discovered": 4, 00:09:55.244 "num_base_bdevs_operational": 4, 00:09:55.244 "base_bdevs_list": [ 00:09:55.244 { 00:09:55.244 "name": "NewBaseBdev", 00:09:55.244 "uuid": "9d2660e9-c8e2-42e2-8820-cb3af5b7e5ec", 00:09:55.244 "is_configured": true, 00:09:55.244 "data_offset": 0, 00:09:55.244 "data_size": 65536 00:09:55.244 }, 00:09:55.244 { 00:09:55.244 "name": "BaseBdev2", 00:09:55.244 "uuid": "d8c630d6-cda4-4dea-b338-ada656de458c", 00:09:55.244 "is_configured": true, 00:09:55.244 "data_offset": 0, 00:09:55.244 "data_size": 65536 00:09:55.244 }, 00:09:55.244 { 00:09:55.244 "name": "BaseBdev3", 00:09:55.244 "uuid": "23ddc50f-e8f5-4eed-ae5f-164b59ab9a0e", 00:09:55.244 "is_configured": true, 00:09:55.244 "data_offset": 0, 00:09:55.244 "data_size": 65536 00:09:55.244 }, 00:09:55.244 { 00:09:55.244 "name": "BaseBdev4", 00:09:55.244 "uuid": "aa840108-0a37-4337-ac98-77500cd6ba01", 00:09:55.244 "is_configured": true, 00:09:55.244 "data_offset": 0, 00:09:55.244 "data_size": 65536 00:09:55.244 } 00:09:55.244 ] 00:09:55.244 } 00:09:55.244 } 00:09:55.244 }' 00:09:55.244 23:43:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:55.503 23:43:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:09:55.503 BaseBdev2 00:09:55.503 BaseBdev3 00:09:55.503 BaseBdev4' 00:09:55.503 23:43:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:55.503 23:43:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:55.503 23:43:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:55.503 23:43:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:09:55.504 23:43:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:55.504 23:43:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:55.504 23:43:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:55.504 23:43:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:55.504 23:43:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:55.504 23:43:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:55.504 23:43:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:55.504 23:43:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:55.504 23:43:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:09:55.504 23:43:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:55.504 23:43:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:55.504 23:43:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:55.504 23:43:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:55.504 23:43:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:55.504 23:43:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:55.504 23:43:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:09:55.504 23:43:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:55.504 23:43:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:55.504 23:43:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:55.504 23:43:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:55.504 23:43:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:55.504 23:43:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:55.504 23:43:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:55.504 23:43:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:09:55.504 23:43:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:55.504 23:43:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:55.504 23:43:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:55.504 23:43:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:55.504 23:43:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:55.504 23:43:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:55.504 23:43:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:55.504 23:43:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:55.504 23:43:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:55.504 [2024-11-26 23:43:43.605677] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:55.504 [2024-11-26 23:43:43.605703] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:55.504 [2024-11-26 23:43:43.605777] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:55.504 [2024-11-26 23:43:43.606020] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:55.504 [2024-11-26 23:43:43.606034] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name Existed_Raid, state offline 00:09:55.504 23:43:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:55.504 23:43:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 83690 00:09:55.504 23:43:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # '[' -z 83690 ']' 00:09:55.504 23:43:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@958 -- # kill -0 83690 00:09:55.504 23:43:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # uname 00:09:55.504 23:43:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:09:55.504 23:43:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 83690 00:09:55.763 23:43:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:09:55.763 23:43:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:09:55.763 23:43:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 83690' 00:09:55.763 killing process with pid 83690 00:09:55.763 23:43:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@973 -- # kill 83690 00:09:55.763 [2024-11-26 23:43:43.654424] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:09:55.763 23:43:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@978 -- # wait 83690 00:09:55.763 [2024-11-26 23:43:43.695517] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:09:56.021 23:43:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:09:56.021 00:09:56.021 real 0m9.258s 00:09:56.021 user 0m15.885s 00:09:56.021 sys 0m1.868s 00:09:56.021 23:43:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:09:56.021 23:43:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:56.021 ************************************ 00:09:56.021 END TEST raid_state_function_test 00:09:56.021 ************************************ 00:09:56.021 23:43:43 bdev_raid -- bdev/bdev_raid.sh@969 -- # run_test raid_state_function_test_sb raid_state_function_test raid1 4 true 00:09:56.021 23:43:43 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:09:56.021 23:43:43 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:09:56.021 23:43:43 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:09:56.021 ************************************ 00:09:56.021 START TEST raid_state_function_test_sb 00:09:56.021 ************************************ 00:09:56.021 23:43:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1129 -- # raid_state_function_test raid1 4 true 00:09:56.021 23:43:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=raid1 00:09:56.021 23:43:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=4 00:09:56.021 23:43:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:09:56.021 23:43:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:09:56.021 23:43:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:09:56.021 23:43:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:56.021 23:43:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:09:56.021 23:43:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:56.021 23:43:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:56.021 23:43:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:09:56.021 23:43:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:56.021 23:43:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:56.021 23:43:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:09:56.021 23:43:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:56.021 23:43:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:56.021 23:43:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev4 00:09:56.021 23:43:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:56.021 23:43:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:56.021 23:43:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:09:56.021 23:43:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:09:56.021 23:43:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:09:56.021 23:43:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:09:56.021 23:43:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:09:56.021 23:43:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:09:56.021 23:43:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' raid1 '!=' raid1 ']' 00:09:56.021 23:43:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@219 -- # strip_size=0 00:09:56.021 23:43:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:09:56.021 23:43:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:09:56.021 Process raid pid: 84334 00:09:56.021 23:43:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=84334 00:09:56.022 23:43:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:09:56.022 23:43:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 84334' 00:09:56.022 23:43:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 84334 00:09:56.022 23:43:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@835 -- # '[' -z 84334 ']' 00:09:56.022 23:43:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:56.022 23:43:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@840 -- # local max_retries=100 00:09:56.022 23:43:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:56.022 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:56.022 23:43:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@844 -- # xtrace_disable 00:09:56.022 23:43:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:56.022 [2024-11-26 23:43:44.074611] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:09:56.022 [2024-11-26 23:43:44.074821] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:09:56.280 [2024-11-26 23:43:44.229558] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:56.280 [2024-11-26 23:43:44.254472] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:09:56.280 [2024-11-26 23:43:44.296457] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:56.280 [2024-11-26 23:43:44.296578] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:56.848 23:43:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:09:56.848 23:43:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@868 -- # return 0 00:09:56.848 23:43:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:09:56.848 23:43:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:56.848 23:43:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:56.848 [2024-11-26 23:43:44.898912] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:56.848 [2024-11-26 23:43:44.898964] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:56.848 [2024-11-26 23:43:44.898982] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:56.848 [2024-11-26 23:43:44.898993] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:56.848 [2024-11-26 23:43:44.898999] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:56.848 [2024-11-26 23:43:44.899009] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:56.848 [2024-11-26 23:43:44.899015] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:09:56.848 [2024-11-26 23:43:44.899024] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:09:56.848 23:43:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:56.848 23:43:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:09:56.848 23:43:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:56.848 23:43:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:56.848 23:43:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:56.848 23:43:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:56.848 23:43:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:56.848 23:43:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:56.848 23:43:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:56.848 23:43:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:56.848 23:43:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:56.848 23:43:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:56.848 23:43:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:56.848 23:43:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:56.848 23:43:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:56.848 23:43:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:56.848 23:43:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:56.848 "name": "Existed_Raid", 00:09:56.848 "uuid": "85589fbf-4856-4199-b4ac-3d1dd8f19baf", 00:09:56.848 "strip_size_kb": 0, 00:09:56.848 "state": "configuring", 00:09:56.848 "raid_level": "raid1", 00:09:56.848 "superblock": true, 00:09:56.848 "num_base_bdevs": 4, 00:09:56.848 "num_base_bdevs_discovered": 0, 00:09:56.848 "num_base_bdevs_operational": 4, 00:09:56.848 "base_bdevs_list": [ 00:09:56.848 { 00:09:56.848 "name": "BaseBdev1", 00:09:56.848 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:56.848 "is_configured": false, 00:09:56.848 "data_offset": 0, 00:09:56.848 "data_size": 0 00:09:56.848 }, 00:09:56.848 { 00:09:56.848 "name": "BaseBdev2", 00:09:56.848 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:56.848 "is_configured": false, 00:09:56.848 "data_offset": 0, 00:09:56.848 "data_size": 0 00:09:56.848 }, 00:09:56.848 { 00:09:56.848 "name": "BaseBdev3", 00:09:56.848 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:56.848 "is_configured": false, 00:09:56.848 "data_offset": 0, 00:09:56.848 "data_size": 0 00:09:56.848 }, 00:09:56.848 { 00:09:56.848 "name": "BaseBdev4", 00:09:56.848 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:56.848 "is_configured": false, 00:09:56.848 "data_offset": 0, 00:09:56.848 "data_size": 0 00:09:56.848 } 00:09:56.848 ] 00:09:56.848 }' 00:09:56.848 23:43:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:56.848 23:43:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:57.417 23:43:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:57.417 23:43:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:57.417 23:43:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:57.417 [2024-11-26 23:43:45.314136] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:57.417 [2024-11-26 23:43:45.314237] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:09:57.417 23:43:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:57.417 23:43:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:09:57.417 23:43:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:57.417 23:43:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:57.417 [2024-11-26 23:43:45.326129] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:57.417 [2024-11-26 23:43:45.326231] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:57.417 [2024-11-26 23:43:45.326258] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:57.417 [2024-11-26 23:43:45.326281] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:57.417 [2024-11-26 23:43:45.326298] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:57.417 [2024-11-26 23:43:45.326318] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:57.417 [2024-11-26 23:43:45.326335] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:09:57.417 [2024-11-26 23:43:45.326367] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:09:57.417 23:43:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:57.417 23:43:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:09:57.417 23:43:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:57.417 23:43:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:57.417 [2024-11-26 23:43:45.347008] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:57.417 BaseBdev1 00:09:57.417 23:43:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:57.417 23:43:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:09:57.417 23:43:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:09:57.417 23:43:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:57.417 23:43:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:09:57.417 23:43:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:57.417 23:43:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:57.417 23:43:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:57.417 23:43:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:57.417 23:43:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:57.417 23:43:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:57.417 23:43:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:09:57.417 23:43:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:57.417 23:43:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:57.417 [ 00:09:57.417 { 00:09:57.417 "name": "BaseBdev1", 00:09:57.417 "aliases": [ 00:09:57.417 "12a9b3d5-a51d-4cc9-ae35-f3f0e67f18b0" 00:09:57.417 ], 00:09:57.417 "product_name": "Malloc disk", 00:09:57.417 "block_size": 512, 00:09:57.417 "num_blocks": 65536, 00:09:57.417 "uuid": "12a9b3d5-a51d-4cc9-ae35-f3f0e67f18b0", 00:09:57.417 "assigned_rate_limits": { 00:09:57.417 "rw_ios_per_sec": 0, 00:09:57.417 "rw_mbytes_per_sec": 0, 00:09:57.417 "r_mbytes_per_sec": 0, 00:09:57.417 "w_mbytes_per_sec": 0 00:09:57.417 }, 00:09:57.417 "claimed": true, 00:09:57.417 "claim_type": "exclusive_write", 00:09:57.417 "zoned": false, 00:09:57.417 "supported_io_types": { 00:09:57.417 "read": true, 00:09:57.417 "write": true, 00:09:57.417 "unmap": true, 00:09:57.417 "flush": true, 00:09:57.417 "reset": true, 00:09:57.417 "nvme_admin": false, 00:09:57.417 "nvme_io": false, 00:09:57.417 "nvme_io_md": false, 00:09:57.417 "write_zeroes": true, 00:09:57.417 "zcopy": true, 00:09:57.417 "get_zone_info": false, 00:09:57.417 "zone_management": false, 00:09:57.417 "zone_append": false, 00:09:57.417 "compare": false, 00:09:57.417 "compare_and_write": false, 00:09:57.417 "abort": true, 00:09:57.417 "seek_hole": false, 00:09:57.417 "seek_data": false, 00:09:57.417 "copy": true, 00:09:57.417 "nvme_iov_md": false 00:09:57.417 }, 00:09:57.417 "memory_domains": [ 00:09:57.417 { 00:09:57.417 "dma_device_id": "system", 00:09:57.417 "dma_device_type": 1 00:09:57.417 }, 00:09:57.417 { 00:09:57.417 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:57.417 "dma_device_type": 2 00:09:57.417 } 00:09:57.417 ], 00:09:57.417 "driver_specific": {} 00:09:57.417 } 00:09:57.417 ] 00:09:57.417 23:43:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:57.417 23:43:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:09:57.417 23:43:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:09:57.417 23:43:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:57.417 23:43:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:57.417 23:43:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:57.417 23:43:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:57.417 23:43:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:57.417 23:43:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:57.417 23:43:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:57.417 23:43:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:57.417 23:43:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:57.417 23:43:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:57.417 23:43:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:57.417 23:43:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:57.417 23:43:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:57.417 23:43:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:57.417 23:43:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:57.417 "name": "Existed_Raid", 00:09:57.417 "uuid": "095940eb-0d1c-44ab-877b-8f049022219a", 00:09:57.417 "strip_size_kb": 0, 00:09:57.417 "state": "configuring", 00:09:57.417 "raid_level": "raid1", 00:09:57.417 "superblock": true, 00:09:57.417 "num_base_bdevs": 4, 00:09:57.417 "num_base_bdevs_discovered": 1, 00:09:57.417 "num_base_bdevs_operational": 4, 00:09:57.417 "base_bdevs_list": [ 00:09:57.417 { 00:09:57.417 "name": "BaseBdev1", 00:09:57.417 "uuid": "12a9b3d5-a51d-4cc9-ae35-f3f0e67f18b0", 00:09:57.417 "is_configured": true, 00:09:57.417 "data_offset": 2048, 00:09:57.417 "data_size": 63488 00:09:57.417 }, 00:09:57.417 { 00:09:57.417 "name": "BaseBdev2", 00:09:57.417 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:57.417 "is_configured": false, 00:09:57.417 "data_offset": 0, 00:09:57.417 "data_size": 0 00:09:57.417 }, 00:09:57.417 { 00:09:57.417 "name": "BaseBdev3", 00:09:57.417 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:57.417 "is_configured": false, 00:09:57.417 "data_offset": 0, 00:09:57.417 "data_size": 0 00:09:57.417 }, 00:09:57.417 { 00:09:57.417 "name": "BaseBdev4", 00:09:57.417 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:57.417 "is_configured": false, 00:09:57.417 "data_offset": 0, 00:09:57.418 "data_size": 0 00:09:57.418 } 00:09:57.418 ] 00:09:57.418 }' 00:09:57.418 23:43:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:57.418 23:43:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:57.985 23:43:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:57.985 23:43:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:57.985 23:43:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:57.985 [2024-11-26 23:43:45.830244] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:57.985 [2024-11-26 23:43:45.830377] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:09:57.985 23:43:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:57.985 23:43:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:09:57.985 23:43:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:57.985 23:43:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:57.985 [2024-11-26 23:43:45.842276] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:57.985 [2024-11-26 23:43:45.844127] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:57.985 [2024-11-26 23:43:45.844164] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:57.985 [2024-11-26 23:43:45.844173] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:57.985 [2024-11-26 23:43:45.844182] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:57.985 [2024-11-26 23:43:45.844188] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:09:57.985 [2024-11-26 23:43:45.844196] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:09:57.985 23:43:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:57.985 23:43:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:09:57.985 23:43:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:57.985 23:43:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:09:57.985 23:43:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:57.985 23:43:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:57.985 23:43:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:57.985 23:43:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:57.985 23:43:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:57.985 23:43:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:57.985 23:43:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:57.985 23:43:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:57.985 23:43:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:57.985 23:43:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:57.985 23:43:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:57.985 23:43:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:57.985 23:43:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:57.985 23:43:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:57.985 23:43:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:57.985 "name": "Existed_Raid", 00:09:57.985 "uuid": "69ec4f5c-23b9-42c5-b1aa-4abd32961f6d", 00:09:57.985 "strip_size_kb": 0, 00:09:57.985 "state": "configuring", 00:09:57.985 "raid_level": "raid1", 00:09:57.985 "superblock": true, 00:09:57.985 "num_base_bdevs": 4, 00:09:57.985 "num_base_bdevs_discovered": 1, 00:09:57.985 "num_base_bdevs_operational": 4, 00:09:57.985 "base_bdevs_list": [ 00:09:57.985 { 00:09:57.985 "name": "BaseBdev1", 00:09:57.985 "uuid": "12a9b3d5-a51d-4cc9-ae35-f3f0e67f18b0", 00:09:57.985 "is_configured": true, 00:09:57.985 "data_offset": 2048, 00:09:57.985 "data_size": 63488 00:09:57.985 }, 00:09:57.985 { 00:09:57.985 "name": "BaseBdev2", 00:09:57.985 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:57.985 "is_configured": false, 00:09:57.985 "data_offset": 0, 00:09:57.985 "data_size": 0 00:09:57.985 }, 00:09:57.985 { 00:09:57.985 "name": "BaseBdev3", 00:09:57.985 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:57.985 "is_configured": false, 00:09:57.985 "data_offset": 0, 00:09:57.985 "data_size": 0 00:09:57.985 }, 00:09:57.985 { 00:09:57.985 "name": "BaseBdev4", 00:09:57.985 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:57.985 "is_configured": false, 00:09:57.985 "data_offset": 0, 00:09:57.985 "data_size": 0 00:09:57.985 } 00:09:57.985 ] 00:09:57.985 }' 00:09:57.985 23:43:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:57.985 23:43:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:58.245 23:43:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:09:58.245 23:43:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:58.245 23:43:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:58.245 [2024-11-26 23:43:46.300542] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:58.245 BaseBdev2 00:09:58.245 23:43:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:58.245 23:43:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:09:58.245 23:43:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:09:58.245 23:43:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:58.245 23:43:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:09:58.245 23:43:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:58.245 23:43:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:58.245 23:43:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:58.245 23:43:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:58.245 23:43:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:58.245 23:43:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:58.245 23:43:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:09:58.245 23:43:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:58.245 23:43:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:58.245 [ 00:09:58.245 { 00:09:58.245 "name": "BaseBdev2", 00:09:58.245 "aliases": [ 00:09:58.245 "6399de16-6542-4bd7-8ad8-8bc2959500fc" 00:09:58.245 ], 00:09:58.245 "product_name": "Malloc disk", 00:09:58.245 "block_size": 512, 00:09:58.245 "num_blocks": 65536, 00:09:58.245 "uuid": "6399de16-6542-4bd7-8ad8-8bc2959500fc", 00:09:58.245 "assigned_rate_limits": { 00:09:58.245 "rw_ios_per_sec": 0, 00:09:58.245 "rw_mbytes_per_sec": 0, 00:09:58.245 "r_mbytes_per_sec": 0, 00:09:58.245 "w_mbytes_per_sec": 0 00:09:58.245 }, 00:09:58.245 "claimed": true, 00:09:58.245 "claim_type": "exclusive_write", 00:09:58.245 "zoned": false, 00:09:58.245 "supported_io_types": { 00:09:58.245 "read": true, 00:09:58.245 "write": true, 00:09:58.245 "unmap": true, 00:09:58.245 "flush": true, 00:09:58.245 "reset": true, 00:09:58.245 "nvme_admin": false, 00:09:58.245 "nvme_io": false, 00:09:58.245 "nvme_io_md": false, 00:09:58.245 "write_zeroes": true, 00:09:58.245 "zcopy": true, 00:09:58.245 "get_zone_info": false, 00:09:58.245 "zone_management": false, 00:09:58.245 "zone_append": false, 00:09:58.245 "compare": false, 00:09:58.245 "compare_and_write": false, 00:09:58.245 "abort": true, 00:09:58.245 "seek_hole": false, 00:09:58.245 "seek_data": false, 00:09:58.245 "copy": true, 00:09:58.245 "nvme_iov_md": false 00:09:58.245 }, 00:09:58.245 "memory_domains": [ 00:09:58.245 { 00:09:58.245 "dma_device_id": "system", 00:09:58.245 "dma_device_type": 1 00:09:58.245 }, 00:09:58.245 { 00:09:58.245 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:58.245 "dma_device_type": 2 00:09:58.245 } 00:09:58.245 ], 00:09:58.245 "driver_specific": {} 00:09:58.245 } 00:09:58.245 ] 00:09:58.245 23:43:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:58.245 23:43:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:09:58.245 23:43:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:09:58.245 23:43:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:58.245 23:43:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:09:58.245 23:43:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:58.245 23:43:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:58.245 23:43:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:58.245 23:43:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:58.245 23:43:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:58.245 23:43:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:58.245 23:43:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:58.245 23:43:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:58.245 23:43:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:58.245 23:43:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:58.245 23:43:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:58.245 23:43:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:58.245 23:43:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:58.245 23:43:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:58.504 23:43:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:58.504 "name": "Existed_Raid", 00:09:58.504 "uuid": "69ec4f5c-23b9-42c5-b1aa-4abd32961f6d", 00:09:58.504 "strip_size_kb": 0, 00:09:58.504 "state": "configuring", 00:09:58.504 "raid_level": "raid1", 00:09:58.504 "superblock": true, 00:09:58.504 "num_base_bdevs": 4, 00:09:58.504 "num_base_bdevs_discovered": 2, 00:09:58.504 "num_base_bdevs_operational": 4, 00:09:58.504 "base_bdevs_list": [ 00:09:58.504 { 00:09:58.504 "name": "BaseBdev1", 00:09:58.504 "uuid": "12a9b3d5-a51d-4cc9-ae35-f3f0e67f18b0", 00:09:58.504 "is_configured": true, 00:09:58.504 "data_offset": 2048, 00:09:58.504 "data_size": 63488 00:09:58.504 }, 00:09:58.504 { 00:09:58.504 "name": "BaseBdev2", 00:09:58.504 "uuid": "6399de16-6542-4bd7-8ad8-8bc2959500fc", 00:09:58.504 "is_configured": true, 00:09:58.504 "data_offset": 2048, 00:09:58.504 "data_size": 63488 00:09:58.504 }, 00:09:58.504 { 00:09:58.504 "name": "BaseBdev3", 00:09:58.504 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:58.504 "is_configured": false, 00:09:58.504 "data_offset": 0, 00:09:58.504 "data_size": 0 00:09:58.504 }, 00:09:58.504 { 00:09:58.504 "name": "BaseBdev4", 00:09:58.504 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:58.504 "is_configured": false, 00:09:58.504 "data_offset": 0, 00:09:58.504 "data_size": 0 00:09:58.504 } 00:09:58.504 ] 00:09:58.504 }' 00:09:58.504 23:43:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:58.504 23:43:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:58.763 23:43:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:09:58.763 23:43:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:58.763 23:43:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:58.763 [2024-11-26 23:43:46.812232] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:58.763 BaseBdev3 00:09:58.763 23:43:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:58.763 23:43:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:09:58.763 23:43:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:09:58.763 23:43:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:58.763 23:43:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:09:58.763 23:43:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:58.763 23:43:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:58.763 23:43:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:58.763 23:43:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:58.763 23:43:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:58.763 23:43:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:58.763 23:43:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:09:58.763 23:43:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:58.763 23:43:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:58.763 [ 00:09:58.763 { 00:09:58.763 "name": "BaseBdev3", 00:09:58.763 "aliases": [ 00:09:58.763 "d42f72aa-d487-4fa2-8a4e-861ab8ba0e60" 00:09:58.763 ], 00:09:58.763 "product_name": "Malloc disk", 00:09:58.763 "block_size": 512, 00:09:58.763 "num_blocks": 65536, 00:09:58.763 "uuid": "d42f72aa-d487-4fa2-8a4e-861ab8ba0e60", 00:09:58.763 "assigned_rate_limits": { 00:09:58.763 "rw_ios_per_sec": 0, 00:09:58.763 "rw_mbytes_per_sec": 0, 00:09:58.763 "r_mbytes_per_sec": 0, 00:09:58.763 "w_mbytes_per_sec": 0 00:09:58.763 }, 00:09:58.763 "claimed": true, 00:09:58.763 "claim_type": "exclusive_write", 00:09:58.763 "zoned": false, 00:09:58.763 "supported_io_types": { 00:09:58.763 "read": true, 00:09:58.763 "write": true, 00:09:58.763 "unmap": true, 00:09:58.763 "flush": true, 00:09:58.763 "reset": true, 00:09:58.763 "nvme_admin": false, 00:09:58.763 "nvme_io": false, 00:09:58.763 "nvme_io_md": false, 00:09:58.763 "write_zeroes": true, 00:09:58.763 "zcopy": true, 00:09:58.763 "get_zone_info": false, 00:09:58.763 "zone_management": false, 00:09:58.763 "zone_append": false, 00:09:58.763 "compare": false, 00:09:58.763 "compare_and_write": false, 00:09:58.763 "abort": true, 00:09:58.763 "seek_hole": false, 00:09:58.763 "seek_data": false, 00:09:58.763 "copy": true, 00:09:58.763 "nvme_iov_md": false 00:09:58.763 }, 00:09:58.763 "memory_domains": [ 00:09:58.763 { 00:09:58.763 "dma_device_id": "system", 00:09:58.763 "dma_device_type": 1 00:09:58.763 }, 00:09:58.763 { 00:09:58.763 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:58.763 "dma_device_type": 2 00:09:58.763 } 00:09:58.763 ], 00:09:58.763 "driver_specific": {} 00:09:58.763 } 00:09:58.763 ] 00:09:58.763 23:43:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:58.763 23:43:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:09:58.763 23:43:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:09:58.763 23:43:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:58.763 23:43:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:09:58.763 23:43:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:58.763 23:43:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:58.763 23:43:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:58.763 23:43:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:58.763 23:43:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:58.763 23:43:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:58.763 23:43:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:58.763 23:43:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:58.763 23:43:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:58.763 23:43:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:58.763 23:43:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:58.763 23:43:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:58.763 23:43:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:58.763 23:43:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:59.022 23:43:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:59.022 "name": "Existed_Raid", 00:09:59.022 "uuid": "69ec4f5c-23b9-42c5-b1aa-4abd32961f6d", 00:09:59.022 "strip_size_kb": 0, 00:09:59.022 "state": "configuring", 00:09:59.022 "raid_level": "raid1", 00:09:59.022 "superblock": true, 00:09:59.022 "num_base_bdevs": 4, 00:09:59.022 "num_base_bdevs_discovered": 3, 00:09:59.022 "num_base_bdevs_operational": 4, 00:09:59.022 "base_bdevs_list": [ 00:09:59.022 { 00:09:59.022 "name": "BaseBdev1", 00:09:59.022 "uuid": "12a9b3d5-a51d-4cc9-ae35-f3f0e67f18b0", 00:09:59.022 "is_configured": true, 00:09:59.022 "data_offset": 2048, 00:09:59.022 "data_size": 63488 00:09:59.022 }, 00:09:59.022 { 00:09:59.022 "name": "BaseBdev2", 00:09:59.022 "uuid": "6399de16-6542-4bd7-8ad8-8bc2959500fc", 00:09:59.022 "is_configured": true, 00:09:59.022 "data_offset": 2048, 00:09:59.022 "data_size": 63488 00:09:59.022 }, 00:09:59.022 { 00:09:59.022 "name": "BaseBdev3", 00:09:59.022 "uuid": "d42f72aa-d487-4fa2-8a4e-861ab8ba0e60", 00:09:59.022 "is_configured": true, 00:09:59.022 "data_offset": 2048, 00:09:59.022 "data_size": 63488 00:09:59.022 }, 00:09:59.022 { 00:09:59.022 "name": "BaseBdev4", 00:09:59.022 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:59.022 "is_configured": false, 00:09:59.022 "data_offset": 0, 00:09:59.022 "data_size": 0 00:09:59.022 } 00:09:59.022 ] 00:09:59.022 }' 00:09:59.022 23:43:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:59.022 23:43:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:59.281 23:43:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:09:59.281 23:43:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:59.281 23:43:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:59.281 [2024-11-26 23:43:47.306326] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:09:59.281 [2024-11-26 23:43:47.306538] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:09:59.281 [2024-11-26 23:43:47.306560] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:09:59.281 BaseBdev4 00:09:59.281 [2024-11-26 23:43:47.306852] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:09:59.281 [2024-11-26 23:43:47.306992] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:09:59.281 [2024-11-26 23:43:47.307011] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:09:59.281 [2024-11-26 23:43:47.307150] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:59.281 23:43:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:59.281 23:43:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev4 00:09:59.281 23:43:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:09:59.281 23:43:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:59.281 23:43:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:09:59.281 23:43:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:59.281 23:43:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:59.281 23:43:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:59.281 23:43:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:59.281 23:43:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:59.281 23:43:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:59.281 23:43:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:09:59.281 23:43:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:59.281 23:43:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:59.281 [ 00:09:59.281 { 00:09:59.281 "name": "BaseBdev4", 00:09:59.281 "aliases": [ 00:09:59.281 "2e22f21a-7f1c-47ca-af0c-8bd216d29a40" 00:09:59.281 ], 00:09:59.281 "product_name": "Malloc disk", 00:09:59.281 "block_size": 512, 00:09:59.281 "num_blocks": 65536, 00:09:59.281 "uuid": "2e22f21a-7f1c-47ca-af0c-8bd216d29a40", 00:09:59.281 "assigned_rate_limits": { 00:09:59.281 "rw_ios_per_sec": 0, 00:09:59.281 "rw_mbytes_per_sec": 0, 00:09:59.281 "r_mbytes_per_sec": 0, 00:09:59.281 "w_mbytes_per_sec": 0 00:09:59.281 }, 00:09:59.281 "claimed": true, 00:09:59.281 "claim_type": "exclusive_write", 00:09:59.281 "zoned": false, 00:09:59.281 "supported_io_types": { 00:09:59.281 "read": true, 00:09:59.281 "write": true, 00:09:59.281 "unmap": true, 00:09:59.281 "flush": true, 00:09:59.281 "reset": true, 00:09:59.281 "nvme_admin": false, 00:09:59.281 "nvme_io": false, 00:09:59.281 "nvme_io_md": false, 00:09:59.281 "write_zeroes": true, 00:09:59.281 "zcopy": true, 00:09:59.281 "get_zone_info": false, 00:09:59.281 "zone_management": false, 00:09:59.281 "zone_append": false, 00:09:59.281 "compare": false, 00:09:59.281 "compare_and_write": false, 00:09:59.281 "abort": true, 00:09:59.281 "seek_hole": false, 00:09:59.281 "seek_data": false, 00:09:59.281 "copy": true, 00:09:59.281 "nvme_iov_md": false 00:09:59.281 }, 00:09:59.281 "memory_domains": [ 00:09:59.281 { 00:09:59.281 "dma_device_id": "system", 00:09:59.281 "dma_device_type": 1 00:09:59.281 }, 00:09:59.281 { 00:09:59.281 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:59.281 "dma_device_type": 2 00:09:59.281 } 00:09:59.281 ], 00:09:59.281 "driver_specific": {} 00:09:59.281 } 00:09:59.281 ] 00:09:59.281 23:43:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:59.281 23:43:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:09:59.281 23:43:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:09:59.281 23:43:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:59.281 23:43:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid1 0 4 00:09:59.281 23:43:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:59.281 23:43:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:59.281 23:43:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:59.281 23:43:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:59.281 23:43:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:59.281 23:43:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:59.281 23:43:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:59.281 23:43:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:59.281 23:43:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:59.281 23:43:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:59.281 23:43:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:59.281 23:43:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:59.281 23:43:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:59.281 23:43:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:59.281 23:43:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:59.281 "name": "Existed_Raid", 00:09:59.281 "uuid": "69ec4f5c-23b9-42c5-b1aa-4abd32961f6d", 00:09:59.281 "strip_size_kb": 0, 00:09:59.281 "state": "online", 00:09:59.281 "raid_level": "raid1", 00:09:59.281 "superblock": true, 00:09:59.281 "num_base_bdevs": 4, 00:09:59.281 "num_base_bdevs_discovered": 4, 00:09:59.281 "num_base_bdevs_operational": 4, 00:09:59.281 "base_bdevs_list": [ 00:09:59.281 { 00:09:59.281 "name": "BaseBdev1", 00:09:59.281 "uuid": "12a9b3d5-a51d-4cc9-ae35-f3f0e67f18b0", 00:09:59.281 "is_configured": true, 00:09:59.281 "data_offset": 2048, 00:09:59.281 "data_size": 63488 00:09:59.281 }, 00:09:59.281 { 00:09:59.281 "name": "BaseBdev2", 00:09:59.281 "uuid": "6399de16-6542-4bd7-8ad8-8bc2959500fc", 00:09:59.281 "is_configured": true, 00:09:59.281 "data_offset": 2048, 00:09:59.281 "data_size": 63488 00:09:59.281 }, 00:09:59.281 { 00:09:59.281 "name": "BaseBdev3", 00:09:59.281 "uuid": "d42f72aa-d487-4fa2-8a4e-861ab8ba0e60", 00:09:59.281 "is_configured": true, 00:09:59.281 "data_offset": 2048, 00:09:59.281 "data_size": 63488 00:09:59.281 }, 00:09:59.281 { 00:09:59.281 "name": "BaseBdev4", 00:09:59.281 "uuid": "2e22f21a-7f1c-47ca-af0c-8bd216d29a40", 00:09:59.281 "is_configured": true, 00:09:59.281 "data_offset": 2048, 00:09:59.281 "data_size": 63488 00:09:59.281 } 00:09:59.281 ] 00:09:59.281 }' 00:09:59.281 23:43:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:59.281 23:43:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:59.849 23:43:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:09:59.849 23:43:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:09:59.849 23:43:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:59.849 23:43:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:59.849 23:43:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:09:59.849 23:43:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:59.849 23:43:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:09:59.849 23:43:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:59.849 23:43:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:59.849 23:43:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:59.849 [2024-11-26 23:43:47.781926] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:59.849 23:43:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:59.849 23:43:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:59.849 "name": "Existed_Raid", 00:09:59.849 "aliases": [ 00:09:59.849 "69ec4f5c-23b9-42c5-b1aa-4abd32961f6d" 00:09:59.849 ], 00:09:59.849 "product_name": "Raid Volume", 00:09:59.849 "block_size": 512, 00:09:59.849 "num_blocks": 63488, 00:09:59.849 "uuid": "69ec4f5c-23b9-42c5-b1aa-4abd32961f6d", 00:09:59.849 "assigned_rate_limits": { 00:09:59.849 "rw_ios_per_sec": 0, 00:09:59.849 "rw_mbytes_per_sec": 0, 00:09:59.849 "r_mbytes_per_sec": 0, 00:09:59.849 "w_mbytes_per_sec": 0 00:09:59.849 }, 00:09:59.849 "claimed": false, 00:09:59.849 "zoned": false, 00:09:59.849 "supported_io_types": { 00:09:59.849 "read": true, 00:09:59.849 "write": true, 00:09:59.849 "unmap": false, 00:09:59.849 "flush": false, 00:09:59.849 "reset": true, 00:09:59.849 "nvme_admin": false, 00:09:59.849 "nvme_io": false, 00:09:59.849 "nvme_io_md": false, 00:09:59.849 "write_zeroes": true, 00:09:59.849 "zcopy": false, 00:09:59.849 "get_zone_info": false, 00:09:59.849 "zone_management": false, 00:09:59.849 "zone_append": false, 00:09:59.849 "compare": false, 00:09:59.849 "compare_and_write": false, 00:09:59.849 "abort": false, 00:09:59.849 "seek_hole": false, 00:09:59.849 "seek_data": false, 00:09:59.849 "copy": false, 00:09:59.849 "nvme_iov_md": false 00:09:59.849 }, 00:09:59.849 "memory_domains": [ 00:09:59.849 { 00:09:59.849 "dma_device_id": "system", 00:09:59.849 "dma_device_type": 1 00:09:59.849 }, 00:09:59.849 { 00:09:59.849 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:59.849 "dma_device_type": 2 00:09:59.849 }, 00:09:59.849 { 00:09:59.849 "dma_device_id": "system", 00:09:59.849 "dma_device_type": 1 00:09:59.849 }, 00:09:59.849 { 00:09:59.849 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:59.849 "dma_device_type": 2 00:09:59.849 }, 00:09:59.849 { 00:09:59.849 "dma_device_id": "system", 00:09:59.849 "dma_device_type": 1 00:09:59.849 }, 00:09:59.849 { 00:09:59.849 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:59.849 "dma_device_type": 2 00:09:59.849 }, 00:09:59.849 { 00:09:59.849 "dma_device_id": "system", 00:09:59.849 "dma_device_type": 1 00:09:59.849 }, 00:09:59.849 { 00:09:59.849 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:59.849 "dma_device_type": 2 00:09:59.849 } 00:09:59.849 ], 00:09:59.849 "driver_specific": { 00:09:59.849 "raid": { 00:09:59.849 "uuid": "69ec4f5c-23b9-42c5-b1aa-4abd32961f6d", 00:09:59.849 "strip_size_kb": 0, 00:09:59.849 "state": "online", 00:09:59.849 "raid_level": "raid1", 00:09:59.849 "superblock": true, 00:09:59.849 "num_base_bdevs": 4, 00:09:59.849 "num_base_bdevs_discovered": 4, 00:09:59.849 "num_base_bdevs_operational": 4, 00:09:59.849 "base_bdevs_list": [ 00:09:59.849 { 00:09:59.849 "name": "BaseBdev1", 00:09:59.849 "uuid": "12a9b3d5-a51d-4cc9-ae35-f3f0e67f18b0", 00:09:59.849 "is_configured": true, 00:09:59.849 "data_offset": 2048, 00:09:59.849 "data_size": 63488 00:09:59.849 }, 00:09:59.849 { 00:09:59.849 "name": "BaseBdev2", 00:09:59.849 "uuid": "6399de16-6542-4bd7-8ad8-8bc2959500fc", 00:09:59.849 "is_configured": true, 00:09:59.849 "data_offset": 2048, 00:09:59.849 "data_size": 63488 00:09:59.849 }, 00:09:59.849 { 00:09:59.849 "name": "BaseBdev3", 00:09:59.849 "uuid": "d42f72aa-d487-4fa2-8a4e-861ab8ba0e60", 00:09:59.849 "is_configured": true, 00:09:59.849 "data_offset": 2048, 00:09:59.849 "data_size": 63488 00:09:59.849 }, 00:09:59.849 { 00:09:59.849 "name": "BaseBdev4", 00:09:59.849 "uuid": "2e22f21a-7f1c-47ca-af0c-8bd216d29a40", 00:09:59.849 "is_configured": true, 00:09:59.849 "data_offset": 2048, 00:09:59.849 "data_size": 63488 00:09:59.849 } 00:09:59.849 ] 00:09:59.849 } 00:09:59.849 } 00:09:59.849 }' 00:09:59.849 23:43:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:59.849 23:43:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:09:59.849 BaseBdev2 00:09:59.849 BaseBdev3 00:09:59.849 BaseBdev4' 00:09:59.849 23:43:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:59.849 23:43:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:59.849 23:43:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:59.849 23:43:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:09:59.849 23:43:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:59.849 23:43:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:59.850 23:43:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:59.850 23:43:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:59.850 23:43:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:59.850 23:43:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:59.850 23:43:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:59.850 23:43:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:09:59.850 23:43:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:59.850 23:43:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:59.850 23:43:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:59.850 23:43:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:00.109 23:43:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:00.109 23:43:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:00.109 23:43:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:00.109 23:43:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:10:00.109 23:43:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:00.109 23:43:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:00.109 23:43:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:00.109 23:43:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:00.109 23:43:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:00.109 23:43:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:00.109 23:43:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:00.109 23:43:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:00.109 23:43:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:10:00.109 23:43:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:00.109 23:43:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:00.109 23:43:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:00.109 23:43:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:00.109 23:43:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:00.109 23:43:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:10:00.109 23:43:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:00.109 23:43:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:00.109 [2024-11-26 23:43:48.105067] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:10:00.109 23:43:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:00.109 23:43:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:10:00.109 23:43:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy raid1 00:10:00.109 23:43:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:10:00.109 23:43:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@199 -- # return 0 00:10:00.109 23:43:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:10:00.109 23:43:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid1 0 3 00:10:00.109 23:43:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:00.109 23:43:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:00.109 23:43:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:00.109 23:43:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:00.109 23:43:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:00.109 23:43:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:00.109 23:43:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:00.109 23:43:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:00.109 23:43:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:00.109 23:43:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:00.109 23:43:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:00.109 23:43:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:00.109 23:43:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:00.109 23:43:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:00.109 23:43:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:00.109 "name": "Existed_Raid", 00:10:00.109 "uuid": "69ec4f5c-23b9-42c5-b1aa-4abd32961f6d", 00:10:00.109 "strip_size_kb": 0, 00:10:00.109 "state": "online", 00:10:00.109 "raid_level": "raid1", 00:10:00.109 "superblock": true, 00:10:00.109 "num_base_bdevs": 4, 00:10:00.109 "num_base_bdevs_discovered": 3, 00:10:00.109 "num_base_bdevs_operational": 3, 00:10:00.109 "base_bdevs_list": [ 00:10:00.109 { 00:10:00.109 "name": null, 00:10:00.109 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:00.109 "is_configured": false, 00:10:00.109 "data_offset": 0, 00:10:00.109 "data_size": 63488 00:10:00.109 }, 00:10:00.109 { 00:10:00.109 "name": "BaseBdev2", 00:10:00.109 "uuid": "6399de16-6542-4bd7-8ad8-8bc2959500fc", 00:10:00.109 "is_configured": true, 00:10:00.109 "data_offset": 2048, 00:10:00.109 "data_size": 63488 00:10:00.109 }, 00:10:00.109 { 00:10:00.109 "name": "BaseBdev3", 00:10:00.109 "uuid": "d42f72aa-d487-4fa2-8a4e-861ab8ba0e60", 00:10:00.109 "is_configured": true, 00:10:00.109 "data_offset": 2048, 00:10:00.109 "data_size": 63488 00:10:00.109 }, 00:10:00.109 { 00:10:00.109 "name": "BaseBdev4", 00:10:00.109 "uuid": "2e22f21a-7f1c-47ca-af0c-8bd216d29a40", 00:10:00.109 "is_configured": true, 00:10:00.109 "data_offset": 2048, 00:10:00.109 "data_size": 63488 00:10:00.109 } 00:10:00.109 ] 00:10:00.109 }' 00:10:00.109 23:43:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:00.109 23:43:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:00.678 23:43:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:10:00.678 23:43:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:00.678 23:43:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:10:00.678 23:43:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:00.678 23:43:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:00.678 23:43:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:00.678 23:43:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:00.678 23:43:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:10:00.678 23:43:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:10:00.678 23:43:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:10:00.678 23:43:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:00.678 23:43:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:00.678 [2024-11-26 23:43:48.567483] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:10:00.678 23:43:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:00.678 23:43:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:10:00.678 23:43:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:00.678 23:43:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:00.678 23:43:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:10:00.678 23:43:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:00.678 23:43:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:00.678 23:43:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:00.678 23:43:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:10:00.678 23:43:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:10:00.678 23:43:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:10:00.678 23:43:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:00.678 23:43:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:00.678 [2024-11-26 23:43:48.634852] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:10:00.678 23:43:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:00.678 23:43:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:10:00.678 23:43:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:00.678 23:43:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:00.678 23:43:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:10:00.678 23:43:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:00.678 23:43:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:00.678 23:43:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:00.678 23:43:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:10:00.678 23:43:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:10:00.678 23:43:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev4 00:10:00.678 23:43:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:00.678 23:43:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:00.678 [2024-11-26 23:43:48.701910] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev4 00:10:00.678 [2024-11-26 23:43:48.702052] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:00.678 [2024-11-26 23:43:48.713644] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:00.678 [2024-11-26 23:43:48.713691] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:00.678 [2024-11-26 23:43:48.713702] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:10:00.678 23:43:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:00.678 23:43:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:10:00.678 23:43:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:00.678 23:43:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:00.678 23:43:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:10:00.678 23:43:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:00.678 23:43:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:00.678 23:43:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:00.678 23:43:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:10:00.678 23:43:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:10:00.678 23:43:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 4 -gt 2 ']' 00:10:00.678 23:43:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:10:00.678 23:43:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:00.678 23:43:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:10:00.678 23:43:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:00.678 23:43:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:00.678 BaseBdev2 00:10:00.678 23:43:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:00.678 23:43:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:10:00.678 23:43:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:10:00.678 23:43:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:00.678 23:43:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:10:00.678 23:43:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:00.678 23:43:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:00.678 23:43:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:00.678 23:43:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:00.678 23:43:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:00.678 23:43:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:00.678 23:43:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:10:00.678 23:43:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:00.678 23:43:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:00.678 [ 00:10:00.678 { 00:10:00.678 "name": "BaseBdev2", 00:10:00.678 "aliases": [ 00:10:00.678 "9d5f8e3d-79a6-4c5f-93a0-903568a10261" 00:10:00.678 ], 00:10:00.678 "product_name": "Malloc disk", 00:10:00.678 "block_size": 512, 00:10:00.938 "num_blocks": 65536, 00:10:00.938 "uuid": "9d5f8e3d-79a6-4c5f-93a0-903568a10261", 00:10:00.938 "assigned_rate_limits": { 00:10:00.938 "rw_ios_per_sec": 0, 00:10:00.938 "rw_mbytes_per_sec": 0, 00:10:00.938 "r_mbytes_per_sec": 0, 00:10:00.938 "w_mbytes_per_sec": 0 00:10:00.938 }, 00:10:00.938 "claimed": false, 00:10:00.938 "zoned": false, 00:10:00.938 "supported_io_types": { 00:10:00.938 "read": true, 00:10:00.938 "write": true, 00:10:00.938 "unmap": true, 00:10:00.938 "flush": true, 00:10:00.938 "reset": true, 00:10:00.938 "nvme_admin": false, 00:10:00.938 "nvme_io": false, 00:10:00.938 "nvme_io_md": false, 00:10:00.938 "write_zeroes": true, 00:10:00.938 "zcopy": true, 00:10:00.938 "get_zone_info": false, 00:10:00.938 "zone_management": false, 00:10:00.938 "zone_append": false, 00:10:00.938 "compare": false, 00:10:00.938 "compare_and_write": false, 00:10:00.938 "abort": true, 00:10:00.938 "seek_hole": false, 00:10:00.938 "seek_data": false, 00:10:00.938 "copy": true, 00:10:00.938 "nvme_iov_md": false 00:10:00.938 }, 00:10:00.938 "memory_domains": [ 00:10:00.938 { 00:10:00.938 "dma_device_id": "system", 00:10:00.938 "dma_device_type": 1 00:10:00.938 }, 00:10:00.938 { 00:10:00.938 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:00.938 "dma_device_type": 2 00:10:00.938 } 00:10:00.938 ], 00:10:00.938 "driver_specific": {} 00:10:00.938 } 00:10:00.938 ] 00:10:00.938 23:43:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:00.938 23:43:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:10:00.938 23:43:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:10:00.938 23:43:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:00.938 23:43:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:10:00.938 23:43:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:00.938 23:43:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:00.938 BaseBdev3 00:10:00.938 23:43:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:00.938 23:43:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:10:00.938 23:43:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:10:00.938 23:43:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:00.938 23:43:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:10:00.938 23:43:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:00.938 23:43:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:00.938 23:43:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:00.938 23:43:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:00.938 23:43:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:00.938 23:43:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:00.938 23:43:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:10:00.938 23:43:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:00.938 23:43:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:00.938 [ 00:10:00.938 { 00:10:00.938 "name": "BaseBdev3", 00:10:00.938 "aliases": [ 00:10:00.938 "0d43e9de-d1fb-4d3a-b984-bfe07f021194" 00:10:00.938 ], 00:10:00.938 "product_name": "Malloc disk", 00:10:00.938 "block_size": 512, 00:10:00.938 "num_blocks": 65536, 00:10:00.938 "uuid": "0d43e9de-d1fb-4d3a-b984-bfe07f021194", 00:10:00.938 "assigned_rate_limits": { 00:10:00.938 "rw_ios_per_sec": 0, 00:10:00.938 "rw_mbytes_per_sec": 0, 00:10:00.938 "r_mbytes_per_sec": 0, 00:10:00.938 "w_mbytes_per_sec": 0 00:10:00.938 }, 00:10:00.938 "claimed": false, 00:10:00.938 "zoned": false, 00:10:00.938 "supported_io_types": { 00:10:00.938 "read": true, 00:10:00.938 "write": true, 00:10:00.938 "unmap": true, 00:10:00.938 "flush": true, 00:10:00.938 "reset": true, 00:10:00.938 "nvme_admin": false, 00:10:00.938 "nvme_io": false, 00:10:00.938 "nvme_io_md": false, 00:10:00.938 "write_zeroes": true, 00:10:00.938 "zcopy": true, 00:10:00.938 "get_zone_info": false, 00:10:00.938 "zone_management": false, 00:10:00.938 "zone_append": false, 00:10:00.938 "compare": false, 00:10:00.938 "compare_and_write": false, 00:10:00.938 "abort": true, 00:10:00.938 "seek_hole": false, 00:10:00.938 "seek_data": false, 00:10:00.938 "copy": true, 00:10:00.938 "nvme_iov_md": false 00:10:00.938 }, 00:10:00.938 "memory_domains": [ 00:10:00.938 { 00:10:00.938 "dma_device_id": "system", 00:10:00.938 "dma_device_type": 1 00:10:00.938 }, 00:10:00.938 { 00:10:00.938 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:00.938 "dma_device_type": 2 00:10:00.938 } 00:10:00.938 ], 00:10:00.938 "driver_specific": {} 00:10:00.938 } 00:10:00.938 ] 00:10:00.938 23:43:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:00.938 23:43:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:10:00.938 23:43:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:10:00.938 23:43:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:00.938 23:43:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:10:00.938 23:43:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:00.938 23:43:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:00.938 BaseBdev4 00:10:00.938 23:43:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:00.938 23:43:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev4 00:10:00.938 23:43:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:10:00.938 23:43:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:00.938 23:43:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:10:00.939 23:43:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:00.939 23:43:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:00.939 23:43:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:00.939 23:43:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:00.939 23:43:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:00.939 23:43:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:00.939 23:43:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:10:00.939 23:43:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:00.939 23:43:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:00.939 [ 00:10:00.939 { 00:10:00.939 "name": "BaseBdev4", 00:10:00.939 "aliases": [ 00:10:00.939 "0f999cee-895d-4f31-84e7-a808f30c4f0a" 00:10:00.939 ], 00:10:00.939 "product_name": "Malloc disk", 00:10:00.939 "block_size": 512, 00:10:00.939 "num_blocks": 65536, 00:10:00.939 "uuid": "0f999cee-895d-4f31-84e7-a808f30c4f0a", 00:10:00.939 "assigned_rate_limits": { 00:10:00.939 "rw_ios_per_sec": 0, 00:10:00.939 "rw_mbytes_per_sec": 0, 00:10:00.939 "r_mbytes_per_sec": 0, 00:10:00.939 "w_mbytes_per_sec": 0 00:10:00.939 }, 00:10:00.939 "claimed": false, 00:10:00.939 "zoned": false, 00:10:00.939 "supported_io_types": { 00:10:00.939 "read": true, 00:10:00.939 "write": true, 00:10:00.939 "unmap": true, 00:10:00.939 "flush": true, 00:10:00.939 "reset": true, 00:10:00.939 "nvme_admin": false, 00:10:00.939 "nvme_io": false, 00:10:00.939 "nvme_io_md": false, 00:10:00.939 "write_zeroes": true, 00:10:00.939 "zcopy": true, 00:10:00.939 "get_zone_info": false, 00:10:00.939 "zone_management": false, 00:10:00.939 "zone_append": false, 00:10:00.939 "compare": false, 00:10:00.939 "compare_and_write": false, 00:10:00.939 "abort": true, 00:10:00.939 "seek_hole": false, 00:10:00.939 "seek_data": false, 00:10:00.939 "copy": true, 00:10:00.939 "nvme_iov_md": false 00:10:00.939 }, 00:10:00.939 "memory_domains": [ 00:10:00.939 { 00:10:00.939 "dma_device_id": "system", 00:10:00.939 "dma_device_type": 1 00:10:00.939 }, 00:10:00.939 { 00:10:00.939 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:00.939 "dma_device_type": 2 00:10:00.939 } 00:10:00.939 ], 00:10:00.939 "driver_specific": {} 00:10:00.939 } 00:10:00.939 ] 00:10:00.939 23:43:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:00.939 23:43:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:10:00.939 23:43:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:10:00.939 23:43:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:00.939 23:43:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:10:00.939 23:43:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:00.939 23:43:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:00.939 [2024-11-26 23:43:48.929301] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:10:00.939 [2024-11-26 23:43:48.929407] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:10:00.939 [2024-11-26 23:43:48.929446] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:00.939 [2024-11-26 23:43:48.931221] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:00.939 [2024-11-26 23:43:48.931316] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:10:00.939 23:43:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:00.939 23:43:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:00.939 23:43:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:00.939 23:43:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:00.939 23:43:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:00.939 23:43:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:00.939 23:43:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:00.939 23:43:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:00.939 23:43:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:00.939 23:43:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:00.939 23:43:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:00.939 23:43:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:00.939 23:43:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:00.939 23:43:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:00.939 23:43:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:00.939 23:43:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:00.939 23:43:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:00.939 "name": "Existed_Raid", 00:10:00.939 "uuid": "e969c527-39df-4a0d-aac6-f89a15acab42", 00:10:00.939 "strip_size_kb": 0, 00:10:00.939 "state": "configuring", 00:10:00.939 "raid_level": "raid1", 00:10:00.939 "superblock": true, 00:10:00.939 "num_base_bdevs": 4, 00:10:00.939 "num_base_bdevs_discovered": 3, 00:10:00.939 "num_base_bdevs_operational": 4, 00:10:00.939 "base_bdevs_list": [ 00:10:00.939 { 00:10:00.939 "name": "BaseBdev1", 00:10:00.939 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:00.939 "is_configured": false, 00:10:00.939 "data_offset": 0, 00:10:00.939 "data_size": 0 00:10:00.939 }, 00:10:00.939 { 00:10:00.939 "name": "BaseBdev2", 00:10:00.939 "uuid": "9d5f8e3d-79a6-4c5f-93a0-903568a10261", 00:10:00.939 "is_configured": true, 00:10:00.939 "data_offset": 2048, 00:10:00.939 "data_size": 63488 00:10:00.939 }, 00:10:00.939 { 00:10:00.939 "name": "BaseBdev3", 00:10:00.939 "uuid": "0d43e9de-d1fb-4d3a-b984-bfe07f021194", 00:10:00.939 "is_configured": true, 00:10:00.939 "data_offset": 2048, 00:10:00.939 "data_size": 63488 00:10:00.939 }, 00:10:00.939 { 00:10:00.939 "name": "BaseBdev4", 00:10:00.939 "uuid": "0f999cee-895d-4f31-84e7-a808f30c4f0a", 00:10:00.939 "is_configured": true, 00:10:00.939 "data_offset": 2048, 00:10:00.939 "data_size": 63488 00:10:00.939 } 00:10:00.939 ] 00:10:00.939 }' 00:10:00.939 23:43:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:00.939 23:43:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:01.513 23:43:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:10:01.513 23:43:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:01.513 23:43:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:01.513 [2024-11-26 23:43:49.396555] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:10:01.513 23:43:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:01.513 23:43:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:01.513 23:43:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:01.513 23:43:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:01.513 23:43:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:01.513 23:43:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:01.513 23:43:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:01.513 23:43:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:01.513 23:43:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:01.513 23:43:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:01.513 23:43:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:01.513 23:43:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:01.513 23:43:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:01.513 23:43:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:01.513 23:43:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:01.513 23:43:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:01.513 23:43:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:01.513 "name": "Existed_Raid", 00:10:01.513 "uuid": "e969c527-39df-4a0d-aac6-f89a15acab42", 00:10:01.513 "strip_size_kb": 0, 00:10:01.513 "state": "configuring", 00:10:01.513 "raid_level": "raid1", 00:10:01.513 "superblock": true, 00:10:01.513 "num_base_bdevs": 4, 00:10:01.513 "num_base_bdevs_discovered": 2, 00:10:01.513 "num_base_bdevs_operational": 4, 00:10:01.513 "base_bdevs_list": [ 00:10:01.514 { 00:10:01.514 "name": "BaseBdev1", 00:10:01.514 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:01.514 "is_configured": false, 00:10:01.514 "data_offset": 0, 00:10:01.514 "data_size": 0 00:10:01.514 }, 00:10:01.514 { 00:10:01.514 "name": null, 00:10:01.514 "uuid": "9d5f8e3d-79a6-4c5f-93a0-903568a10261", 00:10:01.514 "is_configured": false, 00:10:01.514 "data_offset": 0, 00:10:01.514 "data_size": 63488 00:10:01.514 }, 00:10:01.514 { 00:10:01.514 "name": "BaseBdev3", 00:10:01.514 "uuid": "0d43e9de-d1fb-4d3a-b984-bfe07f021194", 00:10:01.514 "is_configured": true, 00:10:01.514 "data_offset": 2048, 00:10:01.514 "data_size": 63488 00:10:01.514 }, 00:10:01.514 { 00:10:01.514 "name": "BaseBdev4", 00:10:01.514 "uuid": "0f999cee-895d-4f31-84e7-a808f30c4f0a", 00:10:01.514 "is_configured": true, 00:10:01.514 "data_offset": 2048, 00:10:01.514 "data_size": 63488 00:10:01.514 } 00:10:01.514 ] 00:10:01.514 }' 00:10:01.514 23:43:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:01.514 23:43:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:01.773 23:43:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:10:01.773 23:43:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:01.773 23:43:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:01.773 23:43:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:01.773 23:43:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:01.773 23:43:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:10:01.773 23:43:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:10:01.773 23:43:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:01.773 23:43:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:01.773 [2024-11-26 23:43:49.878571] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:01.773 BaseBdev1 00:10:01.773 23:43:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:01.773 23:43:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:10:01.773 23:43:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:10:01.773 23:43:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:01.773 23:43:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:10:01.773 23:43:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:01.773 23:43:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:01.773 23:43:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:01.773 23:43:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:01.773 23:43:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:01.773 23:43:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:01.773 23:43:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:10:01.773 23:43:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:01.773 23:43:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:02.044 [ 00:10:02.044 { 00:10:02.044 "name": "BaseBdev1", 00:10:02.044 "aliases": [ 00:10:02.044 "16525f75-33e9-4e8a-9784-73f81b5875ae" 00:10:02.044 ], 00:10:02.044 "product_name": "Malloc disk", 00:10:02.044 "block_size": 512, 00:10:02.044 "num_blocks": 65536, 00:10:02.044 "uuid": "16525f75-33e9-4e8a-9784-73f81b5875ae", 00:10:02.044 "assigned_rate_limits": { 00:10:02.044 "rw_ios_per_sec": 0, 00:10:02.044 "rw_mbytes_per_sec": 0, 00:10:02.044 "r_mbytes_per_sec": 0, 00:10:02.044 "w_mbytes_per_sec": 0 00:10:02.044 }, 00:10:02.044 "claimed": true, 00:10:02.044 "claim_type": "exclusive_write", 00:10:02.044 "zoned": false, 00:10:02.044 "supported_io_types": { 00:10:02.044 "read": true, 00:10:02.044 "write": true, 00:10:02.044 "unmap": true, 00:10:02.044 "flush": true, 00:10:02.044 "reset": true, 00:10:02.044 "nvme_admin": false, 00:10:02.044 "nvme_io": false, 00:10:02.044 "nvme_io_md": false, 00:10:02.044 "write_zeroes": true, 00:10:02.044 "zcopy": true, 00:10:02.044 "get_zone_info": false, 00:10:02.044 "zone_management": false, 00:10:02.044 "zone_append": false, 00:10:02.044 "compare": false, 00:10:02.044 "compare_and_write": false, 00:10:02.044 "abort": true, 00:10:02.044 "seek_hole": false, 00:10:02.044 "seek_data": false, 00:10:02.044 "copy": true, 00:10:02.044 "nvme_iov_md": false 00:10:02.044 }, 00:10:02.044 "memory_domains": [ 00:10:02.044 { 00:10:02.044 "dma_device_id": "system", 00:10:02.044 "dma_device_type": 1 00:10:02.044 }, 00:10:02.044 { 00:10:02.044 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:02.044 "dma_device_type": 2 00:10:02.044 } 00:10:02.044 ], 00:10:02.044 "driver_specific": {} 00:10:02.044 } 00:10:02.044 ] 00:10:02.044 23:43:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:02.044 23:43:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:10:02.044 23:43:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:02.044 23:43:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:02.044 23:43:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:02.044 23:43:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:02.044 23:43:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:02.044 23:43:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:02.044 23:43:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:02.044 23:43:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:02.044 23:43:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:02.044 23:43:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:02.044 23:43:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:02.044 23:43:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:02.044 23:43:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:02.044 23:43:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:02.044 23:43:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:02.044 23:43:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:02.044 "name": "Existed_Raid", 00:10:02.044 "uuid": "e969c527-39df-4a0d-aac6-f89a15acab42", 00:10:02.044 "strip_size_kb": 0, 00:10:02.044 "state": "configuring", 00:10:02.044 "raid_level": "raid1", 00:10:02.044 "superblock": true, 00:10:02.044 "num_base_bdevs": 4, 00:10:02.044 "num_base_bdevs_discovered": 3, 00:10:02.044 "num_base_bdevs_operational": 4, 00:10:02.044 "base_bdevs_list": [ 00:10:02.044 { 00:10:02.044 "name": "BaseBdev1", 00:10:02.044 "uuid": "16525f75-33e9-4e8a-9784-73f81b5875ae", 00:10:02.044 "is_configured": true, 00:10:02.044 "data_offset": 2048, 00:10:02.044 "data_size": 63488 00:10:02.044 }, 00:10:02.044 { 00:10:02.044 "name": null, 00:10:02.044 "uuid": "9d5f8e3d-79a6-4c5f-93a0-903568a10261", 00:10:02.044 "is_configured": false, 00:10:02.044 "data_offset": 0, 00:10:02.044 "data_size": 63488 00:10:02.044 }, 00:10:02.044 { 00:10:02.044 "name": "BaseBdev3", 00:10:02.044 "uuid": "0d43e9de-d1fb-4d3a-b984-bfe07f021194", 00:10:02.044 "is_configured": true, 00:10:02.044 "data_offset": 2048, 00:10:02.044 "data_size": 63488 00:10:02.044 }, 00:10:02.044 { 00:10:02.044 "name": "BaseBdev4", 00:10:02.044 "uuid": "0f999cee-895d-4f31-84e7-a808f30c4f0a", 00:10:02.044 "is_configured": true, 00:10:02.044 "data_offset": 2048, 00:10:02.044 "data_size": 63488 00:10:02.044 } 00:10:02.044 ] 00:10:02.044 }' 00:10:02.044 23:43:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:02.044 23:43:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:02.302 23:43:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:02.302 23:43:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:02.302 23:43:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:02.302 23:43:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:10:02.302 23:43:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:02.302 23:43:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:10:02.302 23:43:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:10:02.302 23:43:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:02.302 23:43:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:02.302 [2024-11-26 23:43:50.397798] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:10:02.302 23:43:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:02.302 23:43:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:02.302 23:43:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:02.302 23:43:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:02.302 23:43:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:02.302 23:43:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:02.302 23:43:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:02.302 23:43:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:02.302 23:43:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:02.302 23:43:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:02.302 23:43:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:02.302 23:43:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:02.302 23:43:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:02.302 23:43:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:02.302 23:43:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:02.302 23:43:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:02.561 23:43:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:02.561 "name": "Existed_Raid", 00:10:02.561 "uuid": "e969c527-39df-4a0d-aac6-f89a15acab42", 00:10:02.561 "strip_size_kb": 0, 00:10:02.561 "state": "configuring", 00:10:02.561 "raid_level": "raid1", 00:10:02.561 "superblock": true, 00:10:02.561 "num_base_bdevs": 4, 00:10:02.561 "num_base_bdevs_discovered": 2, 00:10:02.561 "num_base_bdevs_operational": 4, 00:10:02.561 "base_bdevs_list": [ 00:10:02.561 { 00:10:02.561 "name": "BaseBdev1", 00:10:02.561 "uuid": "16525f75-33e9-4e8a-9784-73f81b5875ae", 00:10:02.561 "is_configured": true, 00:10:02.561 "data_offset": 2048, 00:10:02.561 "data_size": 63488 00:10:02.561 }, 00:10:02.561 { 00:10:02.561 "name": null, 00:10:02.561 "uuid": "9d5f8e3d-79a6-4c5f-93a0-903568a10261", 00:10:02.561 "is_configured": false, 00:10:02.561 "data_offset": 0, 00:10:02.561 "data_size": 63488 00:10:02.561 }, 00:10:02.561 { 00:10:02.561 "name": null, 00:10:02.561 "uuid": "0d43e9de-d1fb-4d3a-b984-bfe07f021194", 00:10:02.561 "is_configured": false, 00:10:02.561 "data_offset": 0, 00:10:02.561 "data_size": 63488 00:10:02.561 }, 00:10:02.561 { 00:10:02.561 "name": "BaseBdev4", 00:10:02.561 "uuid": "0f999cee-895d-4f31-84e7-a808f30c4f0a", 00:10:02.561 "is_configured": true, 00:10:02.561 "data_offset": 2048, 00:10:02.561 "data_size": 63488 00:10:02.561 } 00:10:02.561 ] 00:10:02.561 }' 00:10:02.561 23:43:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:02.561 23:43:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:02.820 23:43:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:02.820 23:43:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:02.820 23:43:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:02.820 23:43:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:10:02.820 23:43:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:02.820 23:43:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:10:02.820 23:43:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:10:02.820 23:43:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:02.820 23:43:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:02.820 [2024-11-26 23:43:50.833026] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:02.820 23:43:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:02.820 23:43:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:02.820 23:43:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:02.820 23:43:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:02.820 23:43:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:02.820 23:43:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:02.820 23:43:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:02.820 23:43:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:02.820 23:43:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:02.820 23:43:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:02.820 23:43:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:02.820 23:43:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:02.820 23:43:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:02.820 23:43:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:02.820 23:43:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:02.820 23:43:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:02.821 23:43:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:02.821 "name": "Existed_Raid", 00:10:02.821 "uuid": "e969c527-39df-4a0d-aac6-f89a15acab42", 00:10:02.821 "strip_size_kb": 0, 00:10:02.821 "state": "configuring", 00:10:02.821 "raid_level": "raid1", 00:10:02.821 "superblock": true, 00:10:02.821 "num_base_bdevs": 4, 00:10:02.821 "num_base_bdevs_discovered": 3, 00:10:02.821 "num_base_bdevs_operational": 4, 00:10:02.821 "base_bdevs_list": [ 00:10:02.821 { 00:10:02.821 "name": "BaseBdev1", 00:10:02.821 "uuid": "16525f75-33e9-4e8a-9784-73f81b5875ae", 00:10:02.821 "is_configured": true, 00:10:02.821 "data_offset": 2048, 00:10:02.821 "data_size": 63488 00:10:02.821 }, 00:10:02.821 { 00:10:02.821 "name": null, 00:10:02.821 "uuid": "9d5f8e3d-79a6-4c5f-93a0-903568a10261", 00:10:02.821 "is_configured": false, 00:10:02.821 "data_offset": 0, 00:10:02.821 "data_size": 63488 00:10:02.821 }, 00:10:02.821 { 00:10:02.821 "name": "BaseBdev3", 00:10:02.821 "uuid": "0d43e9de-d1fb-4d3a-b984-bfe07f021194", 00:10:02.821 "is_configured": true, 00:10:02.821 "data_offset": 2048, 00:10:02.821 "data_size": 63488 00:10:02.821 }, 00:10:02.821 { 00:10:02.821 "name": "BaseBdev4", 00:10:02.821 "uuid": "0f999cee-895d-4f31-84e7-a808f30c4f0a", 00:10:02.821 "is_configured": true, 00:10:02.821 "data_offset": 2048, 00:10:02.821 "data_size": 63488 00:10:02.821 } 00:10:02.821 ] 00:10:02.821 }' 00:10:02.821 23:43:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:02.821 23:43:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:03.389 23:43:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:03.389 23:43:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:03.389 23:43:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:03.389 23:43:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:10:03.389 23:43:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:03.389 23:43:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:10:03.389 23:43:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:10:03.389 23:43:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:03.389 23:43:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:03.389 [2024-11-26 23:43:51.316239] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:10:03.389 23:43:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:03.389 23:43:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:03.389 23:43:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:03.389 23:43:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:03.389 23:43:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:03.389 23:43:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:03.389 23:43:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:03.389 23:43:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:03.389 23:43:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:03.389 23:43:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:03.389 23:43:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:03.389 23:43:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:03.389 23:43:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:03.389 23:43:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:03.389 23:43:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:03.389 23:43:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:03.389 23:43:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:03.389 "name": "Existed_Raid", 00:10:03.389 "uuid": "e969c527-39df-4a0d-aac6-f89a15acab42", 00:10:03.389 "strip_size_kb": 0, 00:10:03.389 "state": "configuring", 00:10:03.389 "raid_level": "raid1", 00:10:03.389 "superblock": true, 00:10:03.389 "num_base_bdevs": 4, 00:10:03.389 "num_base_bdevs_discovered": 2, 00:10:03.389 "num_base_bdevs_operational": 4, 00:10:03.389 "base_bdevs_list": [ 00:10:03.389 { 00:10:03.389 "name": null, 00:10:03.389 "uuid": "16525f75-33e9-4e8a-9784-73f81b5875ae", 00:10:03.389 "is_configured": false, 00:10:03.389 "data_offset": 0, 00:10:03.389 "data_size": 63488 00:10:03.389 }, 00:10:03.389 { 00:10:03.389 "name": null, 00:10:03.389 "uuid": "9d5f8e3d-79a6-4c5f-93a0-903568a10261", 00:10:03.389 "is_configured": false, 00:10:03.389 "data_offset": 0, 00:10:03.389 "data_size": 63488 00:10:03.389 }, 00:10:03.389 { 00:10:03.389 "name": "BaseBdev3", 00:10:03.389 "uuid": "0d43e9de-d1fb-4d3a-b984-bfe07f021194", 00:10:03.389 "is_configured": true, 00:10:03.389 "data_offset": 2048, 00:10:03.389 "data_size": 63488 00:10:03.389 }, 00:10:03.389 { 00:10:03.389 "name": "BaseBdev4", 00:10:03.389 "uuid": "0f999cee-895d-4f31-84e7-a808f30c4f0a", 00:10:03.389 "is_configured": true, 00:10:03.389 "data_offset": 2048, 00:10:03.390 "data_size": 63488 00:10:03.390 } 00:10:03.390 ] 00:10:03.390 }' 00:10:03.390 23:43:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:03.390 23:43:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:03.646 23:43:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:03.646 23:43:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:03.646 23:43:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:10:03.646 23:43:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:03.904 23:43:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:03.904 23:43:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:10:03.904 23:43:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:10:03.904 23:43:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:03.904 23:43:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:03.904 [2024-11-26 23:43:51.813988] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:03.905 23:43:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:03.905 23:43:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:03.905 23:43:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:03.905 23:43:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:03.905 23:43:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:03.905 23:43:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:03.905 23:43:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:03.905 23:43:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:03.905 23:43:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:03.905 23:43:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:03.905 23:43:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:03.905 23:43:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:03.905 23:43:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:03.905 23:43:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:03.905 23:43:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:03.905 23:43:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:03.905 23:43:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:03.905 "name": "Existed_Raid", 00:10:03.905 "uuid": "e969c527-39df-4a0d-aac6-f89a15acab42", 00:10:03.905 "strip_size_kb": 0, 00:10:03.905 "state": "configuring", 00:10:03.905 "raid_level": "raid1", 00:10:03.905 "superblock": true, 00:10:03.905 "num_base_bdevs": 4, 00:10:03.905 "num_base_bdevs_discovered": 3, 00:10:03.905 "num_base_bdevs_operational": 4, 00:10:03.905 "base_bdevs_list": [ 00:10:03.905 { 00:10:03.905 "name": null, 00:10:03.905 "uuid": "16525f75-33e9-4e8a-9784-73f81b5875ae", 00:10:03.905 "is_configured": false, 00:10:03.905 "data_offset": 0, 00:10:03.905 "data_size": 63488 00:10:03.905 }, 00:10:03.905 { 00:10:03.905 "name": "BaseBdev2", 00:10:03.905 "uuid": "9d5f8e3d-79a6-4c5f-93a0-903568a10261", 00:10:03.905 "is_configured": true, 00:10:03.905 "data_offset": 2048, 00:10:03.905 "data_size": 63488 00:10:03.905 }, 00:10:03.905 { 00:10:03.905 "name": "BaseBdev3", 00:10:03.905 "uuid": "0d43e9de-d1fb-4d3a-b984-bfe07f021194", 00:10:03.905 "is_configured": true, 00:10:03.905 "data_offset": 2048, 00:10:03.905 "data_size": 63488 00:10:03.905 }, 00:10:03.905 { 00:10:03.905 "name": "BaseBdev4", 00:10:03.905 "uuid": "0f999cee-895d-4f31-84e7-a808f30c4f0a", 00:10:03.905 "is_configured": true, 00:10:03.905 "data_offset": 2048, 00:10:03.905 "data_size": 63488 00:10:03.905 } 00:10:03.905 ] 00:10:03.905 }' 00:10:03.905 23:43:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:03.905 23:43:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:04.164 23:43:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:04.164 23:43:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:04.164 23:43:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:10:04.164 23:43:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:04.164 23:43:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:04.423 23:43:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:10:04.423 23:43:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:04.423 23:43:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:04.423 23:43:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:04.423 23:43:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:10:04.423 23:43:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:04.424 23:43:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 16525f75-33e9-4e8a-9784-73f81b5875ae 00:10:04.424 23:43:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:04.424 23:43:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:04.424 [2024-11-26 23:43:52.356104] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:10:04.424 [2024-11-26 23:43:52.356322] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:10:04.424 [2024-11-26 23:43:52.356340] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:10:04.424 NewBaseBdev 00:10:04.424 [2024-11-26 23:43:52.356633] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002a10 00:10:04.424 [2024-11-26 23:43:52.356800] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:10:04.424 [2024-11-26 23:43:52.356816] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001c80 00:10:04.424 [2024-11-26 23:43:52.356926] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:04.424 23:43:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:04.424 23:43:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:10:04.424 23:43:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:10:04.424 23:43:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:04.424 23:43:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:10:04.424 23:43:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:04.424 23:43:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:04.424 23:43:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:04.424 23:43:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:04.424 23:43:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:04.424 23:43:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:04.424 23:43:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:10:04.424 23:43:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:04.424 23:43:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:04.424 [ 00:10:04.424 { 00:10:04.424 "name": "NewBaseBdev", 00:10:04.424 "aliases": [ 00:10:04.424 "16525f75-33e9-4e8a-9784-73f81b5875ae" 00:10:04.424 ], 00:10:04.424 "product_name": "Malloc disk", 00:10:04.424 "block_size": 512, 00:10:04.424 "num_blocks": 65536, 00:10:04.424 "uuid": "16525f75-33e9-4e8a-9784-73f81b5875ae", 00:10:04.424 "assigned_rate_limits": { 00:10:04.424 "rw_ios_per_sec": 0, 00:10:04.424 "rw_mbytes_per_sec": 0, 00:10:04.424 "r_mbytes_per_sec": 0, 00:10:04.424 "w_mbytes_per_sec": 0 00:10:04.424 }, 00:10:04.424 "claimed": true, 00:10:04.424 "claim_type": "exclusive_write", 00:10:04.424 "zoned": false, 00:10:04.424 "supported_io_types": { 00:10:04.424 "read": true, 00:10:04.424 "write": true, 00:10:04.424 "unmap": true, 00:10:04.424 "flush": true, 00:10:04.424 "reset": true, 00:10:04.424 "nvme_admin": false, 00:10:04.424 "nvme_io": false, 00:10:04.424 "nvme_io_md": false, 00:10:04.424 "write_zeroes": true, 00:10:04.424 "zcopy": true, 00:10:04.424 "get_zone_info": false, 00:10:04.424 "zone_management": false, 00:10:04.424 "zone_append": false, 00:10:04.424 "compare": false, 00:10:04.424 "compare_and_write": false, 00:10:04.424 "abort": true, 00:10:04.424 "seek_hole": false, 00:10:04.424 "seek_data": false, 00:10:04.424 "copy": true, 00:10:04.424 "nvme_iov_md": false 00:10:04.424 }, 00:10:04.424 "memory_domains": [ 00:10:04.424 { 00:10:04.424 "dma_device_id": "system", 00:10:04.424 "dma_device_type": 1 00:10:04.424 }, 00:10:04.424 { 00:10:04.424 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:04.424 "dma_device_type": 2 00:10:04.424 } 00:10:04.424 ], 00:10:04.424 "driver_specific": {} 00:10:04.424 } 00:10:04.424 ] 00:10:04.424 23:43:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:04.424 23:43:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:10:04.424 23:43:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid1 0 4 00:10:04.424 23:43:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:04.424 23:43:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:04.424 23:43:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:04.424 23:43:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:04.424 23:43:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:04.424 23:43:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:04.424 23:43:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:04.424 23:43:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:04.424 23:43:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:04.424 23:43:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:04.424 23:43:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:04.424 23:43:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:04.424 23:43:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:04.424 23:43:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:04.424 23:43:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:04.424 "name": "Existed_Raid", 00:10:04.424 "uuid": "e969c527-39df-4a0d-aac6-f89a15acab42", 00:10:04.424 "strip_size_kb": 0, 00:10:04.424 "state": "online", 00:10:04.424 "raid_level": "raid1", 00:10:04.424 "superblock": true, 00:10:04.424 "num_base_bdevs": 4, 00:10:04.424 "num_base_bdevs_discovered": 4, 00:10:04.424 "num_base_bdevs_operational": 4, 00:10:04.424 "base_bdevs_list": [ 00:10:04.424 { 00:10:04.424 "name": "NewBaseBdev", 00:10:04.424 "uuid": "16525f75-33e9-4e8a-9784-73f81b5875ae", 00:10:04.424 "is_configured": true, 00:10:04.424 "data_offset": 2048, 00:10:04.424 "data_size": 63488 00:10:04.424 }, 00:10:04.424 { 00:10:04.424 "name": "BaseBdev2", 00:10:04.424 "uuid": "9d5f8e3d-79a6-4c5f-93a0-903568a10261", 00:10:04.424 "is_configured": true, 00:10:04.424 "data_offset": 2048, 00:10:04.424 "data_size": 63488 00:10:04.424 }, 00:10:04.424 { 00:10:04.424 "name": "BaseBdev3", 00:10:04.424 "uuid": "0d43e9de-d1fb-4d3a-b984-bfe07f021194", 00:10:04.424 "is_configured": true, 00:10:04.424 "data_offset": 2048, 00:10:04.424 "data_size": 63488 00:10:04.424 }, 00:10:04.424 { 00:10:04.424 "name": "BaseBdev4", 00:10:04.424 "uuid": "0f999cee-895d-4f31-84e7-a808f30c4f0a", 00:10:04.424 "is_configured": true, 00:10:04.424 "data_offset": 2048, 00:10:04.424 "data_size": 63488 00:10:04.424 } 00:10:04.424 ] 00:10:04.424 }' 00:10:04.424 23:43:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:04.424 23:43:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:04.992 23:43:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:10:04.992 23:43:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:10:04.992 23:43:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:10:04.992 23:43:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:10:04.992 23:43:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:10:04.992 23:43:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:10:04.992 23:43:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:10:04.992 23:43:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:04.992 23:43:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:10:04.992 23:43:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:04.992 [2024-11-26 23:43:52.863682] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:04.992 23:43:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:04.992 23:43:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:10:04.992 "name": "Existed_Raid", 00:10:04.992 "aliases": [ 00:10:04.992 "e969c527-39df-4a0d-aac6-f89a15acab42" 00:10:04.992 ], 00:10:04.992 "product_name": "Raid Volume", 00:10:04.992 "block_size": 512, 00:10:04.992 "num_blocks": 63488, 00:10:04.992 "uuid": "e969c527-39df-4a0d-aac6-f89a15acab42", 00:10:04.992 "assigned_rate_limits": { 00:10:04.992 "rw_ios_per_sec": 0, 00:10:04.992 "rw_mbytes_per_sec": 0, 00:10:04.992 "r_mbytes_per_sec": 0, 00:10:04.992 "w_mbytes_per_sec": 0 00:10:04.992 }, 00:10:04.992 "claimed": false, 00:10:04.992 "zoned": false, 00:10:04.992 "supported_io_types": { 00:10:04.992 "read": true, 00:10:04.992 "write": true, 00:10:04.992 "unmap": false, 00:10:04.992 "flush": false, 00:10:04.992 "reset": true, 00:10:04.992 "nvme_admin": false, 00:10:04.992 "nvme_io": false, 00:10:04.992 "nvme_io_md": false, 00:10:04.992 "write_zeroes": true, 00:10:04.992 "zcopy": false, 00:10:04.992 "get_zone_info": false, 00:10:04.992 "zone_management": false, 00:10:04.992 "zone_append": false, 00:10:04.992 "compare": false, 00:10:04.992 "compare_and_write": false, 00:10:04.992 "abort": false, 00:10:04.992 "seek_hole": false, 00:10:04.992 "seek_data": false, 00:10:04.992 "copy": false, 00:10:04.992 "nvme_iov_md": false 00:10:04.992 }, 00:10:04.992 "memory_domains": [ 00:10:04.992 { 00:10:04.992 "dma_device_id": "system", 00:10:04.992 "dma_device_type": 1 00:10:04.992 }, 00:10:04.992 { 00:10:04.992 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:04.992 "dma_device_type": 2 00:10:04.992 }, 00:10:04.992 { 00:10:04.992 "dma_device_id": "system", 00:10:04.992 "dma_device_type": 1 00:10:04.992 }, 00:10:04.992 { 00:10:04.992 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:04.992 "dma_device_type": 2 00:10:04.992 }, 00:10:04.992 { 00:10:04.992 "dma_device_id": "system", 00:10:04.992 "dma_device_type": 1 00:10:04.992 }, 00:10:04.992 { 00:10:04.992 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:04.992 "dma_device_type": 2 00:10:04.992 }, 00:10:04.992 { 00:10:04.992 "dma_device_id": "system", 00:10:04.992 "dma_device_type": 1 00:10:04.992 }, 00:10:04.992 { 00:10:04.992 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:04.992 "dma_device_type": 2 00:10:04.992 } 00:10:04.992 ], 00:10:04.992 "driver_specific": { 00:10:04.992 "raid": { 00:10:04.992 "uuid": "e969c527-39df-4a0d-aac6-f89a15acab42", 00:10:04.992 "strip_size_kb": 0, 00:10:04.992 "state": "online", 00:10:04.992 "raid_level": "raid1", 00:10:04.992 "superblock": true, 00:10:04.992 "num_base_bdevs": 4, 00:10:04.992 "num_base_bdevs_discovered": 4, 00:10:04.992 "num_base_bdevs_operational": 4, 00:10:04.992 "base_bdevs_list": [ 00:10:04.992 { 00:10:04.992 "name": "NewBaseBdev", 00:10:04.992 "uuid": "16525f75-33e9-4e8a-9784-73f81b5875ae", 00:10:04.992 "is_configured": true, 00:10:04.992 "data_offset": 2048, 00:10:04.992 "data_size": 63488 00:10:04.992 }, 00:10:04.992 { 00:10:04.992 "name": "BaseBdev2", 00:10:04.992 "uuid": "9d5f8e3d-79a6-4c5f-93a0-903568a10261", 00:10:04.992 "is_configured": true, 00:10:04.992 "data_offset": 2048, 00:10:04.992 "data_size": 63488 00:10:04.992 }, 00:10:04.992 { 00:10:04.992 "name": "BaseBdev3", 00:10:04.992 "uuid": "0d43e9de-d1fb-4d3a-b984-bfe07f021194", 00:10:04.992 "is_configured": true, 00:10:04.992 "data_offset": 2048, 00:10:04.992 "data_size": 63488 00:10:04.992 }, 00:10:04.992 { 00:10:04.992 "name": "BaseBdev4", 00:10:04.992 "uuid": "0f999cee-895d-4f31-84e7-a808f30c4f0a", 00:10:04.992 "is_configured": true, 00:10:04.992 "data_offset": 2048, 00:10:04.992 "data_size": 63488 00:10:04.992 } 00:10:04.992 ] 00:10:04.992 } 00:10:04.992 } 00:10:04.992 }' 00:10:04.992 23:43:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:10:04.992 23:43:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:10:04.992 BaseBdev2 00:10:04.992 BaseBdev3 00:10:04.992 BaseBdev4' 00:10:04.992 23:43:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:04.992 23:43:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:10:04.992 23:43:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:04.992 23:43:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:10:04.992 23:43:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:04.992 23:43:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:04.992 23:43:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:04.992 23:43:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:04.992 23:43:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:04.992 23:43:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:04.993 23:43:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:04.993 23:43:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:04.993 23:43:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:10:04.993 23:43:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:04.993 23:43:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:04.993 23:43:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:04.993 23:43:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:04.993 23:43:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:04.993 23:43:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:04.993 23:43:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:10:04.993 23:43:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:04.993 23:43:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:04.993 23:43:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:04.993 23:43:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:05.251 23:43:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:05.251 23:43:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:05.251 23:43:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:05.251 23:43:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:10:05.251 23:43:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:05.251 23:43:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:05.251 23:43:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:05.251 23:43:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:05.251 23:43:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:05.251 23:43:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:05.251 23:43:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:10:05.251 23:43:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:05.251 23:43:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:05.251 [2024-11-26 23:43:53.202697] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:10:05.251 [2024-11-26 23:43:53.202771] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:05.251 [2024-11-26 23:43:53.202874] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:05.251 [2024-11-26 23:43:53.203172] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:05.251 [2024-11-26 23:43:53.203231] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name Existed_Raid, state offline 00:10:05.251 23:43:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:05.251 23:43:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 84334 00:10:05.251 23:43:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # '[' -z 84334 ']' 00:10:05.251 23:43:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@958 -- # kill -0 84334 00:10:05.251 23:43:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # uname 00:10:05.251 23:43:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:10:05.251 23:43:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 84334 00:10:05.251 23:43:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:10:05.251 23:43:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:10:05.251 23:43:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@972 -- # echo 'killing process with pid 84334' 00:10:05.251 killing process with pid 84334 00:10:05.251 23:43:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@973 -- # kill 84334 00:10:05.251 [2024-11-26 23:43:53.250600] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:10:05.251 23:43:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@978 -- # wait 84334 00:10:05.251 [2024-11-26 23:43:53.290893] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:10:05.517 23:43:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:10:05.517 00:10:05.517 real 0m9.522s 00:10:05.517 user 0m16.382s 00:10:05.517 sys 0m1.908s 00:10:05.517 ************************************ 00:10:05.517 END TEST raid_state_function_test_sb 00:10:05.517 ************************************ 00:10:05.517 23:43:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1130 -- # xtrace_disable 00:10:05.517 23:43:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:05.517 23:43:53 bdev_raid -- bdev/bdev_raid.sh@970 -- # run_test raid_superblock_test raid_superblock_test raid1 4 00:10:05.517 23:43:53 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:10:05.517 23:43:53 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:10:05.517 23:43:53 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:10:05.517 ************************************ 00:10:05.517 START TEST raid_superblock_test 00:10:05.517 ************************************ 00:10:05.517 23:43:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1129 -- # raid_superblock_test raid1 4 00:10:05.517 23:43:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=raid1 00:10:05.517 23:43:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=4 00:10:05.517 23:43:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:10:05.517 23:43:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:10:05.517 23:43:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:10:05.517 23:43:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:10:05.517 23:43:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:10:05.517 23:43:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:10:05.517 23:43:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:10:05.517 23:43:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:10:05.517 23:43:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:10:05.518 23:43:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:10:05.518 23:43:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:10:05.518 23:43:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' raid1 '!=' raid1 ']' 00:10:05.518 23:43:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@408 -- # strip_size=0 00:10:05.518 23:43:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=84983 00:10:05.518 23:43:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:10:05.518 23:43:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 84983 00:10:05.518 23:43:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@835 -- # '[' -z 84983 ']' 00:10:05.518 23:43:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:05.518 23:43:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:10:05.518 23:43:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:05.518 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:05.518 23:43:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:10:05.518 23:43:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:05.779 [2024-11-26 23:43:53.660773] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:10:05.779 [2024-11-26 23:43:53.660903] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid84983 ] 00:10:05.779 [2024-11-26 23:43:53.815899] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:10:05.779 [2024-11-26 23:43:53.841648] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:10:05.780 [2024-11-26 23:43:53.883777] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:05.780 [2024-11-26 23:43:53.883814] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:06.715 23:43:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:10:06.715 23:43:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@868 -- # return 0 00:10:06.715 23:43:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:10:06.715 23:43:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:10:06.715 23:43:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:10:06.715 23:43:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:10:06.715 23:43:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:10:06.715 23:43:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:10:06.715 23:43:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:10:06.715 23:43:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:10:06.715 23:43:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:10:06.715 23:43:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:06.715 23:43:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:06.715 malloc1 00:10:06.715 23:43:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:06.715 23:43:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:10:06.715 23:43:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:06.715 23:43:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:06.715 [2024-11-26 23:43:54.519497] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:10:06.715 [2024-11-26 23:43:54.519614] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:06.715 [2024-11-26 23:43:54.519676] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:10:06.715 [2024-11-26 23:43:54.519716] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:06.715 [2024-11-26 23:43:54.521851] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:06.715 [2024-11-26 23:43:54.521930] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:10:06.715 pt1 00:10:06.715 23:43:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:06.715 23:43:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:10:06.716 23:43:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:10:06.716 23:43:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:10:06.716 23:43:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:10:06.716 23:43:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:10:06.716 23:43:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:10:06.716 23:43:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:10:06.716 23:43:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:10:06.716 23:43:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:10:06.716 23:43:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:06.716 23:43:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:06.716 malloc2 00:10:06.716 23:43:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:06.716 23:43:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:10:06.716 23:43:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:06.716 23:43:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:06.716 [2024-11-26 23:43:54.552095] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:10:06.716 [2024-11-26 23:43:54.552213] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:06.716 [2024-11-26 23:43:54.552251] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:10:06.716 [2024-11-26 23:43:54.552286] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:06.716 [2024-11-26 23:43:54.554408] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:06.716 [2024-11-26 23:43:54.554484] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:10:06.716 pt2 00:10:06.716 23:43:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:06.716 23:43:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:10:06.716 23:43:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:10:06.716 23:43:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc3 00:10:06.716 23:43:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt3 00:10:06.716 23:43:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000003 00:10:06.716 23:43:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:10:06.716 23:43:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:10:06.716 23:43:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:10:06.716 23:43:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc3 00:10:06.716 23:43:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:06.716 23:43:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:06.716 malloc3 00:10:06.716 23:43:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:06.716 23:43:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:10:06.716 23:43:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:06.716 23:43:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:06.716 [2024-11-26 23:43:54.580542] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:10:06.716 [2024-11-26 23:43:54.580635] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:06.716 [2024-11-26 23:43:54.580686] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:10:06.716 [2024-11-26 23:43:54.580714] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:06.716 [2024-11-26 23:43:54.582756] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:06.716 [2024-11-26 23:43:54.582845] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:10:06.716 pt3 00:10:06.716 23:43:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:06.716 23:43:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:10:06.716 23:43:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:10:06.716 23:43:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc4 00:10:06.716 23:43:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt4 00:10:06.716 23:43:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000004 00:10:06.716 23:43:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:10:06.716 23:43:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:10:06.716 23:43:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:10:06.716 23:43:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc4 00:10:06.716 23:43:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:06.716 23:43:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:06.716 malloc4 00:10:06.716 23:43:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:06.716 23:43:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:10:06.716 23:43:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:06.716 23:43:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:06.716 [2024-11-26 23:43:54.633784] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:10:06.716 [2024-11-26 23:43:54.633963] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:06.716 [2024-11-26 23:43:54.634000] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:10:06.716 [2024-11-26 23:43:54.634026] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:06.716 [2024-11-26 23:43:54.638212] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:06.716 [2024-11-26 23:43:54.638279] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:10:06.716 pt4 00:10:06.716 23:43:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:06.716 23:43:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:10:06.716 23:43:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:10:06.716 23:43:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''pt1 pt2 pt3 pt4'\''' -n raid_bdev1 -s 00:10:06.716 23:43:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:06.716 23:43:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:06.716 [2024-11-26 23:43:54.646539] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:10:06.716 [2024-11-26 23:43:54.649050] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:10:06.716 [2024-11-26 23:43:54.649208] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:10:06.716 [2024-11-26 23:43:54.649313] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:10:06.716 [2024-11-26 23:43:54.649542] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:10:06.716 [2024-11-26 23:43:54.649563] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:10:06.716 [2024-11-26 23:43:54.649921] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:10:06.716 [2024-11-26 23:43:54.650136] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:10:06.716 [2024-11-26 23:43:54.650150] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:10:06.716 [2024-11-26 23:43:54.650323] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:06.716 23:43:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:06.716 23:43:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:10:06.716 23:43:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:06.716 23:43:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:06.716 23:43:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:06.716 23:43:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:06.716 23:43:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:06.716 23:43:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:06.716 23:43:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:06.716 23:43:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:06.716 23:43:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:06.716 23:43:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:06.716 23:43:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:06.716 23:43:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:06.716 23:43:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:06.716 23:43:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:06.716 23:43:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:06.716 "name": "raid_bdev1", 00:10:06.716 "uuid": "5d0721fd-b970-445d-80e4-2670a6d1e0e3", 00:10:06.716 "strip_size_kb": 0, 00:10:06.716 "state": "online", 00:10:06.716 "raid_level": "raid1", 00:10:06.716 "superblock": true, 00:10:06.716 "num_base_bdevs": 4, 00:10:06.716 "num_base_bdevs_discovered": 4, 00:10:06.716 "num_base_bdevs_operational": 4, 00:10:06.716 "base_bdevs_list": [ 00:10:06.716 { 00:10:06.716 "name": "pt1", 00:10:06.716 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:06.716 "is_configured": true, 00:10:06.716 "data_offset": 2048, 00:10:06.716 "data_size": 63488 00:10:06.716 }, 00:10:06.716 { 00:10:06.716 "name": "pt2", 00:10:06.716 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:06.716 "is_configured": true, 00:10:06.716 "data_offset": 2048, 00:10:06.716 "data_size": 63488 00:10:06.716 }, 00:10:06.716 { 00:10:06.716 "name": "pt3", 00:10:06.716 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:06.716 "is_configured": true, 00:10:06.716 "data_offset": 2048, 00:10:06.716 "data_size": 63488 00:10:06.716 }, 00:10:06.716 { 00:10:06.716 "name": "pt4", 00:10:06.717 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:06.717 "is_configured": true, 00:10:06.717 "data_offset": 2048, 00:10:06.717 "data_size": 63488 00:10:06.717 } 00:10:06.717 ] 00:10:06.717 }' 00:10:06.717 23:43:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:06.717 23:43:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:07.283 23:43:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:10:07.283 23:43:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:10:07.283 23:43:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:10:07.283 23:43:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:10:07.283 23:43:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:10:07.283 23:43:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:10:07.283 23:43:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:10:07.283 23:43:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:07.283 23:43:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:07.283 23:43:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:10:07.283 [2024-11-26 23:43:55.122059] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:07.283 23:43:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:07.283 23:43:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:10:07.283 "name": "raid_bdev1", 00:10:07.283 "aliases": [ 00:10:07.283 "5d0721fd-b970-445d-80e4-2670a6d1e0e3" 00:10:07.283 ], 00:10:07.283 "product_name": "Raid Volume", 00:10:07.283 "block_size": 512, 00:10:07.283 "num_blocks": 63488, 00:10:07.283 "uuid": "5d0721fd-b970-445d-80e4-2670a6d1e0e3", 00:10:07.283 "assigned_rate_limits": { 00:10:07.283 "rw_ios_per_sec": 0, 00:10:07.283 "rw_mbytes_per_sec": 0, 00:10:07.283 "r_mbytes_per_sec": 0, 00:10:07.283 "w_mbytes_per_sec": 0 00:10:07.283 }, 00:10:07.283 "claimed": false, 00:10:07.283 "zoned": false, 00:10:07.283 "supported_io_types": { 00:10:07.283 "read": true, 00:10:07.283 "write": true, 00:10:07.283 "unmap": false, 00:10:07.283 "flush": false, 00:10:07.283 "reset": true, 00:10:07.283 "nvme_admin": false, 00:10:07.283 "nvme_io": false, 00:10:07.283 "nvme_io_md": false, 00:10:07.283 "write_zeroes": true, 00:10:07.283 "zcopy": false, 00:10:07.283 "get_zone_info": false, 00:10:07.283 "zone_management": false, 00:10:07.283 "zone_append": false, 00:10:07.283 "compare": false, 00:10:07.283 "compare_and_write": false, 00:10:07.283 "abort": false, 00:10:07.283 "seek_hole": false, 00:10:07.283 "seek_data": false, 00:10:07.283 "copy": false, 00:10:07.283 "nvme_iov_md": false 00:10:07.283 }, 00:10:07.283 "memory_domains": [ 00:10:07.283 { 00:10:07.283 "dma_device_id": "system", 00:10:07.283 "dma_device_type": 1 00:10:07.283 }, 00:10:07.283 { 00:10:07.283 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:07.283 "dma_device_type": 2 00:10:07.283 }, 00:10:07.283 { 00:10:07.283 "dma_device_id": "system", 00:10:07.283 "dma_device_type": 1 00:10:07.283 }, 00:10:07.283 { 00:10:07.283 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:07.283 "dma_device_type": 2 00:10:07.283 }, 00:10:07.283 { 00:10:07.283 "dma_device_id": "system", 00:10:07.283 "dma_device_type": 1 00:10:07.283 }, 00:10:07.283 { 00:10:07.283 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:07.283 "dma_device_type": 2 00:10:07.283 }, 00:10:07.283 { 00:10:07.283 "dma_device_id": "system", 00:10:07.283 "dma_device_type": 1 00:10:07.283 }, 00:10:07.283 { 00:10:07.283 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:07.283 "dma_device_type": 2 00:10:07.283 } 00:10:07.283 ], 00:10:07.283 "driver_specific": { 00:10:07.283 "raid": { 00:10:07.283 "uuid": "5d0721fd-b970-445d-80e4-2670a6d1e0e3", 00:10:07.283 "strip_size_kb": 0, 00:10:07.283 "state": "online", 00:10:07.283 "raid_level": "raid1", 00:10:07.283 "superblock": true, 00:10:07.283 "num_base_bdevs": 4, 00:10:07.283 "num_base_bdevs_discovered": 4, 00:10:07.283 "num_base_bdevs_operational": 4, 00:10:07.283 "base_bdevs_list": [ 00:10:07.283 { 00:10:07.283 "name": "pt1", 00:10:07.283 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:07.283 "is_configured": true, 00:10:07.283 "data_offset": 2048, 00:10:07.283 "data_size": 63488 00:10:07.283 }, 00:10:07.283 { 00:10:07.283 "name": "pt2", 00:10:07.283 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:07.283 "is_configured": true, 00:10:07.283 "data_offset": 2048, 00:10:07.283 "data_size": 63488 00:10:07.283 }, 00:10:07.283 { 00:10:07.283 "name": "pt3", 00:10:07.283 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:07.283 "is_configured": true, 00:10:07.283 "data_offset": 2048, 00:10:07.283 "data_size": 63488 00:10:07.283 }, 00:10:07.283 { 00:10:07.283 "name": "pt4", 00:10:07.283 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:07.283 "is_configured": true, 00:10:07.284 "data_offset": 2048, 00:10:07.284 "data_size": 63488 00:10:07.284 } 00:10:07.284 ] 00:10:07.284 } 00:10:07.284 } 00:10:07.284 }' 00:10:07.284 23:43:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:10:07.284 23:43:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:10:07.284 pt2 00:10:07.284 pt3 00:10:07.284 pt4' 00:10:07.284 23:43:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:07.284 23:43:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:10:07.284 23:43:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:07.284 23:43:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:10:07.284 23:43:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:07.284 23:43:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:07.284 23:43:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:07.284 23:43:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:07.284 23:43:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:07.284 23:43:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:07.284 23:43:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:07.284 23:43:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:10:07.284 23:43:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:07.284 23:43:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:07.284 23:43:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:07.284 23:43:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:07.284 23:43:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:07.284 23:43:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:07.284 23:43:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:07.284 23:43:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:10:07.284 23:43:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:07.284 23:43:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:07.284 23:43:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:07.284 23:43:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:07.284 23:43:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:07.284 23:43:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:07.284 23:43:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:07.284 23:43:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt4 00:10:07.284 23:43:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:07.284 23:43:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:07.284 23:43:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:07.284 23:43:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:07.543 23:43:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:07.543 23:43:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:07.543 23:43:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:10:07.543 23:43:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:10:07.543 23:43:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:07.543 23:43:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:07.543 [2024-11-26 23:43:55.433494] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:07.543 23:43:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:07.543 23:43:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=5d0721fd-b970-445d-80e4-2670a6d1e0e3 00:10:07.543 23:43:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z 5d0721fd-b970-445d-80e4-2670a6d1e0e3 ']' 00:10:07.543 23:43:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:10:07.543 23:43:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:07.543 23:43:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:07.543 [2024-11-26 23:43:55.477136] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:10:07.543 [2024-11-26 23:43:55.477209] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:07.543 [2024-11-26 23:43:55.477286] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:07.543 [2024-11-26 23:43:55.477446] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:07.543 [2024-11-26 23:43:55.477501] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:10:07.543 23:43:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:07.543 23:43:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:07.543 23:43:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:10:07.543 23:43:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:07.543 23:43:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:07.543 23:43:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:07.543 23:43:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:10:07.543 23:43:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:10:07.543 23:43:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:10:07.543 23:43:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:10:07.543 23:43:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:07.543 23:43:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:07.543 23:43:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:07.543 23:43:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:10:07.543 23:43:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:10:07.543 23:43:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:07.543 23:43:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:07.543 23:43:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:07.543 23:43:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:10:07.543 23:43:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt3 00:10:07.543 23:43:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:07.543 23:43:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:07.543 23:43:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:07.543 23:43:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:10:07.543 23:43:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt4 00:10:07.543 23:43:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:07.543 23:43:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:07.543 23:43:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:07.543 23:43:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:10:07.543 23:43:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:10:07.543 23:43:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:07.543 23:43:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:07.543 23:43:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:07.543 23:43:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:10:07.543 23:43:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:10:07.543 23:43:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@652 -- # local es=0 00:10:07.543 23:43:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:10:07.543 23:43:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:10:07.543 23:43:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:10:07.543 23:43:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:10:07.543 23:43:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:10:07.543 23:43:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:10:07.543 23:43:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:07.543 23:43:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:07.543 [2024-11-26 23:43:55.636904] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:10:07.543 [2024-11-26 23:43:55.638841] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:10:07.543 [2024-11-26 23:43:55.638889] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc3 is claimed 00:10:07.543 [2024-11-26 23:43:55.638917] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc4 is claimed 00:10:07.543 [2024-11-26 23:43:55.638963] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:10:07.543 [2024-11-26 23:43:55.639009] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:10:07.543 [2024-11-26 23:43:55.639028] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc3 00:10:07.543 [2024-11-26 23:43:55.639043] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc4 00:10:07.543 [2024-11-26 23:43:55.639056] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:10:07.543 [2024-11-26 23:43:55.639066] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state configuring 00:10:07.543 request: 00:10:07.543 { 00:10:07.543 "name": "raid_bdev1", 00:10:07.543 "raid_level": "raid1", 00:10:07.543 "base_bdevs": [ 00:10:07.543 "malloc1", 00:10:07.543 "malloc2", 00:10:07.543 "malloc3", 00:10:07.543 "malloc4" 00:10:07.543 ], 00:10:07.543 "superblock": false, 00:10:07.543 "method": "bdev_raid_create", 00:10:07.543 "req_id": 1 00:10:07.543 } 00:10:07.543 Got JSON-RPC error response 00:10:07.543 response: 00:10:07.543 { 00:10:07.543 "code": -17, 00:10:07.543 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:10:07.543 } 00:10:07.543 23:43:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:10:07.543 23:43:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # es=1 00:10:07.543 23:43:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:10:07.543 23:43:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:10:07.543 23:43:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:10:07.544 23:43:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:07.544 23:43:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:07.544 23:43:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:07.544 23:43:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:10:07.544 23:43:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:07.803 23:43:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:10:07.803 23:43:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:10:07.803 23:43:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:10:07.803 23:43:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:07.803 23:43:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:07.803 [2024-11-26 23:43:55.700750] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:10:07.803 [2024-11-26 23:43:55.700841] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:07.803 [2024-11-26 23:43:55.700877] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:10:07.803 [2024-11-26 23:43:55.700904] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:07.803 [2024-11-26 23:43:55.703076] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:07.803 [2024-11-26 23:43:55.703144] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:10:07.803 [2024-11-26 23:43:55.703242] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:10:07.803 [2024-11-26 23:43:55.703319] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:10:07.803 pt1 00:10:07.803 23:43:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:07.803 23:43:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 4 00:10:07.803 23:43:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:07.803 23:43:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:07.803 23:43:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:07.803 23:43:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:07.803 23:43:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:07.803 23:43:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:07.803 23:43:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:07.803 23:43:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:07.803 23:43:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:07.803 23:43:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:07.803 23:43:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:07.803 23:43:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:07.803 23:43:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:07.803 23:43:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:07.803 23:43:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:07.803 "name": "raid_bdev1", 00:10:07.803 "uuid": "5d0721fd-b970-445d-80e4-2670a6d1e0e3", 00:10:07.803 "strip_size_kb": 0, 00:10:07.803 "state": "configuring", 00:10:07.803 "raid_level": "raid1", 00:10:07.803 "superblock": true, 00:10:07.803 "num_base_bdevs": 4, 00:10:07.803 "num_base_bdevs_discovered": 1, 00:10:07.803 "num_base_bdevs_operational": 4, 00:10:07.803 "base_bdevs_list": [ 00:10:07.803 { 00:10:07.803 "name": "pt1", 00:10:07.803 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:07.803 "is_configured": true, 00:10:07.803 "data_offset": 2048, 00:10:07.803 "data_size": 63488 00:10:07.803 }, 00:10:07.803 { 00:10:07.803 "name": null, 00:10:07.803 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:07.803 "is_configured": false, 00:10:07.803 "data_offset": 2048, 00:10:07.803 "data_size": 63488 00:10:07.803 }, 00:10:07.803 { 00:10:07.803 "name": null, 00:10:07.803 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:07.803 "is_configured": false, 00:10:07.803 "data_offset": 2048, 00:10:07.803 "data_size": 63488 00:10:07.803 }, 00:10:07.803 { 00:10:07.803 "name": null, 00:10:07.803 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:07.803 "is_configured": false, 00:10:07.803 "data_offset": 2048, 00:10:07.803 "data_size": 63488 00:10:07.803 } 00:10:07.803 ] 00:10:07.803 }' 00:10:07.803 23:43:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:07.803 23:43:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:08.062 23:43:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 4 -gt 2 ']' 00:10:08.062 23:43:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@472 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:10:08.062 23:43:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:08.062 23:43:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:08.321 [2024-11-26 23:43:56.195915] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:10:08.321 [2024-11-26 23:43:56.196033] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:08.321 [2024-11-26 23:43:56.196059] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009c80 00:10:08.321 [2024-11-26 23:43:56.196068] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:08.321 [2024-11-26 23:43:56.196521] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:08.321 [2024-11-26 23:43:56.196540] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:10:08.321 [2024-11-26 23:43:56.196616] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:10:08.321 [2024-11-26 23:43:56.196638] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:10:08.321 pt2 00:10:08.321 23:43:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:08.321 23:43:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@473 -- # rpc_cmd bdev_passthru_delete pt2 00:10:08.321 23:43:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:08.321 23:43:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:08.321 [2024-11-26 23:43:56.207904] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt2 00:10:08.321 23:43:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:08.321 23:43:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@474 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 4 00:10:08.321 23:43:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:08.321 23:43:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:08.321 23:43:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:08.321 23:43:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:08.321 23:43:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:08.321 23:43:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:08.321 23:43:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:08.321 23:43:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:08.321 23:43:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:08.321 23:43:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:08.321 23:43:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:08.321 23:43:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:08.321 23:43:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:08.321 23:43:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:08.321 23:43:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:08.321 "name": "raid_bdev1", 00:10:08.321 "uuid": "5d0721fd-b970-445d-80e4-2670a6d1e0e3", 00:10:08.321 "strip_size_kb": 0, 00:10:08.321 "state": "configuring", 00:10:08.321 "raid_level": "raid1", 00:10:08.321 "superblock": true, 00:10:08.321 "num_base_bdevs": 4, 00:10:08.321 "num_base_bdevs_discovered": 1, 00:10:08.321 "num_base_bdevs_operational": 4, 00:10:08.321 "base_bdevs_list": [ 00:10:08.321 { 00:10:08.321 "name": "pt1", 00:10:08.321 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:08.321 "is_configured": true, 00:10:08.321 "data_offset": 2048, 00:10:08.321 "data_size": 63488 00:10:08.321 }, 00:10:08.321 { 00:10:08.321 "name": null, 00:10:08.321 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:08.321 "is_configured": false, 00:10:08.321 "data_offset": 0, 00:10:08.321 "data_size": 63488 00:10:08.321 }, 00:10:08.321 { 00:10:08.321 "name": null, 00:10:08.321 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:08.321 "is_configured": false, 00:10:08.321 "data_offset": 2048, 00:10:08.321 "data_size": 63488 00:10:08.321 }, 00:10:08.321 { 00:10:08.321 "name": null, 00:10:08.321 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:08.321 "is_configured": false, 00:10:08.321 "data_offset": 2048, 00:10:08.321 "data_size": 63488 00:10:08.321 } 00:10:08.321 ] 00:10:08.321 }' 00:10:08.321 23:43:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:08.321 23:43:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:08.581 23:43:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:10:08.581 23:43:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:10:08.581 23:43:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:10:08.581 23:43:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:08.581 23:43:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:08.581 [2024-11-26 23:43:56.635180] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:10:08.581 [2024-11-26 23:43:56.635310] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:08.581 [2024-11-26 23:43:56.635353] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009f80 00:10:08.581 [2024-11-26 23:43:56.635384] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:08.581 [2024-11-26 23:43:56.635821] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:08.581 [2024-11-26 23:43:56.635882] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:10:08.581 [2024-11-26 23:43:56.635989] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:10:08.581 [2024-11-26 23:43:56.636044] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:10:08.581 pt2 00:10:08.581 23:43:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:08.581 23:43:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:10:08.581 23:43:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:10:08.581 23:43:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:10:08.581 23:43:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:08.581 23:43:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:08.581 [2024-11-26 23:43:56.647111] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:10:08.582 [2024-11-26 23:43:56.647161] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:08.582 [2024-11-26 23:43:56.647176] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:10:08.582 [2024-11-26 23:43:56.647185] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:08.582 [2024-11-26 23:43:56.647539] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:08.582 [2024-11-26 23:43:56.647562] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:10:08.582 [2024-11-26 23:43:56.647618] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:10:08.582 [2024-11-26 23:43:56.647653] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:10:08.582 pt3 00:10:08.582 23:43:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:08.582 23:43:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:10:08.582 23:43:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:10:08.582 23:43:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:10:08.582 23:43:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:08.582 23:43:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:08.582 [2024-11-26 23:43:56.659093] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:10:08.582 [2024-11-26 23:43:56.659144] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:08.582 [2024-11-26 23:43:56.659157] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a580 00:10:08.582 [2024-11-26 23:43:56.659167] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:08.582 [2024-11-26 23:43:56.659487] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:08.582 [2024-11-26 23:43:56.659507] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:10:08.582 [2024-11-26 23:43:56.659560] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt4 00:10:08.582 [2024-11-26 23:43:56.659579] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:10:08.582 [2024-11-26 23:43:56.659688] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:10:08.582 [2024-11-26 23:43:56.659709] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:10:08.582 [2024-11-26 23:43:56.659926] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:10:08.582 [2024-11-26 23:43:56.660049] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:10:08.582 [2024-11-26 23:43:56.660065] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:10:08.582 [2024-11-26 23:43:56.660164] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:08.582 pt4 00:10:08.582 23:43:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:08.582 23:43:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:10:08.582 23:43:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:10:08.582 23:43:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:10:08.582 23:43:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:08.582 23:43:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:08.582 23:43:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:08.582 23:43:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:08.582 23:43:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:08.582 23:43:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:08.582 23:43:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:08.582 23:43:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:08.582 23:43:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:08.582 23:43:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:08.582 23:43:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:08.582 23:43:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:08.582 23:43:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:08.582 23:43:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:08.841 23:43:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:08.841 "name": "raid_bdev1", 00:10:08.841 "uuid": "5d0721fd-b970-445d-80e4-2670a6d1e0e3", 00:10:08.841 "strip_size_kb": 0, 00:10:08.841 "state": "online", 00:10:08.841 "raid_level": "raid1", 00:10:08.841 "superblock": true, 00:10:08.841 "num_base_bdevs": 4, 00:10:08.841 "num_base_bdevs_discovered": 4, 00:10:08.841 "num_base_bdevs_operational": 4, 00:10:08.841 "base_bdevs_list": [ 00:10:08.841 { 00:10:08.841 "name": "pt1", 00:10:08.841 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:08.841 "is_configured": true, 00:10:08.841 "data_offset": 2048, 00:10:08.841 "data_size": 63488 00:10:08.841 }, 00:10:08.841 { 00:10:08.841 "name": "pt2", 00:10:08.841 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:08.841 "is_configured": true, 00:10:08.841 "data_offset": 2048, 00:10:08.841 "data_size": 63488 00:10:08.841 }, 00:10:08.841 { 00:10:08.841 "name": "pt3", 00:10:08.841 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:08.841 "is_configured": true, 00:10:08.841 "data_offset": 2048, 00:10:08.841 "data_size": 63488 00:10:08.841 }, 00:10:08.841 { 00:10:08.841 "name": "pt4", 00:10:08.841 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:08.841 "is_configured": true, 00:10:08.841 "data_offset": 2048, 00:10:08.841 "data_size": 63488 00:10:08.841 } 00:10:08.841 ] 00:10:08.841 }' 00:10:08.841 23:43:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:08.841 23:43:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:09.099 23:43:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:10:09.099 23:43:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:10:09.099 23:43:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:10:09.099 23:43:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:10:09.099 23:43:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:10:09.099 23:43:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:10:09.099 23:43:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:10:09.099 23:43:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:09.099 23:43:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:09.099 23:43:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:10:09.099 [2024-11-26 23:43:57.094738] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:09.099 23:43:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:09.099 23:43:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:10:09.099 "name": "raid_bdev1", 00:10:09.099 "aliases": [ 00:10:09.099 "5d0721fd-b970-445d-80e4-2670a6d1e0e3" 00:10:09.099 ], 00:10:09.099 "product_name": "Raid Volume", 00:10:09.099 "block_size": 512, 00:10:09.099 "num_blocks": 63488, 00:10:09.099 "uuid": "5d0721fd-b970-445d-80e4-2670a6d1e0e3", 00:10:09.099 "assigned_rate_limits": { 00:10:09.099 "rw_ios_per_sec": 0, 00:10:09.099 "rw_mbytes_per_sec": 0, 00:10:09.099 "r_mbytes_per_sec": 0, 00:10:09.099 "w_mbytes_per_sec": 0 00:10:09.099 }, 00:10:09.099 "claimed": false, 00:10:09.099 "zoned": false, 00:10:09.099 "supported_io_types": { 00:10:09.099 "read": true, 00:10:09.099 "write": true, 00:10:09.099 "unmap": false, 00:10:09.099 "flush": false, 00:10:09.099 "reset": true, 00:10:09.099 "nvme_admin": false, 00:10:09.099 "nvme_io": false, 00:10:09.099 "nvme_io_md": false, 00:10:09.099 "write_zeroes": true, 00:10:09.099 "zcopy": false, 00:10:09.099 "get_zone_info": false, 00:10:09.099 "zone_management": false, 00:10:09.099 "zone_append": false, 00:10:09.099 "compare": false, 00:10:09.099 "compare_and_write": false, 00:10:09.099 "abort": false, 00:10:09.099 "seek_hole": false, 00:10:09.099 "seek_data": false, 00:10:09.099 "copy": false, 00:10:09.099 "nvme_iov_md": false 00:10:09.099 }, 00:10:09.099 "memory_domains": [ 00:10:09.099 { 00:10:09.099 "dma_device_id": "system", 00:10:09.099 "dma_device_type": 1 00:10:09.099 }, 00:10:09.099 { 00:10:09.099 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:09.099 "dma_device_type": 2 00:10:09.099 }, 00:10:09.099 { 00:10:09.099 "dma_device_id": "system", 00:10:09.099 "dma_device_type": 1 00:10:09.099 }, 00:10:09.099 { 00:10:09.099 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:09.099 "dma_device_type": 2 00:10:09.099 }, 00:10:09.099 { 00:10:09.099 "dma_device_id": "system", 00:10:09.099 "dma_device_type": 1 00:10:09.099 }, 00:10:09.099 { 00:10:09.099 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:09.099 "dma_device_type": 2 00:10:09.099 }, 00:10:09.099 { 00:10:09.099 "dma_device_id": "system", 00:10:09.099 "dma_device_type": 1 00:10:09.099 }, 00:10:09.099 { 00:10:09.099 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:09.099 "dma_device_type": 2 00:10:09.099 } 00:10:09.099 ], 00:10:09.099 "driver_specific": { 00:10:09.099 "raid": { 00:10:09.099 "uuid": "5d0721fd-b970-445d-80e4-2670a6d1e0e3", 00:10:09.099 "strip_size_kb": 0, 00:10:09.099 "state": "online", 00:10:09.099 "raid_level": "raid1", 00:10:09.099 "superblock": true, 00:10:09.099 "num_base_bdevs": 4, 00:10:09.099 "num_base_bdevs_discovered": 4, 00:10:09.099 "num_base_bdevs_operational": 4, 00:10:09.099 "base_bdevs_list": [ 00:10:09.099 { 00:10:09.099 "name": "pt1", 00:10:09.099 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:09.099 "is_configured": true, 00:10:09.099 "data_offset": 2048, 00:10:09.099 "data_size": 63488 00:10:09.099 }, 00:10:09.099 { 00:10:09.099 "name": "pt2", 00:10:09.099 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:09.099 "is_configured": true, 00:10:09.099 "data_offset": 2048, 00:10:09.099 "data_size": 63488 00:10:09.099 }, 00:10:09.099 { 00:10:09.099 "name": "pt3", 00:10:09.099 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:09.099 "is_configured": true, 00:10:09.099 "data_offset": 2048, 00:10:09.099 "data_size": 63488 00:10:09.099 }, 00:10:09.099 { 00:10:09.099 "name": "pt4", 00:10:09.099 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:09.099 "is_configured": true, 00:10:09.099 "data_offset": 2048, 00:10:09.099 "data_size": 63488 00:10:09.099 } 00:10:09.100 ] 00:10:09.100 } 00:10:09.100 } 00:10:09.100 }' 00:10:09.100 23:43:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:10:09.100 23:43:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:10:09.100 pt2 00:10:09.100 pt3 00:10:09.100 pt4' 00:10:09.100 23:43:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:09.357 23:43:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:10:09.357 23:43:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:09.357 23:43:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:10:09.357 23:43:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:09.357 23:43:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:09.357 23:43:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:09.357 23:43:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:09.357 23:43:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:09.357 23:43:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:09.357 23:43:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:09.357 23:43:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:10:09.357 23:43:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:09.357 23:43:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:09.357 23:43:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:09.357 23:43:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:09.357 23:43:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:09.357 23:43:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:09.357 23:43:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:09.357 23:43:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:10:09.357 23:43:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:09.357 23:43:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:09.357 23:43:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:09.357 23:43:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:09.357 23:43:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:09.357 23:43:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:09.357 23:43:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:09.357 23:43:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:09.358 23:43:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt4 00:10:09.358 23:43:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:09.358 23:43:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:09.358 23:43:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:09.358 23:43:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:09.358 23:43:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:09.358 23:43:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:10:09.358 23:43:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:10:09.358 23:43:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:09.358 23:43:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:09.358 [2024-11-26 23:43:57.434139] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:09.358 23:43:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:09.358 23:43:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' 5d0721fd-b970-445d-80e4-2670a6d1e0e3 '!=' 5d0721fd-b970-445d-80e4-2670a6d1e0e3 ']' 00:10:09.358 23:43:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy raid1 00:10:09.358 23:43:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:10:09.358 23:43:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@199 -- # return 0 00:10:09.358 23:43:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@493 -- # rpc_cmd bdev_passthru_delete pt1 00:10:09.358 23:43:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:09.358 23:43:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:09.358 [2024-11-26 23:43:57.477825] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt1 00:10:09.358 23:43:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:09.358 23:43:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@496 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:10:09.358 23:43:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:09.358 23:43:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:09.358 23:43:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:09.358 23:43:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:09.358 23:43:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:09.358 23:43:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:09.358 23:43:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:09.358 23:43:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:09.358 23:43:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:09.615 23:43:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:09.615 23:43:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:09.615 23:43:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:09.615 23:43:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:09.615 23:43:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:09.615 23:43:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:09.615 "name": "raid_bdev1", 00:10:09.615 "uuid": "5d0721fd-b970-445d-80e4-2670a6d1e0e3", 00:10:09.615 "strip_size_kb": 0, 00:10:09.615 "state": "online", 00:10:09.615 "raid_level": "raid1", 00:10:09.615 "superblock": true, 00:10:09.615 "num_base_bdevs": 4, 00:10:09.615 "num_base_bdevs_discovered": 3, 00:10:09.615 "num_base_bdevs_operational": 3, 00:10:09.615 "base_bdevs_list": [ 00:10:09.615 { 00:10:09.615 "name": null, 00:10:09.615 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:09.615 "is_configured": false, 00:10:09.615 "data_offset": 0, 00:10:09.615 "data_size": 63488 00:10:09.615 }, 00:10:09.615 { 00:10:09.615 "name": "pt2", 00:10:09.615 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:09.615 "is_configured": true, 00:10:09.615 "data_offset": 2048, 00:10:09.615 "data_size": 63488 00:10:09.615 }, 00:10:09.615 { 00:10:09.615 "name": "pt3", 00:10:09.615 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:09.615 "is_configured": true, 00:10:09.615 "data_offset": 2048, 00:10:09.615 "data_size": 63488 00:10:09.615 }, 00:10:09.615 { 00:10:09.615 "name": "pt4", 00:10:09.615 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:09.615 "is_configured": true, 00:10:09.615 "data_offset": 2048, 00:10:09.615 "data_size": 63488 00:10:09.615 } 00:10:09.615 ] 00:10:09.615 }' 00:10:09.615 23:43:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:09.615 23:43:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:09.873 23:43:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@499 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:10:09.873 23:43:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:09.873 23:43:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:09.873 [2024-11-26 23:43:57.917022] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:10:09.873 [2024-11-26 23:43:57.917113] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:09.873 [2024-11-26 23:43:57.917227] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:09.873 [2024-11-26 23:43:57.917359] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:09.873 [2024-11-26 23:43:57.917414] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:10:09.873 23:43:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:09.873 23:43:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:09.873 23:43:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # jq -r '.[]' 00:10:09.873 23:43:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:09.873 23:43:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:09.873 23:43:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:09.873 23:43:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # raid_bdev= 00:10:09.873 23:43:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@501 -- # '[' -n '' ']' 00:10:09.873 23:43:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i = 1 )) 00:10:09.873 23:43:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:10:09.873 23:43:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt2 00:10:09.873 23:43:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:09.873 23:43:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:09.873 23:43:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:09.873 23:43:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:10:09.873 23:43:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:10:09.873 23:43:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt3 00:10:09.873 23:43:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:09.873 23:43:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:09.873 23:43:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:09.873 23:43:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:10:09.873 23:43:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:10:09.873 23:43:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt4 00:10:09.873 23:43:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:09.873 23:43:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:10.179 23:43:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:10.179 23:43:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:10:10.179 23:43:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:10:10.179 23:43:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i = 1 )) 00:10:10.179 23:43:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:10:10.179 23:43:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@512 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:10:10.179 23:43:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:10.179 23:43:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:10.179 [2024-11-26 23:43:58.012817] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:10:10.179 [2024-11-26 23:43:58.012873] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:10.179 [2024-11-26 23:43:58.012888] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a880 00:10:10.179 [2024-11-26 23:43:58.012898] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:10.179 [2024-11-26 23:43:58.015196] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:10.179 [2024-11-26 23:43:58.015236] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:10:10.179 [2024-11-26 23:43:58.015303] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:10:10.179 [2024-11-26 23:43:58.015338] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:10:10.179 pt2 00:10:10.179 23:43:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:10.179 23:43:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@515 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 3 00:10:10.179 23:43:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:10.179 23:43:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:10.179 23:43:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:10.179 23:43:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:10.179 23:43:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:10.179 23:43:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:10.179 23:43:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:10.179 23:43:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:10.179 23:43:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:10.179 23:43:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:10.179 23:43:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:10.179 23:43:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:10.179 23:43:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:10.179 23:43:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:10.179 23:43:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:10.179 "name": "raid_bdev1", 00:10:10.179 "uuid": "5d0721fd-b970-445d-80e4-2670a6d1e0e3", 00:10:10.179 "strip_size_kb": 0, 00:10:10.179 "state": "configuring", 00:10:10.179 "raid_level": "raid1", 00:10:10.179 "superblock": true, 00:10:10.179 "num_base_bdevs": 4, 00:10:10.179 "num_base_bdevs_discovered": 1, 00:10:10.179 "num_base_bdevs_operational": 3, 00:10:10.179 "base_bdevs_list": [ 00:10:10.179 { 00:10:10.179 "name": null, 00:10:10.179 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:10.179 "is_configured": false, 00:10:10.179 "data_offset": 2048, 00:10:10.179 "data_size": 63488 00:10:10.179 }, 00:10:10.179 { 00:10:10.179 "name": "pt2", 00:10:10.180 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:10.180 "is_configured": true, 00:10:10.180 "data_offset": 2048, 00:10:10.180 "data_size": 63488 00:10:10.180 }, 00:10:10.180 { 00:10:10.180 "name": null, 00:10:10.180 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:10.180 "is_configured": false, 00:10:10.180 "data_offset": 2048, 00:10:10.180 "data_size": 63488 00:10:10.180 }, 00:10:10.180 { 00:10:10.180 "name": null, 00:10:10.180 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:10.180 "is_configured": false, 00:10:10.180 "data_offset": 2048, 00:10:10.180 "data_size": 63488 00:10:10.180 } 00:10:10.180 ] 00:10:10.180 }' 00:10:10.180 23:43:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:10.180 23:43:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:10.437 23:43:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i++ )) 00:10:10.437 23:43:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:10:10.437 23:43:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@512 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:10:10.437 23:43:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:10.437 23:43:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:10.437 [2024-11-26 23:43:58.416171] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:10:10.437 [2024-11-26 23:43:58.416301] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:10.437 [2024-11-26 23:43:58.416348] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ae80 00:10:10.437 [2024-11-26 23:43:58.416384] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:10.437 [2024-11-26 23:43:58.416800] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:10.437 [2024-11-26 23:43:58.416859] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:10:10.437 [2024-11-26 23:43:58.416971] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:10:10.437 [2024-11-26 23:43:58.417026] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:10:10.437 pt3 00:10:10.437 23:43:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:10.437 23:43:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@515 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 3 00:10:10.437 23:43:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:10.437 23:43:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:10.437 23:43:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:10.437 23:43:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:10.437 23:43:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:10.437 23:43:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:10.437 23:43:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:10.437 23:43:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:10.437 23:43:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:10.437 23:43:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:10.437 23:43:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:10.437 23:43:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:10.437 23:43:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:10.437 23:43:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:10.437 23:43:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:10.437 "name": "raid_bdev1", 00:10:10.437 "uuid": "5d0721fd-b970-445d-80e4-2670a6d1e0e3", 00:10:10.437 "strip_size_kb": 0, 00:10:10.437 "state": "configuring", 00:10:10.437 "raid_level": "raid1", 00:10:10.437 "superblock": true, 00:10:10.437 "num_base_bdevs": 4, 00:10:10.437 "num_base_bdevs_discovered": 2, 00:10:10.437 "num_base_bdevs_operational": 3, 00:10:10.437 "base_bdevs_list": [ 00:10:10.437 { 00:10:10.437 "name": null, 00:10:10.438 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:10.438 "is_configured": false, 00:10:10.438 "data_offset": 2048, 00:10:10.438 "data_size": 63488 00:10:10.438 }, 00:10:10.438 { 00:10:10.438 "name": "pt2", 00:10:10.438 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:10.438 "is_configured": true, 00:10:10.438 "data_offset": 2048, 00:10:10.438 "data_size": 63488 00:10:10.438 }, 00:10:10.438 { 00:10:10.438 "name": "pt3", 00:10:10.438 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:10.438 "is_configured": true, 00:10:10.438 "data_offset": 2048, 00:10:10.438 "data_size": 63488 00:10:10.438 }, 00:10:10.438 { 00:10:10.438 "name": null, 00:10:10.438 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:10.438 "is_configured": false, 00:10:10.438 "data_offset": 2048, 00:10:10.438 "data_size": 63488 00:10:10.438 } 00:10:10.438 ] 00:10:10.438 }' 00:10:10.438 23:43:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:10.438 23:43:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:11.003 23:43:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i++ )) 00:10:11.004 23:43:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:10:11.004 23:43:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@519 -- # i=3 00:10:11.004 23:43:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@520 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:10:11.004 23:43:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:11.004 23:43:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:11.004 [2024-11-26 23:43:58.843410] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:10:11.004 [2024-11-26 23:43:58.843483] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:11.004 [2024-11-26 23:43:58.843503] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b180 00:10:11.004 [2024-11-26 23:43:58.843513] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:11.004 [2024-11-26 23:43:58.843923] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:11.004 [2024-11-26 23:43:58.843944] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:10:11.004 [2024-11-26 23:43:58.844020] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt4 00:10:11.004 [2024-11-26 23:43:58.844043] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:10:11.004 [2024-11-26 23:43:58.844137] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:10:11.004 [2024-11-26 23:43:58.844148] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:10:11.004 [2024-11-26 23:43:58.844406] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000026d0 00:10:11.004 [2024-11-26 23:43:58.844536] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:10:11.004 [2024-11-26 23:43:58.844544] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001c80 00:10:11.004 [2024-11-26 23:43:58.844647] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:11.004 pt4 00:10:11.004 23:43:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:11.004 23:43:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@523 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:10:11.004 23:43:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:11.004 23:43:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:11.004 23:43:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:11.004 23:43:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:11.004 23:43:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:11.004 23:43:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:11.004 23:43:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:11.004 23:43:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:11.004 23:43:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:11.004 23:43:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:11.004 23:43:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:11.004 23:43:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:11.004 23:43:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:11.004 23:43:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:11.004 23:43:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:11.004 "name": "raid_bdev1", 00:10:11.004 "uuid": "5d0721fd-b970-445d-80e4-2670a6d1e0e3", 00:10:11.004 "strip_size_kb": 0, 00:10:11.004 "state": "online", 00:10:11.004 "raid_level": "raid1", 00:10:11.004 "superblock": true, 00:10:11.004 "num_base_bdevs": 4, 00:10:11.004 "num_base_bdevs_discovered": 3, 00:10:11.004 "num_base_bdevs_operational": 3, 00:10:11.004 "base_bdevs_list": [ 00:10:11.004 { 00:10:11.004 "name": null, 00:10:11.004 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:11.004 "is_configured": false, 00:10:11.004 "data_offset": 2048, 00:10:11.004 "data_size": 63488 00:10:11.004 }, 00:10:11.004 { 00:10:11.004 "name": "pt2", 00:10:11.004 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:11.004 "is_configured": true, 00:10:11.004 "data_offset": 2048, 00:10:11.004 "data_size": 63488 00:10:11.004 }, 00:10:11.004 { 00:10:11.004 "name": "pt3", 00:10:11.004 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:11.004 "is_configured": true, 00:10:11.004 "data_offset": 2048, 00:10:11.004 "data_size": 63488 00:10:11.004 }, 00:10:11.004 { 00:10:11.004 "name": "pt4", 00:10:11.004 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:11.004 "is_configured": true, 00:10:11.004 "data_offset": 2048, 00:10:11.004 "data_size": 63488 00:10:11.004 } 00:10:11.004 ] 00:10:11.004 }' 00:10:11.004 23:43:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:11.004 23:43:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:11.262 23:43:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@526 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:10:11.262 23:43:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:11.262 23:43:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:11.262 [2024-11-26 23:43:59.294602] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:10:11.262 [2024-11-26 23:43:59.294692] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:11.262 [2024-11-26 23:43:59.294790] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:11.262 [2024-11-26 23:43:59.294880] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:11.262 [2024-11-26 23:43:59.294929] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name raid_bdev1, state offline 00:10:11.262 23:43:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:11.262 23:43:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:11.262 23:43:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # jq -r '.[]' 00:10:11.262 23:43:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:11.262 23:43:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:11.262 23:43:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:11.262 23:43:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # raid_bdev= 00:10:11.262 23:43:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@528 -- # '[' -n '' ']' 00:10:11.262 23:43:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@532 -- # '[' 4 -gt 2 ']' 00:10:11.262 23:43:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@534 -- # i=3 00:10:11.262 23:43:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@535 -- # rpc_cmd bdev_passthru_delete pt4 00:10:11.262 23:43:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:11.262 23:43:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:11.262 23:43:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:11.262 23:43:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@540 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:10:11.262 23:43:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:11.262 23:43:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:11.262 [2024-11-26 23:43:59.366480] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:10:11.262 [2024-11-26 23:43:59.366567] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:11.262 [2024-11-26 23:43:59.366617] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b480 00:10:11.262 [2024-11-26 23:43:59.366651] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:11.262 [2024-11-26 23:43:59.368893] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:11.262 [2024-11-26 23:43:59.368963] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:10:11.262 [2024-11-26 23:43:59.369086] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:10:11.262 [2024-11-26 23:43:59.369158] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:10:11.262 [2024-11-26 23:43:59.369329] bdev_raid.c:3685:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev pt2 (4) greater than existing raid bdev raid_bdev1 (2) 00:10:11.262 [2024-11-26 23:43:59.369404] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:10:11.262 [2024-11-26 23:43:59.369462] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002000 name raid_bdev1, state configuring 00:10:11.262 [2024-11-26 23:43:59.369549] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:10:11.262 [2024-11-26 23:43:59.369683] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:10:11.262 pt1 00:10:11.262 23:43:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:11.262 23:43:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@542 -- # '[' 4 -gt 2 ']' 00:10:11.262 23:43:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@545 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 3 00:10:11.262 23:43:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:11.262 23:43:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:11.262 23:43:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:11.262 23:43:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:11.262 23:43:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:11.262 23:43:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:11.262 23:43:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:11.262 23:43:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:11.262 23:43:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:11.262 23:43:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:11.262 23:43:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:11.262 23:43:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:11.262 23:43:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:11.521 23:43:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:11.521 23:43:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:11.521 "name": "raid_bdev1", 00:10:11.521 "uuid": "5d0721fd-b970-445d-80e4-2670a6d1e0e3", 00:10:11.521 "strip_size_kb": 0, 00:10:11.521 "state": "configuring", 00:10:11.521 "raid_level": "raid1", 00:10:11.521 "superblock": true, 00:10:11.521 "num_base_bdevs": 4, 00:10:11.521 "num_base_bdevs_discovered": 2, 00:10:11.521 "num_base_bdevs_operational": 3, 00:10:11.521 "base_bdevs_list": [ 00:10:11.521 { 00:10:11.521 "name": null, 00:10:11.521 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:11.521 "is_configured": false, 00:10:11.521 "data_offset": 2048, 00:10:11.521 "data_size": 63488 00:10:11.521 }, 00:10:11.521 { 00:10:11.521 "name": "pt2", 00:10:11.521 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:11.521 "is_configured": true, 00:10:11.521 "data_offset": 2048, 00:10:11.521 "data_size": 63488 00:10:11.521 }, 00:10:11.521 { 00:10:11.521 "name": "pt3", 00:10:11.521 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:11.521 "is_configured": true, 00:10:11.521 "data_offset": 2048, 00:10:11.521 "data_size": 63488 00:10:11.521 }, 00:10:11.521 { 00:10:11.521 "name": null, 00:10:11.521 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:11.521 "is_configured": false, 00:10:11.521 "data_offset": 2048, 00:10:11.521 "data_size": 63488 00:10:11.521 } 00:10:11.521 ] 00:10:11.521 }' 00:10:11.521 23:43:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:11.521 23:43:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:11.781 23:43:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@546 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:10:11.781 23:43:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@546 -- # rpc_cmd bdev_raid_get_bdevs configuring 00:10:11.781 23:43:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:11.781 23:43:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:11.781 23:43:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:11.781 23:43:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@546 -- # [[ false == \f\a\l\s\e ]] 00:10:11.781 23:43:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@549 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:10:11.781 23:43:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:11.781 23:43:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:11.781 [2024-11-26 23:43:59.845769] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:10:11.781 [2024-11-26 23:43:59.845890] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:11.781 [2024-11-26 23:43:59.845931] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ba80 00:10:11.781 [2024-11-26 23:43:59.845965] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:11.781 [2024-11-26 23:43:59.846473] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:11.781 [2024-11-26 23:43:59.846540] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:10:11.781 [2024-11-26 23:43:59.846652] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt4 00:10:11.781 [2024-11-26 23:43:59.846713] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:10:11.781 [2024-11-26 23:43:59.846852] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000002380 00:10:11.781 [2024-11-26 23:43:59.846899] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:10:11.781 [2024-11-26 23:43:59.847208] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000027a0 00:10:11.781 [2024-11-26 23:43:59.847413] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000002380 00:10:11.781 [2024-11-26 23:43:59.847457] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000002380 00:10:11.781 [2024-11-26 23:43:59.847611] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:11.781 pt4 00:10:11.781 23:43:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:11.781 23:43:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@554 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:10:11.781 23:43:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:11.781 23:43:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:11.781 23:43:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:11.781 23:43:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:11.781 23:43:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:11.781 23:43:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:11.781 23:43:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:11.781 23:43:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:11.781 23:43:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:11.781 23:43:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:11.781 23:43:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:11.781 23:43:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:11.781 23:43:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:11.781 23:43:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:11.781 23:43:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:11.781 "name": "raid_bdev1", 00:10:11.781 "uuid": "5d0721fd-b970-445d-80e4-2670a6d1e0e3", 00:10:11.781 "strip_size_kb": 0, 00:10:11.781 "state": "online", 00:10:11.781 "raid_level": "raid1", 00:10:11.781 "superblock": true, 00:10:11.781 "num_base_bdevs": 4, 00:10:11.781 "num_base_bdevs_discovered": 3, 00:10:11.781 "num_base_bdevs_operational": 3, 00:10:11.781 "base_bdevs_list": [ 00:10:11.781 { 00:10:11.781 "name": null, 00:10:11.781 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:11.781 "is_configured": false, 00:10:11.781 "data_offset": 2048, 00:10:11.781 "data_size": 63488 00:10:11.781 }, 00:10:11.781 { 00:10:11.781 "name": "pt2", 00:10:11.781 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:11.781 "is_configured": true, 00:10:11.781 "data_offset": 2048, 00:10:11.781 "data_size": 63488 00:10:11.781 }, 00:10:11.781 { 00:10:11.781 "name": "pt3", 00:10:11.781 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:11.781 "is_configured": true, 00:10:11.781 "data_offset": 2048, 00:10:11.781 "data_size": 63488 00:10:11.781 }, 00:10:11.781 { 00:10:11.781 "name": "pt4", 00:10:11.781 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:11.781 "is_configured": true, 00:10:11.781 "data_offset": 2048, 00:10:11.781 "data_size": 63488 00:10:11.781 } 00:10:11.781 ] 00:10:11.781 }' 00:10:11.781 23:43:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:11.781 23:43:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:12.368 23:44:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@555 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:10:12.368 23:44:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@555 -- # rpc_cmd bdev_raid_get_bdevs online 00:10:12.368 23:44:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:12.368 23:44:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:12.368 23:44:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:12.368 23:44:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@555 -- # [[ false == \f\a\l\s\e ]] 00:10:12.368 23:44:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@558 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:10:12.368 23:44:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@558 -- # jq -r '.[] | .uuid' 00:10:12.368 23:44:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:12.368 23:44:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:12.368 [2024-11-26 23:44:00.353171] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:12.368 23:44:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:12.368 23:44:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@558 -- # '[' 5d0721fd-b970-445d-80e4-2670a6d1e0e3 '!=' 5d0721fd-b970-445d-80e4-2670a6d1e0e3 ']' 00:10:12.368 23:44:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 84983 00:10:12.368 23:44:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # '[' -z 84983 ']' 00:10:12.368 23:44:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@958 -- # kill -0 84983 00:10:12.368 23:44:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # uname 00:10:12.368 23:44:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:10:12.368 23:44:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 84983 00:10:12.368 23:44:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:10:12.368 23:44:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:10:12.368 killing process with pid 84983 00:10:12.368 23:44:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 84983' 00:10:12.368 23:44:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@973 -- # kill 84983 00:10:12.368 [2024-11-26 23:44:00.435611] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:10:12.368 [2024-11-26 23:44:00.435700] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:12.368 [2024-11-26 23:44:00.435783] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:12.368 [2024-11-26 23:44:00.435792] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002380 name raid_bdev1, state offline 00:10:12.368 23:44:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@978 -- # wait 84983 00:10:12.368 [2024-11-26 23:44:00.479873] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:10:12.626 23:44:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:10:12.626 ************************************ 00:10:12.627 END TEST raid_superblock_test 00:10:12.627 ************************************ 00:10:12.627 00:10:12.627 real 0m7.116s 00:10:12.627 user 0m12.019s 00:10:12.627 sys 0m1.478s 00:10:12.627 23:44:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:10:12.627 23:44:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:12.627 23:44:00 bdev_raid -- bdev/bdev_raid.sh@971 -- # run_test raid_read_error_test raid_io_error_test raid1 4 read 00:10:12.886 23:44:00 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:10:12.886 23:44:00 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:10:12.886 23:44:00 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:10:12.886 ************************************ 00:10:12.886 START TEST raid_read_error_test 00:10:12.886 ************************************ 00:10:12.886 23:44:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test raid1 4 read 00:10:12.886 23:44:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid1 00:10:12.886 23:44:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=4 00:10:12.886 23:44:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=read 00:10:12.886 23:44:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:10:12.886 23:44:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:12.886 23:44:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:10:12.886 23:44:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:12.886 23:44:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:12.886 23:44:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:10:12.886 23:44:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:12.886 23:44:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:12.886 23:44:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:10:12.886 23:44:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:12.886 23:44:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:12.886 23:44:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev4 00:10:12.886 23:44:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:12.886 23:44:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:12.886 23:44:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:10:12.886 23:44:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:10:12.886 23:44:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:10:12.886 23:44:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:10:12.886 23:44:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:10:12.886 23:44:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:10:12.886 23:44:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:10:12.886 23:44:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid1 '!=' raid1 ']' 00:10:12.886 23:44:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@804 -- # strip_size=0 00:10:12.886 23:44:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:10:12.886 23:44:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.4dMRPkI9V4 00:10:12.886 23:44:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=85465 00:10:12.886 23:44:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:10:12.886 23:44:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 85465 00:10:12.886 23:44:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@835 -- # '[' -z 85465 ']' 00:10:12.886 23:44:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:12.886 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:12.886 23:44:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:10:12.886 23:44:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:12.886 23:44:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:10:12.886 23:44:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:12.886 [2024-11-26 23:44:00.870109] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:10:12.886 [2024-11-26 23:44:00.870244] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid85465 ] 00:10:12.886 [2024-11-26 23:44:01.004417] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:10:13.146 [2024-11-26 23:44:01.028605] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:10:13.146 [2024-11-26 23:44:01.071230] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:13.146 [2024-11-26 23:44:01.071265] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:13.713 23:44:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:10:13.713 23:44:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@868 -- # return 0 00:10:13.713 23:44:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:13.713 23:44:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:10:13.713 23:44:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:13.713 23:44:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:13.713 BaseBdev1_malloc 00:10:13.713 23:44:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:13.713 23:44:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:10:13.713 23:44:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:13.713 23:44:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:13.713 true 00:10:13.713 23:44:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:13.714 23:44:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:10:13.714 23:44:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:13.714 23:44:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:13.714 [2024-11-26 23:44:01.734744] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:10:13.714 [2024-11-26 23:44:01.734806] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:13.714 [2024-11-26 23:44:01.734843] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006980 00:10:13.714 [2024-11-26 23:44:01.734854] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:13.714 [2024-11-26 23:44:01.737147] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:13.714 [2024-11-26 23:44:01.737188] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:10:13.714 BaseBdev1 00:10:13.714 23:44:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:13.714 23:44:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:13.714 23:44:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:10:13.714 23:44:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:13.714 23:44:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:13.714 BaseBdev2_malloc 00:10:13.714 23:44:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:13.714 23:44:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:10:13.714 23:44:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:13.714 23:44:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:13.714 true 00:10:13.714 23:44:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:13.714 23:44:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:10:13.714 23:44:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:13.714 23:44:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:13.714 [2024-11-26 23:44:01.775367] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:10:13.714 [2024-11-26 23:44:01.775457] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:13.714 [2024-11-26 23:44:01.775480] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:10:13.714 [2024-11-26 23:44:01.775499] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:13.714 [2024-11-26 23:44:01.777600] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:13.714 [2024-11-26 23:44:01.777637] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:10:13.714 BaseBdev2 00:10:13.714 23:44:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:13.714 23:44:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:13.714 23:44:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:10:13.714 23:44:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:13.714 23:44:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:13.714 BaseBdev3_malloc 00:10:13.714 23:44:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:13.714 23:44:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:10:13.714 23:44:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:13.714 23:44:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:13.714 true 00:10:13.714 23:44:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:13.714 23:44:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:10:13.714 23:44:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:13.714 23:44:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:13.714 [2024-11-26 23:44:01.815796] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:10:13.714 [2024-11-26 23:44:01.815842] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:13.714 [2024-11-26 23:44:01.815858] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008780 00:10:13.714 [2024-11-26 23:44:01.815867] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:13.714 [2024-11-26 23:44:01.817864] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:13.714 [2024-11-26 23:44:01.817898] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:10:13.714 BaseBdev3 00:10:13.714 23:44:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:13.714 23:44:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:13.714 23:44:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:10:13.714 23:44:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:13.714 23:44:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:13.974 BaseBdev4_malloc 00:10:13.974 23:44:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:13.974 23:44:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev4_malloc 00:10:13.974 23:44:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:13.974 23:44:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:13.974 true 00:10:13.974 23:44:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:13.974 23:44:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev4_malloc -p BaseBdev4 00:10:13.974 23:44:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:13.974 23:44:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:13.974 [2024-11-26 23:44:01.866055] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev4_malloc 00:10:13.974 [2024-11-26 23:44:01.866098] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:13.974 [2024-11-26 23:44:01.866133] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:10:13.974 [2024-11-26 23:44:01.866141] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:13.974 [2024-11-26 23:44:01.868095] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:13.974 [2024-11-26 23:44:01.868198] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:10:13.974 BaseBdev4 00:10:13.974 23:44:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:13.974 23:44:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 -s 00:10:13.974 23:44:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:13.974 23:44:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:13.974 [2024-11-26 23:44:01.878073] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:13.974 [2024-11-26 23:44:01.879904] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:13.974 [2024-11-26 23:44:01.879978] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:13.974 [2024-11-26 23:44:01.880027] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:10:13.974 [2024-11-26 23:44:01.880228] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000002000 00:10:13.974 [2024-11-26 23:44:01.880239] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:10:13.974 [2024-11-26 23:44:01.880480] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002ef0 00:10:13.974 [2024-11-26 23:44:01.880611] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000002000 00:10:13.974 [2024-11-26 23:44:01.880622] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000002000 00:10:13.974 [2024-11-26 23:44:01.880739] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:13.974 23:44:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:13.974 23:44:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:10:13.974 23:44:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:13.974 23:44:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:13.974 23:44:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:13.974 23:44:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:13.974 23:44:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:13.974 23:44:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:13.974 23:44:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:13.974 23:44:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:13.974 23:44:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:13.974 23:44:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:13.974 23:44:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:13.974 23:44:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:13.974 23:44:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:13.974 23:44:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:13.974 23:44:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:13.974 "name": "raid_bdev1", 00:10:13.974 "uuid": "f51cccf4-8930-4022-a074-7ebd68f65e61", 00:10:13.974 "strip_size_kb": 0, 00:10:13.974 "state": "online", 00:10:13.974 "raid_level": "raid1", 00:10:13.974 "superblock": true, 00:10:13.974 "num_base_bdevs": 4, 00:10:13.974 "num_base_bdevs_discovered": 4, 00:10:13.974 "num_base_bdevs_operational": 4, 00:10:13.974 "base_bdevs_list": [ 00:10:13.974 { 00:10:13.974 "name": "BaseBdev1", 00:10:13.974 "uuid": "c849a54a-b418-5655-aaa6-b64d650e0d8c", 00:10:13.974 "is_configured": true, 00:10:13.974 "data_offset": 2048, 00:10:13.974 "data_size": 63488 00:10:13.974 }, 00:10:13.974 { 00:10:13.974 "name": "BaseBdev2", 00:10:13.975 "uuid": "3eb8698f-337d-51d4-9830-b8a271efaab6", 00:10:13.975 "is_configured": true, 00:10:13.975 "data_offset": 2048, 00:10:13.975 "data_size": 63488 00:10:13.975 }, 00:10:13.975 { 00:10:13.975 "name": "BaseBdev3", 00:10:13.975 "uuid": "84f53733-4de0-5319-b374-5288e096b102", 00:10:13.975 "is_configured": true, 00:10:13.975 "data_offset": 2048, 00:10:13.975 "data_size": 63488 00:10:13.975 }, 00:10:13.975 { 00:10:13.975 "name": "BaseBdev4", 00:10:13.975 "uuid": "c72967c3-d3d0-5d32-b77c-bc52132baf9c", 00:10:13.975 "is_configured": true, 00:10:13.975 "data_offset": 2048, 00:10:13.975 "data_size": 63488 00:10:13.975 } 00:10:13.975 ] 00:10:13.975 }' 00:10:13.975 23:44:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:13.975 23:44:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:14.235 23:44:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:10:14.235 23:44:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:10:14.495 [2024-11-26 23:44:02.365612] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000003090 00:10:15.434 23:44:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:10:15.434 23:44:03 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:15.434 23:44:03 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:15.434 23:44:03 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:15.434 23:44:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:10:15.434 23:44:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid1 = \r\a\i\d\1 ]] 00:10:15.434 23:44:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ read = \w\r\i\t\e ]] 00:10:15.434 23:44:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=4 00:10:15.434 23:44:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:10:15.434 23:44:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:15.434 23:44:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:15.434 23:44:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:15.434 23:44:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:15.434 23:44:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:15.434 23:44:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:15.434 23:44:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:15.434 23:44:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:15.434 23:44:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:15.434 23:44:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:15.434 23:44:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:15.434 23:44:03 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:15.434 23:44:03 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:15.434 23:44:03 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:15.434 23:44:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:15.434 "name": "raid_bdev1", 00:10:15.434 "uuid": "f51cccf4-8930-4022-a074-7ebd68f65e61", 00:10:15.434 "strip_size_kb": 0, 00:10:15.434 "state": "online", 00:10:15.434 "raid_level": "raid1", 00:10:15.434 "superblock": true, 00:10:15.434 "num_base_bdevs": 4, 00:10:15.434 "num_base_bdevs_discovered": 4, 00:10:15.434 "num_base_bdevs_operational": 4, 00:10:15.434 "base_bdevs_list": [ 00:10:15.434 { 00:10:15.434 "name": "BaseBdev1", 00:10:15.434 "uuid": "c849a54a-b418-5655-aaa6-b64d650e0d8c", 00:10:15.434 "is_configured": true, 00:10:15.434 "data_offset": 2048, 00:10:15.434 "data_size": 63488 00:10:15.434 }, 00:10:15.434 { 00:10:15.434 "name": "BaseBdev2", 00:10:15.434 "uuid": "3eb8698f-337d-51d4-9830-b8a271efaab6", 00:10:15.434 "is_configured": true, 00:10:15.434 "data_offset": 2048, 00:10:15.434 "data_size": 63488 00:10:15.434 }, 00:10:15.434 { 00:10:15.434 "name": "BaseBdev3", 00:10:15.434 "uuid": "84f53733-4de0-5319-b374-5288e096b102", 00:10:15.434 "is_configured": true, 00:10:15.434 "data_offset": 2048, 00:10:15.434 "data_size": 63488 00:10:15.434 }, 00:10:15.434 { 00:10:15.434 "name": "BaseBdev4", 00:10:15.434 "uuid": "c72967c3-d3d0-5d32-b77c-bc52132baf9c", 00:10:15.434 "is_configured": true, 00:10:15.434 "data_offset": 2048, 00:10:15.434 "data_size": 63488 00:10:15.434 } 00:10:15.434 ] 00:10:15.434 }' 00:10:15.434 23:44:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:15.434 23:44:03 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:15.694 23:44:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:10:15.694 23:44:03 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:15.694 23:44:03 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:15.694 [2024-11-26 23:44:03.700234] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:10:15.694 [2024-11-26 23:44:03.700339] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:15.694 [2024-11-26 23:44:03.703010] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:15.694 [2024-11-26 23:44:03.703099] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:15.694 [2024-11-26 23:44:03.703278] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:15.694 [2024-11-26 23:44:03.703331] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002000 name raid_bdev1, state offline 00:10:15.694 { 00:10:15.694 "results": [ 00:10:15.694 { 00:10:15.694 "job": "raid_bdev1", 00:10:15.694 "core_mask": "0x1", 00:10:15.694 "workload": "randrw", 00:10:15.694 "percentage": 50, 00:10:15.694 "status": "finished", 00:10:15.694 "queue_depth": 1, 00:10:15.694 "io_size": 131072, 00:10:15.694 "runtime": 1.33555, 00:10:15.694 "iops": 11453.708210100707, 00:10:15.694 "mibps": 1431.7135262625884, 00:10:15.694 "io_failed": 0, 00:10:15.694 "io_timeout": 0, 00:10:15.694 "avg_latency_us": 84.64714918271784, 00:10:15.694 "min_latency_us": 22.358078602620086, 00:10:15.694 "max_latency_us": 1545.3903930131005 00:10:15.694 } 00:10:15.694 ], 00:10:15.694 "core_count": 1 00:10:15.694 } 00:10:15.694 23:44:03 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:15.694 23:44:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 85465 00:10:15.694 23:44:03 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # '[' -z 85465 ']' 00:10:15.694 23:44:03 bdev_raid.raid_read_error_test -- common/autotest_common.sh@958 -- # kill -0 85465 00:10:15.694 23:44:03 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # uname 00:10:15.694 23:44:03 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:10:15.694 23:44:03 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 85465 00:10:15.694 killing process with pid 85465 00:10:15.694 23:44:03 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:10:15.694 23:44:03 bdev_raid.raid_read_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:10:15.694 23:44:03 bdev_raid.raid_read_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 85465' 00:10:15.694 23:44:03 bdev_raid.raid_read_error_test -- common/autotest_common.sh@973 -- # kill 85465 00:10:15.694 [2024-11-26 23:44:03.750669] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:10:15.694 23:44:03 bdev_raid.raid_read_error_test -- common/autotest_common.sh@978 -- # wait 85465 00:10:15.694 [2024-11-26 23:44:03.786502] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:10:15.954 23:44:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:10:15.954 23:44:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.4dMRPkI9V4 00:10:15.954 23:44:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:10:15.954 23:44:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.00 00:10:15.954 23:44:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid1 00:10:15.954 23:44:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:10:15.954 23:44:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@199 -- # return 0 00:10:15.954 23:44:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@847 -- # [[ 0.00 = \0\.\0\0 ]] 00:10:15.954 00:10:15.954 real 0m3.233s 00:10:15.954 user 0m4.021s 00:10:15.954 sys 0m0.514s 00:10:15.954 23:44:03 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:10:15.954 23:44:03 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:15.954 ************************************ 00:10:15.954 END TEST raid_read_error_test 00:10:15.954 ************************************ 00:10:15.954 23:44:04 bdev_raid -- bdev/bdev_raid.sh@972 -- # run_test raid_write_error_test raid_io_error_test raid1 4 write 00:10:15.954 23:44:04 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:10:15.954 23:44:04 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:10:15.954 23:44:04 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:10:15.954 ************************************ 00:10:15.954 START TEST raid_write_error_test 00:10:15.954 ************************************ 00:10:15.954 23:44:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test raid1 4 write 00:10:15.954 23:44:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid1 00:10:15.954 23:44:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=4 00:10:15.954 23:44:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=write 00:10:15.954 23:44:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:10:15.954 23:44:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:15.954 23:44:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:10:15.954 23:44:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:15.955 23:44:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:15.955 23:44:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:10:15.955 23:44:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:15.955 23:44:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:15.955 23:44:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:10:15.955 23:44:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:15.955 23:44:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:15.955 23:44:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev4 00:10:15.955 23:44:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:15.955 23:44:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:15.955 23:44:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:10:15.955 23:44:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:10:15.955 23:44:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:10:15.955 23:44:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:10:15.955 23:44:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:10:15.955 23:44:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:10:15.955 23:44:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:10:15.955 23:44:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid1 '!=' raid1 ']' 00:10:15.955 23:44:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@804 -- # strip_size=0 00:10:16.214 23:44:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:10:16.214 23:44:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.Ux9bFDiqNT 00:10:16.214 23:44:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=85600 00:10:16.214 23:44:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:10:16.214 23:44:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 85600 00:10:16.214 23:44:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@835 -- # '[' -z 85600 ']' 00:10:16.214 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:16.214 23:44:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:16.214 23:44:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:10:16.214 23:44:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:16.214 23:44:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:10:16.214 23:44:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:16.214 [2024-11-26 23:44:04.171534] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:10:16.214 [2024-11-26 23:44:04.171644] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid85600 ] 00:10:16.214 [2024-11-26 23:44:04.325641] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:10:16.474 [2024-11-26 23:44:04.352002] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:10:16.474 [2024-11-26 23:44:04.394891] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:16.474 [2024-11-26 23:44:04.395019] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:17.044 23:44:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:10:17.044 23:44:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@868 -- # return 0 00:10:17.044 23:44:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:17.044 23:44:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:10:17.044 23:44:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:17.044 23:44:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:17.044 BaseBdev1_malloc 00:10:17.044 23:44:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:17.044 23:44:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:10:17.044 23:44:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:17.044 23:44:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:17.044 true 00:10:17.044 23:44:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:17.044 23:44:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:10:17.044 23:44:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:17.044 23:44:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:17.044 [2024-11-26 23:44:05.038485] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:10:17.044 [2024-11-26 23:44:05.038584] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:17.044 [2024-11-26 23:44:05.038607] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006980 00:10:17.044 [2024-11-26 23:44:05.038642] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:17.044 [2024-11-26 23:44:05.040799] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:17.044 [2024-11-26 23:44:05.040837] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:10:17.044 BaseBdev1 00:10:17.044 23:44:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:17.044 23:44:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:17.044 23:44:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:10:17.044 23:44:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:17.044 23:44:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:17.044 BaseBdev2_malloc 00:10:17.044 23:44:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:17.044 23:44:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:10:17.044 23:44:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:17.044 23:44:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:17.044 true 00:10:17.044 23:44:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:17.044 23:44:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:10:17.044 23:44:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:17.044 23:44:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:17.044 [2024-11-26 23:44:05.079119] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:10:17.044 [2024-11-26 23:44:05.079168] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:17.044 [2024-11-26 23:44:05.079186] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:10:17.044 [2024-11-26 23:44:05.079203] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:17.044 [2024-11-26 23:44:05.081308] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:17.044 [2024-11-26 23:44:05.081363] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:10:17.044 BaseBdev2 00:10:17.044 23:44:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:17.044 23:44:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:17.044 23:44:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:10:17.044 23:44:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:17.044 23:44:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:17.044 BaseBdev3_malloc 00:10:17.044 23:44:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:17.044 23:44:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:10:17.044 23:44:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:17.044 23:44:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:17.044 true 00:10:17.044 23:44:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:17.044 23:44:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:10:17.044 23:44:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:17.044 23:44:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:17.044 [2024-11-26 23:44:05.119692] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:10:17.044 [2024-11-26 23:44:05.119739] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:17.044 [2024-11-26 23:44:05.119775] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008780 00:10:17.044 [2024-11-26 23:44:05.119783] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:17.044 [2024-11-26 23:44:05.121780] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:17.044 [2024-11-26 23:44:05.121869] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:10:17.044 BaseBdev3 00:10:17.044 23:44:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:17.044 23:44:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:17.044 23:44:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:10:17.044 23:44:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:17.044 23:44:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:17.044 BaseBdev4_malloc 00:10:17.044 23:44:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:17.044 23:44:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev4_malloc 00:10:17.044 23:44:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:17.044 23:44:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:17.044 true 00:10:17.044 23:44:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:17.044 23:44:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev4_malloc -p BaseBdev4 00:10:17.044 23:44:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:17.044 23:44:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:17.044 [2024-11-26 23:44:05.168470] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev4_malloc 00:10:17.045 [2024-11-26 23:44:05.168519] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:17.045 [2024-11-26 23:44:05.168556] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:10:17.045 [2024-11-26 23:44:05.168565] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:17.045 [2024-11-26 23:44:05.170606] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:17.045 [2024-11-26 23:44:05.170706] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:10:17.305 BaseBdev4 00:10:17.305 23:44:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:17.305 23:44:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 -s 00:10:17.305 23:44:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:17.305 23:44:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:17.305 [2024-11-26 23:44:05.180494] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:17.305 [2024-11-26 23:44:05.182280] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:17.305 [2024-11-26 23:44:05.182421] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:17.305 [2024-11-26 23:44:05.182480] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:10:17.305 [2024-11-26 23:44:05.182678] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000002000 00:10:17.305 [2024-11-26 23:44:05.182691] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:10:17.305 [2024-11-26 23:44:05.182941] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002ef0 00:10:17.305 [2024-11-26 23:44:05.183084] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000002000 00:10:17.305 [2024-11-26 23:44:05.183096] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000002000 00:10:17.305 [2024-11-26 23:44:05.183206] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:17.305 23:44:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:17.305 23:44:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:10:17.305 23:44:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:17.305 23:44:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:17.305 23:44:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:17.305 23:44:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:17.305 23:44:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:17.305 23:44:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:17.305 23:44:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:17.305 23:44:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:17.305 23:44:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:17.305 23:44:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:17.305 23:44:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:17.305 23:44:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:17.305 23:44:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:17.305 23:44:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:17.305 23:44:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:17.305 "name": "raid_bdev1", 00:10:17.305 "uuid": "26f23325-9b12-470a-95cc-e303d653aa99", 00:10:17.305 "strip_size_kb": 0, 00:10:17.305 "state": "online", 00:10:17.305 "raid_level": "raid1", 00:10:17.305 "superblock": true, 00:10:17.305 "num_base_bdevs": 4, 00:10:17.305 "num_base_bdevs_discovered": 4, 00:10:17.305 "num_base_bdevs_operational": 4, 00:10:17.305 "base_bdevs_list": [ 00:10:17.305 { 00:10:17.305 "name": "BaseBdev1", 00:10:17.305 "uuid": "aa600380-ea8f-5045-a3df-88ce4a2f8226", 00:10:17.305 "is_configured": true, 00:10:17.305 "data_offset": 2048, 00:10:17.305 "data_size": 63488 00:10:17.305 }, 00:10:17.305 { 00:10:17.305 "name": "BaseBdev2", 00:10:17.305 "uuid": "bebe3996-b192-5e29-9f16-e3a3d3b39930", 00:10:17.305 "is_configured": true, 00:10:17.305 "data_offset": 2048, 00:10:17.305 "data_size": 63488 00:10:17.305 }, 00:10:17.305 { 00:10:17.305 "name": "BaseBdev3", 00:10:17.305 "uuid": "001fe03a-0c61-587c-a6a2-307c0ba83a01", 00:10:17.305 "is_configured": true, 00:10:17.305 "data_offset": 2048, 00:10:17.305 "data_size": 63488 00:10:17.305 }, 00:10:17.305 { 00:10:17.305 "name": "BaseBdev4", 00:10:17.305 "uuid": "d150632b-190a-58fc-8d58-c67c5d80a97f", 00:10:17.305 "is_configured": true, 00:10:17.305 "data_offset": 2048, 00:10:17.305 "data_size": 63488 00:10:17.305 } 00:10:17.305 ] 00:10:17.305 }' 00:10:17.305 23:44:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:17.305 23:44:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:17.564 23:44:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:10:17.565 23:44:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:10:17.823 [2024-11-26 23:44:05.727920] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000003090 00:10:18.760 23:44:06 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:10:18.760 23:44:06 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:18.760 23:44:06 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:18.760 [2024-11-26 23:44:06.647181] bdev_raid.c:2276:_raid_bdev_fail_base_bdev: *NOTICE*: Failing base bdev in slot 0 ('BaseBdev1') of raid bdev 'raid_bdev1' 00:10:18.760 [2024-11-26 23:44:06.647328] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:10:18.760 [2024-11-26 23:44:06.647616] bdev_raid.c:1974:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 0 raid_ch: 0x60d000003090 00:10:18.760 23:44:06 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:18.760 23:44:06 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:10:18.760 23:44:06 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid1 = \r\a\i\d\1 ]] 00:10:18.760 23:44:06 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ write = \w\r\i\t\e ]] 00:10:18.760 23:44:06 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@833 -- # expected_num_base_bdevs=3 00:10:18.760 23:44:06 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:10:18.760 23:44:06 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:18.760 23:44:06 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:18.760 23:44:06 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:18.760 23:44:06 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:18.760 23:44:06 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:18.760 23:44:06 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:18.760 23:44:06 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:18.760 23:44:06 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:18.760 23:44:06 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:18.760 23:44:06 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:18.760 23:44:06 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:18.760 23:44:06 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:18.760 23:44:06 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:18.760 23:44:06 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:18.760 23:44:06 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:18.760 "name": "raid_bdev1", 00:10:18.760 "uuid": "26f23325-9b12-470a-95cc-e303d653aa99", 00:10:18.760 "strip_size_kb": 0, 00:10:18.760 "state": "online", 00:10:18.760 "raid_level": "raid1", 00:10:18.760 "superblock": true, 00:10:18.760 "num_base_bdevs": 4, 00:10:18.760 "num_base_bdevs_discovered": 3, 00:10:18.760 "num_base_bdevs_operational": 3, 00:10:18.760 "base_bdevs_list": [ 00:10:18.760 { 00:10:18.760 "name": null, 00:10:18.760 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:18.760 "is_configured": false, 00:10:18.760 "data_offset": 0, 00:10:18.760 "data_size": 63488 00:10:18.760 }, 00:10:18.760 { 00:10:18.760 "name": "BaseBdev2", 00:10:18.760 "uuid": "bebe3996-b192-5e29-9f16-e3a3d3b39930", 00:10:18.760 "is_configured": true, 00:10:18.760 "data_offset": 2048, 00:10:18.760 "data_size": 63488 00:10:18.760 }, 00:10:18.760 { 00:10:18.760 "name": "BaseBdev3", 00:10:18.760 "uuid": "001fe03a-0c61-587c-a6a2-307c0ba83a01", 00:10:18.760 "is_configured": true, 00:10:18.760 "data_offset": 2048, 00:10:18.760 "data_size": 63488 00:10:18.760 }, 00:10:18.760 { 00:10:18.760 "name": "BaseBdev4", 00:10:18.760 "uuid": "d150632b-190a-58fc-8d58-c67c5d80a97f", 00:10:18.760 "is_configured": true, 00:10:18.760 "data_offset": 2048, 00:10:18.760 "data_size": 63488 00:10:18.760 } 00:10:18.760 ] 00:10:18.760 }' 00:10:18.760 23:44:06 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:18.760 23:44:06 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:19.020 23:44:07 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:10:19.020 23:44:07 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:19.020 23:44:07 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:19.020 [2024-11-26 23:44:07.127655] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:10:19.020 [2024-11-26 23:44:07.127687] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:19.020 [2024-11-26 23:44:07.130518] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:19.020 [2024-11-26 23:44:07.130618] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:19.020 [2024-11-26 23:44:07.130725] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:19.020 [2024-11-26 23:44:07.130749] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002000 name raid_bdev1, state offline 00:10:19.020 { 00:10:19.020 "results": [ 00:10:19.020 { 00:10:19.020 "job": "raid_bdev1", 00:10:19.020 "core_mask": "0x1", 00:10:19.020 "workload": "randrw", 00:10:19.020 "percentage": 50, 00:10:19.020 "status": "finished", 00:10:19.020 "queue_depth": 1, 00:10:19.020 "io_size": 131072, 00:10:19.020 "runtime": 1.400623, 00:10:19.020 "iops": 12206.71087080535, 00:10:19.020 "mibps": 1525.8388588506687, 00:10:19.020 "io_failed": 0, 00:10:19.020 "io_timeout": 0, 00:10:19.020 "avg_latency_us": 79.17870889783008, 00:10:19.020 "min_latency_us": 23.699563318777294, 00:10:19.020 "max_latency_us": 1409.4532751091704 00:10:19.020 } 00:10:19.020 ], 00:10:19.020 "core_count": 1 00:10:19.020 } 00:10:19.020 23:44:07 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:19.020 23:44:07 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 85600 00:10:19.020 23:44:07 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # '[' -z 85600 ']' 00:10:19.020 23:44:07 bdev_raid.raid_write_error_test -- common/autotest_common.sh@958 -- # kill -0 85600 00:10:19.020 23:44:07 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # uname 00:10:19.020 23:44:07 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:10:19.020 23:44:07 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 85600 00:10:19.280 killing process with pid 85600 00:10:19.280 23:44:07 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:10:19.280 23:44:07 bdev_raid.raid_write_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:10:19.280 23:44:07 bdev_raid.raid_write_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 85600' 00:10:19.280 23:44:07 bdev_raid.raid_write_error_test -- common/autotest_common.sh@973 -- # kill 85600 00:10:19.280 [2024-11-26 23:44:07.174378] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:10:19.280 23:44:07 bdev_raid.raid_write_error_test -- common/autotest_common.sh@978 -- # wait 85600 00:10:19.280 [2024-11-26 23:44:07.209991] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:10:19.540 23:44:07 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.Ux9bFDiqNT 00:10:19.540 23:44:07 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:10:19.540 23:44:07 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:10:19.540 23:44:07 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.00 00:10:19.540 23:44:07 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid1 00:10:19.540 23:44:07 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:10:19.540 23:44:07 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@199 -- # return 0 00:10:19.540 23:44:07 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@847 -- # [[ 0.00 = \0\.\0\0 ]] 00:10:19.540 00:10:19.540 real 0m3.355s 00:10:19.540 user 0m4.256s 00:10:19.540 sys 0m0.530s 00:10:19.540 ************************************ 00:10:19.540 END TEST raid_write_error_test 00:10:19.540 ************************************ 00:10:19.540 23:44:07 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:10:19.540 23:44:07 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:19.540 23:44:07 bdev_raid -- bdev/bdev_raid.sh@976 -- # '[' true = true ']' 00:10:19.540 23:44:07 bdev_raid -- bdev/bdev_raid.sh@977 -- # for n in 2 4 00:10:19.540 23:44:07 bdev_raid -- bdev/bdev_raid.sh@978 -- # run_test raid_rebuild_test raid_rebuild_test raid1 2 false false true 00:10:19.540 23:44:07 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:10:19.540 23:44:07 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:10:19.540 23:44:07 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:10:19.540 ************************************ 00:10:19.540 START TEST raid_rebuild_test 00:10:19.540 ************************************ 00:10:19.540 23:44:07 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@1129 -- # raid_rebuild_test raid1 2 false false true 00:10:19.540 23:44:07 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:10:19.540 23:44:07 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=2 00:10:19.540 23:44:07 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@571 -- # local superblock=false 00:10:19.540 23:44:07 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:10:19.540 23:44:07 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@573 -- # local verify=true 00:10:19.540 23:44:07 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:10:19.540 23:44:07 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:10:19.540 23:44:07 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:10:19.540 23:44:07 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:10:19.540 23:44:07 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:10:19.540 23:44:07 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:10:19.540 23:44:07 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:10:19.540 23:44:07 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:10:19.540 23:44:07 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:10:19.540 23:44:07 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:10:19.540 23:44:07 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:10:19.540 23:44:07 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@576 -- # local strip_size 00:10:19.540 23:44:07 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@577 -- # local create_arg 00:10:19.540 23:44:07 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:10:19.540 23:44:07 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@579 -- # local data_offset 00:10:19.540 23:44:07 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:10:19.540 23:44:07 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:10:19.540 23:44:07 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@592 -- # '[' false = true ']' 00:10:19.540 23:44:07 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@597 -- # raid_pid=85728 00:10:19.540 23:44:07 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:10:19.540 23:44:07 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@598 -- # waitforlisten 85728 00:10:19.540 23:44:07 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@835 -- # '[' -z 85728 ']' 00:10:19.540 23:44:07 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:19.540 23:44:07 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:10:19.540 23:44:07 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:19.540 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:19.540 23:44:07 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:10:19.540 23:44:07 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:10:19.540 [2024-11-26 23:44:07.590441] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:10:19.540 [2024-11-26 23:44:07.590656] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.ealI/O size of 3145728 is greater than zero copy threshold (65536). 00:10:19.540 Zero copy mechanism will not be used. 00:10:19.540 :6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid85728 ] 00:10:19.804 [2024-11-26 23:44:07.744760] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:10:19.804 [2024-11-26 23:44:07.772063] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:10:19.804 [2024-11-26 23:44:07.814123] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:19.804 [2024-11-26 23:44:07.814157] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:20.380 23:44:08 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:10:20.380 23:44:08 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@868 -- # return 0 00:10:20.380 23:44:08 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:10:20.380 23:44:08 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:10:20.380 23:44:08 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:20.380 23:44:08 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:10:20.380 BaseBdev1_malloc 00:10:20.380 23:44:08 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:20.380 23:44:08 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:10:20.380 23:44:08 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:20.380 23:44:08 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:10:20.380 [2024-11-26 23:44:08.437184] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:10:20.380 [2024-11-26 23:44:08.437251] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:20.380 [2024-11-26 23:44:08.437277] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:10:20.380 [2024-11-26 23:44:08.437288] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:20.380 [2024-11-26 23:44:08.439486] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:20.380 [2024-11-26 23:44:08.439524] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:10:20.380 BaseBdev1 00:10:20.380 23:44:08 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:20.380 23:44:08 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:10:20.380 23:44:08 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:10:20.380 23:44:08 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:20.380 23:44:08 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:10:20.380 BaseBdev2_malloc 00:10:20.380 23:44:08 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:20.381 23:44:08 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:10:20.381 23:44:08 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:20.381 23:44:08 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:10:20.381 [2024-11-26 23:44:08.465744] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:10:20.381 [2024-11-26 23:44:08.465799] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:20.381 [2024-11-26 23:44:08.465838] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:10:20.381 [2024-11-26 23:44:08.465847] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:20.381 [2024-11-26 23:44:08.467976] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:20.381 [2024-11-26 23:44:08.468017] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:10:20.381 BaseBdev2 00:10:20.381 23:44:08 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:20.381 23:44:08 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:10:20.381 23:44:08 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:20.381 23:44:08 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:10:20.381 spare_malloc 00:10:20.381 23:44:08 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:20.381 23:44:08 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:10:20.381 23:44:08 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:20.381 23:44:08 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:10:20.381 spare_delay 00:10:20.381 23:44:08 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:20.381 23:44:08 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:10:20.381 23:44:08 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:20.381 23:44:08 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:10:20.381 [2024-11-26 23:44:08.506327] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:10:20.381 [2024-11-26 23:44:08.506388] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:20.381 [2024-11-26 23:44:08.506423] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:10:20.381 [2024-11-26 23:44:08.506432] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:20.381 [2024-11-26 23:44:08.508481] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:20.381 [2024-11-26 23:44:08.508517] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:10:20.641 spare 00:10:20.641 23:44:08 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:20.641 23:44:08 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 00:10:20.641 23:44:08 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:20.642 23:44:08 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:10:20.642 [2024-11-26 23:44:08.518354] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:20.642 [2024-11-26 23:44:08.520107] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:20.642 [2024-11-26 23:44:08.520194] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:10:20.642 [2024-11-26 23:44:08.520204] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:10:20.642 [2024-11-26 23:44:08.520489] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:10:20.642 [2024-11-26 23:44:08.520628] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:10:20.642 [2024-11-26 23:44:08.520642] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:10:20.642 [2024-11-26 23:44:08.520761] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:20.642 23:44:08 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:20.642 23:44:08 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:10:20.642 23:44:08 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:20.642 23:44:08 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:20.642 23:44:08 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:20.642 23:44:08 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:20.642 23:44:08 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:10:20.642 23:44:08 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:20.642 23:44:08 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:20.642 23:44:08 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:20.642 23:44:08 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:20.642 23:44:08 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:20.642 23:44:08 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:20.642 23:44:08 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:20.642 23:44:08 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:10:20.642 23:44:08 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:20.642 23:44:08 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:20.642 "name": "raid_bdev1", 00:10:20.642 "uuid": "b039aaa7-5122-46ec-8ced-c6f237aa3d51", 00:10:20.642 "strip_size_kb": 0, 00:10:20.642 "state": "online", 00:10:20.642 "raid_level": "raid1", 00:10:20.642 "superblock": false, 00:10:20.642 "num_base_bdevs": 2, 00:10:20.642 "num_base_bdevs_discovered": 2, 00:10:20.642 "num_base_bdevs_operational": 2, 00:10:20.642 "base_bdevs_list": [ 00:10:20.642 { 00:10:20.642 "name": "BaseBdev1", 00:10:20.642 "uuid": "dce8918d-75ad-5b64-88c1-2287eea40791", 00:10:20.642 "is_configured": true, 00:10:20.642 "data_offset": 0, 00:10:20.642 "data_size": 65536 00:10:20.642 }, 00:10:20.642 { 00:10:20.642 "name": "BaseBdev2", 00:10:20.642 "uuid": "d4bb21ce-e99c-5091-bc05-8ecd2175e0d4", 00:10:20.642 "is_configured": true, 00:10:20.642 "data_offset": 0, 00:10:20.642 "data_size": 65536 00:10:20.642 } 00:10:20.642 ] 00:10:20.642 }' 00:10:20.642 23:44:08 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:20.642 23:44:08 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:10:20.902 23:44:08 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:10:20.902 23:44:08 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:10:20.902 23:44:08 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:20.902 23:44:08 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:10:20.902 [2024-11-26 23:44:08.941865] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:20.902 23:44:08 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:20.902 23:44:08 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=65536 00:10:20.902 23:44:08 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:20.902 23:44:08 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:10:20.902 23:44:08 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:20.902 23:44:08 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:10:20.902 23:44:08 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:20.902 23:44:09 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@619 -- # data_offset=0 00:10:20.902 23:44:09 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:10:20.902 23:44:09 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:10:20.902 23:44:09 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:10:20.902 23:44:09 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:10:20.902 23:44:09 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:10:20.902 23:44:09 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:10:20.902 23:44:09 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@10 -- # local bdev_list 00:10:20.902 23:44:09 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:10:20.902 23:44:09 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@11 -- # local nbd_list 00:10:20.902 23:44:09 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@12 -- # local i 00:10:20.902 23:44:09 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:10:20.902 23:44:09 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:10:20.902 23:44:09 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:10:21.162 [2024-11-26 23:44:09.209241] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:10:21.162 /dev/nbd0 00:10:21.162 23:44:09 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:10:21.162 23:44:09 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:10:21.162 23:44:09 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:10:21.162 23:44:09 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@873 -- # local i 00:10:21.162 23:44:09 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:10:21.162 23:44:09 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:10:21.162 23:44:09 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:10:21.162 23:44:09 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@877 -- # break 00:10:21.162 23:44:09 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:10:21.162 23:44:09 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:10:21.162 23:44:09 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:10:21.162 1+0 records in 00:10:21.162 1+0 records out 00:10:21.162 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000308757 s, 13.3 MB/s 00:10:21.162 23:44:09 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:10:21.162 23:44:09 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@890 -- # size=4096 00:10:21.162 23:44:09 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:10:21.162 23:44:09 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:10:21.163 23:44:09 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@893 -- # return 0 00:10:21.163 23:44:09 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:10:21.163 23:44:09 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:10:21.163 23:44:09 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@629 -- # '[' raid1 = raid5f ']' 00:10:21.163 23:44:09 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@633 -- # write_unit_size=1 00:10:21.163 23:44:09 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=512 count=65536 oflag=direct 00:10:25.353 65536+0 records in 00:10:25.353 65536+0 records out 00:10:25.353 33554432 bytes (34 MB, 32 MiB) copied, 3.80775 s, 8.8 MB/s 00:10:25.353 23:44:13 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:10:25.353 23:44:13 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:10:25.353 23:44:13 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:10:25.353 23:44:13 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@50 -- # local nbd_list 00:10:25.353 23:44:13 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@51 -- # local i 00:10:25.353 23:44:13 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:10:25.353 23:44:13 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:10:25.353 [2024-11-26 23:44:13.288236] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:25.353 23:44:13 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:10:25.353 23:44:13 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:10:25.353 23:44:13 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:10:25.353 23:44:13 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:10:25.353 23:44:13 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:10:25.353 23:44:13 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:10:25.353 23:44:13 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:10:25.353 23:44:13 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:10:25.354 23:44:13 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:10:25.354 23:44:13 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:25.354 23:44:13 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:10:25.354 [2024-11-26 23:44:13.328293] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:10:25.354 23:44:13 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:25.354 23:44:13 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:10:25.354 23:44:13 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:25.354 23:44:13 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:25.354 23:44:13 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:25.354 23:44:13 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:25.354 23:44:13 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:10:25.354 23:44:13 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:25.354 23:44:13 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:25.354 23:44:13 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:25.354 23:44:13 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:25.354 23:44:13 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:25.354 23:44:13 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:25.354 23:44:13 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:25.354 23:44:13 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:10:25.354 23:44:13 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:25.354 23:44:13 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:25.354 "name": "raid_bdev1", 00:10:25.354 "uuid": "b039aaa7-5122-46ec-8ced-c6f237aa3d51", 00:10:25.354 "strip_size_kb": 0, 00:10:25.354 "state": "online", 00:10:25.354 "raid_level": "raid1", 00:10:25.354 "superblock": false, 00:10:25.354 "num_base_bdevs": 2, 00:10:25.354 "num_base_bdevs_discovered": 1, 00:10:25.354 "num_base_bdevs_operational": 1, 00:10:25.354 "base_bdevs_list": [ 00:10:25.354 { 00:10:25.354 "name": null, 00:10:25.354 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:25.354 "is_configured": false, 00:10:25.354 "data_offset": 0, 00:10:25.354 "data_size": 65536 00:10:25.354 }, 00:10:25.354 { 00:10:25.354 "name": "BaseBdev2", 00:10:25.355 "uuid": "d4bb21ce-e99c-5091-bc05-8ecd2175e0d4", 00:10:25.355 "is_configured": true, 00:10:25.355 "data_offset": 0, 00:10:25.355 "data_size": 65536 00:10:25.355 } 00:10:25.355 ] 00:10:25.355 }' 00:10:25.355 23:44:13 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:25.355 23:44:13 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:10:25.921 23:44:13 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:10:25.921 23:44:13 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:25.921 23:44:13 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:10:25.921 [2024-11-26 23:44:13.783496] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:10:25.921 [2024-11-26 23:44:13.788587] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000d06220 00:10:25.921 23:44:13 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:25.921 23:44:13 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@647 -- # sleep 1 00:10:25.921 [2024-11-26 23:44:13.790596] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:10:26.855 23:44:14 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:10:26.855 23:44:14 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:10:26.855 23:44:14 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:10:26.855 23:44:14 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:10:26.855 23:44:14 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:10:26.855 23:44:14 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:26.855 23:44:14 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:26.855 23:44:14 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:26.855 23:44:14 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:10:26.855 23:44:14 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:26.855 23:44:14 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:10:26.855 "name": "raid_bdev1", 00:10:26.855 "uuid": "b039aaa7-5122-46ec-8ced-c6f237aa3d51", 00:10:26.855 "strip_size_kb": 0, 00:10:26.855 "state": "online", 00:10:26.855 "raid_level": "raid1", 00:10:26.855 "superblock": false, 00:10:26.855 "num_base_bdevs": 2, 00:10:26.855 "num_base_bdevs_discovered": 2, 00:10:26.855 "num_base_bdevs_operational": 2, 00:10:26.855 "process": { 00:10:26.855 "type": "rebuild", 00:10:26.855 "target": "spare", 00:10:26.855 "progress": { 00:10:26.855 "blocks": 20480, 00:10:26.855 "percent": 31 00:10:26.855 } 00:10:26.855 }, 00:10:26.855 "base_bdevs_list": [ 00:10:26.855 { 00:10:26.855 "name": "spare", 00:10:26.855 "uuid": "964cd690-f042-5333-b6c5-58781675893f", 00:10:26.855 "is_configured": true, 00:10:26.855 "data_offset": 0, 00:10:26.855 "data_size": 65536 00:10:26.855 }, 00:10:26.855 { 00:10:26.855 "name": "BaseBdev2", 00:10:26.855 "uuid": "d4bb21ce-e99c-5091-bc05-8ecd2175e0d4", 00:10:26.855 "is_configured": true, 00:10:26.855 "data_offset": 0, 00:10:26.855 "data_size": 65536 00:10:26.855 } 00:10:26.855 ] 00:10:26.855 }' 00:10:26.855 23:44:14 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:10:26.855 23:44:14 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:10:26.855 23:44:14 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:10:26.855 23:44:14 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:10:26.855 23:44:14 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:10:26.855 23:44:14 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:26.855 23:44:14 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:10:26.855 [2024-11-26 23:44:14.922915] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:10:27.114 [2024-11-26 23:44:14.996037] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:10:27.114 [2024-11-26 23:44:14.996107] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:27.114 [2024-11-26 23:44:14.996126] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:10:27.114 [2024-11-26 23:44:14.996141] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:10:27.114 23:44:15 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:27.114 23:44:15 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:10:27.114 23:44:15 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:27.114 23:44:15 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:27.114 23:44:15 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:27.114 23:44:15 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:27.114 23:44:15 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:10:27.114 23:44:15 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:27.114 23:44:15 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:27.114 23:44:15 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:27.114 23:44:15 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:27.114 23:44:15 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:27.114 23:44:15 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:27.114 23:44:15 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:27.114 23:44:15 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:10:27.114 23:44:15 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:27.114 23:44:15 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:27.114 "name": "raid_bdev1", 00:10:27.114 "uuid": "b039aaa7-5122-46ec-8ced-c6f237aa3d51", 00:10:27.114 "strip_size_kb": 0, 00:10:27.114 "state": "online", 00:10:27.114 "raid_level": "raid1", 00:10:27.114 "superblock": false, 00:10:27.114 "num_base_bdevs": 2, 00:10:27.114 "num_base_bdevs_discovered": 1, 00:10:27.114 "num_base_bdevs_operational": 1, 00:10:27.114 "base_bdevs_list": [ 00:10:27.114 { 00:10:27.114 "name": null, 00:10:27.114 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:27.114 "is_configured": false, 00:10:27.114 "data_offset": 0, 00:10:27.114 "data_size": 65536 00:10:27.114 }, 00:10:27.114 { 00:10:27.114 "name": "BaseBdev2", 00:10:27.114 "uuid": "d4bb21ce-e99c-5091-bc05-8ecd2175e0d4", 00:10:27.114 "is_configured": true, 00:10:27.114 "data_offset": 0, 00:10:27.114 "data_size": 65536 00:10:27.114 } 00:10:27.114 ] 00:10:27.114 }' 00:10:27.114 23:44:15 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:27.114 23:44:15 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:10:27.373 23:44:15 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:10:27.373 23:44:15 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:10:27.373 23:44:15 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:10:27.373 23:44:15 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=none 00:10:27.373 23:44:15 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:10:27.373 23:44:15 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:27.373 23:44:15 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:27.373 23:44:15 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:10:27.373 23:44:15 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:27.373 23:44:15 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:27.373 23:44:15 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:10:27.373 "name": "raid_bdev1", 00:10:27.373 "uuid": "b039aaa7-5122-46ec-8ced-c6f237aa3d51", 00:10:27.373 "strip_size_kb": 0, 00:10:27.373 "state": "online", 00:10:27.373 "raid_level": "raid1", 00:10:27.373 "superblock": false, 00:10:27.373 "num_base_bdevs": 2, 00:10:27.373 "num_base_bdevs_discovered": 1, 00:10:27.373 "num_base_bdevs_operational": 1, 00:10:27.373 "base_bdevs_list": [ 00:10:27.373 { 00:10:27.373 "name": null, 00:10:27.373 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:27.373 "is_configured": false, 00:10:27.373 "data_offset": 0, 00:10:27.373 "data_size": 65536 00:10:27.373 }, 00:10:27.373 { 00:10:27.373 "name": "BaseBdev2", 00:10:27.373 "uuid": "d4bb21ce-e99c-5091-bc05-8ecd2175e0d4", 00:10:27.373 "is_configured": true, 00:10:27.373 "data_offset": 0, 00:10:27.373 "data_size": 65536 00:10:27.373 } 00:10:27.373 ] 00:10:27.373 }' 00:10:27.373 23:44:15 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:10:27.630 23:44:15 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:10:27.630 23:44:15 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:10:27.630 23:44:15 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:10:27.630 23:44:15 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:10:27.630 23:44:15 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:27.630 23:44:15 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:10:27.630 [2024-11-26 23:44:15.580305] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:10:27.630 [2024-11-26 23:44:15.585509] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000d062f0 00:10:27.630 23:44:15 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:27.630 23:44:15 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@663 -- # sleep 1 00:10:27.630 [2024-11-26 23:44:15.587644] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:10:28.563 23:44:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:10:28.563 23:44:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:10:28.563 23:44:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:10:28.563 23:44:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:10:28.563 23:44:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:10:28.563 23:44:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:28.563 23:44:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:28.563 23:44:16 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:28.563 23:44:16 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:10:28.563 23:44:16 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:28.563 23:44:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:10:28.563 "name": "raid_bdev1", 00:10:28.563 "uuid": "b039aaa7-5122-46ec-8ced-c6f237aa3d51", 00:10:28.563 "strip_size_kb": 0, 00:10:28.563 "state": "online", 00:10:28.563 "raid_level": "raid1", 00:10:28.563 "superblock": false, 00:10:28.563 "num_base_bdevs": 2, 00:10:28.563 "num_base_bdevs_discovered": 2, 00:10:28.563 "num_base_bdevs_operational": 2, 00:10:28.563 "process": { 00:10:28.563 "type": "rebuild", 00:10:28.563 "target": "spare", 00:10:28.563 "progress": { 00:10:28.563 "blocks": 20480, 00:10:28.563 "percent": 31 00:10:28.563 } 00:10:28.563 }, 00:10:28.563 "base_bdevs_list": [ 00:10:28.563 { 00:10:28.563 "name": "spare", 00:10:28.563 "uuid": "964cd690-f042-5333-b6c5-58781675893f", 00:10:28.563 "is_configured": true, 00:10:28.563 "data_offset": 0, 00:10:28.563 "data_size": 65536 00:10:28.563 }, 00:10:28.563 { 00:10:28.563 "name": "BaseBdev2", 00:10:28.563 "uuid": "d4bb21ce-e99c-5091-bc05-8ecd2175e0d4", 00:10:28.563 "is_configured": true, 00:10:28.563 "data_offset": 0, 00:10:28.563 "data_size": 65536 00:10:28.563 } 00:10:28.563 ] 00:10:28.563 }' 00:10:28.563 23:44:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:10:28.563 23:44:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:10:28.563 23:44:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:10:28.822 23:44:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:10:28.822 23:44:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@666 -- # '[' false = true ']' 00:10:28.822 23:44:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=2 00:10:28.822 23:44:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:10:28.822 23:44:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@693 -- # '[' 2 -gt 2 ']' 00:10:28.822 23:44:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@706 -- # local timeout=288 00:10:28.822 23:44:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:10:28.822 23:44:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:10:28.822 23:44:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:10:28.822 23:44:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:10:28.822 23:44:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:10:28.822 23:44:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:10:28.822 23:44:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:28.822 23:44:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:28.822 23:44:16 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:28.822 23:44:16 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:10:28.822 23:44:16 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:28.822 23:44:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:10:28.822 "name": "raid_bdev1", 00:10:28.822 "uuid": "b039aaa7-5122-46ec-8ced-c6f237aa3d51", 00:10:28.822 "strip_size_kb": 0, 00:10:28.822 "state": "online", 00:10:28.822 "raid_level": "raid1", 00:10:28.822 "superblock": false, 00:10:28.822 "num_base_bdevs": 2, 00:10:28.822 "num_base_bdevs_discovered": 2, 00:10:28.822 "num_base_bdevs_operational": 2, 00:10:28.822 "process": { 00:10:28.822 "type": "rebuild", 00:10:28.822 "target": "spare", 00:10:28.822 "progress": { 00:10:28.822 "blocks": 22528, 00:10:28.822 "percent": 34 00:10:28.822 } 00:10:28.822 }, 00:10:28.822 "base_bdevs_list": [ 00:10:28.822 { 00:10:28.822 "name": "spare", 00:10:28.822 "uuid": "964cd690-f042-5333-b6c5-58781675893f", 00:10:28.822 "is_configured": true, 00:10:28.822 "data_offset": 0, 00:10:28.822 "data_size": 65536 00:10:28.822 }, 00:10:28.822 { 00:10:28.822 "name": "BaseBdev2", 00:10:28.822 "uuid": "d4bb21ce-e99c-5091-bc05-8ecd2175e0d4", 00:10:28.822 "is_configured": true, 00:10:28.822 "data_offset": 0, 00:10:28.822 "data_size": 65536 00:10:28.822 } 00:10:28.822 ] 00:10:28.822 }' 00:10:28.822 23:44:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:10:28.822 23:44:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:10:28.822 23:44:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:10:28.822 23:44:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:10:28.822 23:44:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:10:29.756 23:44:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:10:29.756 23:44:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:10:29.756 23:44:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:10:29.756 23:44:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:10:29.756 23:44:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:10:29.756 23:44:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:10:29.756 23:44:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:29.756 23:44:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:29.756 23:44:17 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:29.756 23:44:17 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:10:29.756 23:44:17 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:30.015 23:44:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:10:30.015 "name": "raid_bdev1", 00:10:30.015 "uuid": "b039aaa7-5122-46ec-8ced-c6f237aa3d51", 00:10:30.015 "strip_size_kb": 0, 00:10:30.015 "state": "online", 00:10:30.015 "raid_level": "raid1", 00:10:30.015 "superblock": false, 00:10:30.015 "num_base_bdevs": 2, 00:10:30.015 "num_base_bdevs_discovered": 2, 00:10:30.015 "num_base_bdevs_operational": 2, 00:10:30.015 "process": { 00:10:30.015 "type": "rebuild", 00:10:30.015 "target": "spare", 00:10:30.015 "progress": { 00:10:30.015 "blocks": 45056, 00:10:30.015 "percent": 68 00:10:30.015 } 00:10:30.015 }, 00:10:30.015 "base_bdevs_list": [ 00:10:30.015 { 00:10:30.015 "name": "spare", 00:10:30.015 "uuid": "964cd690-f042-5333-b6c5-58781675893f", 00:10:30.015 "is_configured": true, 00:10:30.015 "data_offset": 0, 00:10:30.015 "data_size": 65536 00:10:30.015 }, 00:10:30.015 { 00:10:30.015 "name": "BaseBdev2", 00:10:30.015 "uuid": "d4bb21ce-e99c-5091-bc05-8ecd2175e0d4", 00:10:30.015 "is_configured": true, 00:10:30.015 "data_offset": 0, 00:10:30.015 "data_size": 65536 00:10:30.015 } 00:10:30.015 ] 00:10:30.015 }' 00:10:30.015 23:44:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:10:30.015 23:44:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:10:30.015 23:44:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:10:30.015 23:44:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:10:30.015 23:44:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:10:30.950 [2024-11-26 23:44:18.800989] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:10:30.950 [2024-11-26 23:44:18.801226] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:10:30.950 [2024-11-26 23:44:18.801302] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:30.950 23:44:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:10:30.950 23:44:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:10:30.950 23:44:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:10:30.950 23:44:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:10:30.950 23:44:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:10:30.950 23:44:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:10:30.950 23:44:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:30.950 23:44:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:30.950 23:44:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:30.950 23:44:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:10:30.950 23:44:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:30.950 23:44:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:10:30.950 "name": "raid_bdev1", 00:10:30.950 "uuid": "b039aaa7-5122-46ec-8ced-c6f237aa3d51", 00:10:30.950 "strip_size_kb": 0, 00:10:30.950 "state": "online", 00:10:30.950 "raid_level": "raid1", 00:10:30.950 "superblock": false, 00:10:30.950 "num_base_bdevs": 2, 00:10:30.950 "num_base_bdevs_discovered": 2, 00:10:30.950 "num_base_bdevs_operational": 2, 00:10:30.950 "base_bdevs_list": [ 00:10:30.950 { 00:10:30.950 "name": "spare", 00:10:30.950 "uuid": "964cd690-f042-5333-b6c5-58781675893f", 00:10:30.950 "is_configured": true, 00:10:30.950 "data_offset": 0, 00:10:30.950 "data_size": 65536 00:10:30.950 }, 00:10:30.950 { 00:10:30.950 "name": "BaseBdev2", 00:10:30.950 "uuid": "d4bb21ce-e99c-5091-bc05-8ecd2175e0d4", 00:10:30.950 "is_configured": true, 00:10:30.950 "data_offset": 0, 00:10:30.950 "data_size": 65536 00:10:30.950 } 00:10:30.950 ] 00:10:30.950 }' 00:10:30.950 23:44:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:10:30.950 23:44:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:10:30.950 23:44:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:10:31.207 23:44:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:10:31.207 23:44:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@709 -- # break 00:10:31.207 23:44:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:10:31.207 23:44:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:10:31.207 23:44:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:10:31.207 23:44:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=none 00:10:31.207 23:44:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:10:31.207 23:44:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:31.207 23:44:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:31.207 23:44:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:31.207 23:44:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:10:31.207 23:44:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:31.207 23:44:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:10:31.207 "name": "raid_bdev1", 00:10:31.207 "uuid": "b039aaa7-5122-46ec-8ced-c6f237aa3d51", 00:10:31.207 "strip_size_kb": 0, 00:10:31.207 "state": "online", 00:10:31.207 "raid_level": "raid1", 00:10:31.207 "superblock": false, 00:10:31.207 "num_base_bdevs": 2, 00:10:31.207 "num_base_bdevs_discovered": 2, 00:10:31.207 "num_base_bdevs_operational": 2, 00:10:31.207 "base_bdevs_list": [ 00:10:31.207 { 00:10:31.207 "name": "spare", 00:10:31.207 "uuid": "964cd690-f042-5333-b6c5-58781675893f", 00:10:31.207 "is_configured": true, 00:10:31.207 "data_offset": 0, 00:10:31.207 "data_size": 65536 00:10:31.207 }, 00:10:31.207 { 00:10:31.207 "name": "BaseBdev2", 00:10:31.207 "uuid": "d4bb21ce-e99c-5091-bc05-8ecd2175e0d4", 00:10:31.207 "is_configured": true, 00:10:31.207 "data_offset": 0, 00:10:31.207 "data_size": 65536 00:10:31.207 } 00:10:31.207 ] 00:10:31.207 }' 00:10:31.207 23:44:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:10:31.207 23:44:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:10:31.207 23:44:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:10:31.207 23:44:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:10:31.207 23:44:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:10:31.207 23:44:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:31.207 23:44:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:31.207 23:44:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:31.207 23:44:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:31.207 23:44:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:10:31.207 23:44:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:31.207 23:44:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:31.207 23:44:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:31.207 23:44:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:31.207 23:44:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:31.207 23:44:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:31.207 23:44:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:10:31.207 23:44:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:31.207 23:44:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:31.207 23:44:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:31.207 "name": "raid_bdev1", 00:10:31.207 "uuid": "b039aaa7-5122-46ec-8ced-c6f237aa3d51", 00:10:31.207 "strip_size_kb": 0, 00:10:31.207 "state": "online", 00:10:31.207 "raid_level": "raid1", 00:10:31.207 "superblock": false, 00:10:31.207 "num_base_bdevs": 2, 00:10:31.207 "num_base_bdevs_discovered": 2, 00:10:31.207 "num_base_bdevs_operational": 2, 00:10:31.207 "base_bdevs_list": [ 00:10:31.207 { 00:10:31.207 "name": "spare", 00:10:31.207 "uuid": "964cd690-f042-5333-b6c5-58781675893f", 00:10:31.207 "is_configured": true, 00:10:31.207 "data_offset": 0, 00:10:31.207 "data_size": 65536 00:10:31.207 }, 00:10:31.207 { 00:10:31.207 "name": "BaseBdev2", 00:10:31.207 "uuid": "d4bb21ce-e99c-5091-bc05-8ecd2175e0d4", 00:10:31.207 "is_configured": true, 00:10:31.207 "data_offset": 0, 00:10:31.207 "data_size": 65536 00:10:31.207 } 00:10:31.207 ] 00:10:31.207 }' 00:10:31.207 23:44:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:31.207 23:44:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:10:31.771 23:44:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:10:31.771 23:44:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:31.771 23:44:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:10:31.771 [2024-11-26 23:44:19.636743] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:10:31.771 [2024-11-26 23:44:19.636833] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:31.771 [2024-11-26 23:44:19.636976] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:31.771 [2024-11-26 23:44:19.637080] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:31.771 [2024-11-26 23:44:19.637136] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:10:31.771 23:44:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:31.771 23:44:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:31.771 23:44:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@720 -- # jq length 00:10:31.771 23:44:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:31.771 23:44:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:10:31.771 23:44:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:31.771 23:44:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:10:31.771 23:44:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:10:31.771 23:44:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:10:31.771 23:44:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:10:31.771 23:44:19 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:10:31.771 23:44:19 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:10:31.771 23:44:19 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@10 -- # local bdev_list 00:10:31.771 23:44:19 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:10:31.771 23:44:19 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@11 -- # local nbd_list 00:10:31.771 23:44:19 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@12 -- # local i 00:10:31.772 23:44:19 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:10:31.772 23:44:19 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:10:31.772 23:44:19 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:10:31.772 /dev/nbd0 00:10:32.030 23:44:19 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:10:32.030 23:44:19 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:10:32.030 23:44:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:10:32.030 23:44:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@873 -- # local i 00:10:32.030 23:44:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:10:32.030 23:44:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:10:32.030 23:44:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:10:32.030 23:44:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@877 -- # break 00:10:32.030 23:44:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:10:32.030 23:44:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:10:32.030 23:44:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:10:32.030 1+0 records in 00:10:32.030 1+0 records out 00:10:32.030 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000303905 s, 13.5 MB/s 00:10:32.030 23:44:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:10:32.030 23:44:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@890 -- # size=4096 00:10:32.030 23:44:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:10:32.030 23:44:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:10:32.030 23:44:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@893 -- # return 0 00:10:32.030 23:44:19 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:10:32.030 23:44:19 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:10:32.030 23:44:19 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:10:32.030 /dev/nbd1 00:10:32.289 23:44:20 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:10:32.289 23:44:20 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:10:32.289 23:44:20 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:10:32.289 23:44:20 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@873 -- # local i 00:10:32.289 23:44:20 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:10:32.289 23:44:20 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:10:32.289 23:44:20 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:10:32.289 23:44:20 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@877 -- # break 00:10:32.289 23:44:20 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:10:32.289 23:44:20 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:10:32.289 23:44:20 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:10:32.289 1+0 records in 00:10:32.289 1+0 records out 00:10:32.289 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000372534 s, 11.0 MB/s 00:10:32.289 23:44:20 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:10:32.289 23:44:20 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@890 -- # size=4096 00:10:32.289 23:44:20 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:10:32.289 23:44:20 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:10:32.289 23:44:20 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@893 -- # return 0 00:10:32.289 23:44:20 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:10:32.289 23:44:20 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:10:32.289 23:44:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@738 -- # cmp -i 0 /dev/nbd0 /dev/nbd1 00:10:32.289 23:44:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:10:32.289 23:44:20 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:10:32.289 23:44:20 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:10:32.289 23:44:20 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@50 -- # local nbd_list 00:10:32.289 23:44:20 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@51 -- # local i 00:10:32.289 23:44:20 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:10:32.289 23:44:20 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:10:32.548 23:44:20 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:10:32.548 23:44:20 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:10:32.548 23:44:20 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:10:32.548 23:44:20 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:10:32.548 23:44:20 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:10:32.548 23:44:20 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:10:32.548 23:44:20 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:10:32.548 23:44:20 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:10:32.548 23:44:20 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:10:32.548 23:44:20 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:10:32.806 23:44:20 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:10:32.806 23:44:20 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:10:32.806 23:44:20 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:10:32.806 23:44:20 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:10:32.806 23:44:20 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:10:32.806 23:44:20 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:10:32.806 23:44:20 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:10:32.806 23:44:20 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:10:32.806 23:44:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@743 -- # '[' false = true ']' 00:10:32.806 23:44:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@784 -- # killprocess 85728 00:10:32.806 23:44:20 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@954 -- # '[' -z 85728 ']' 00:10:32.806 23:44:20 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@958 -- # kill -0 85728 00:10:32.806 23:44:20 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@959 -- # uname 00:10:32.806 23:44:20 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:10:32.806 23:44:20 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 85728 00:10:32.806 killing process with pid 85728 00:10:32.806 Received shutdown signal, test time was about 60.000000 seconds 00:10:32.806 00:10:32.806 Latency(us) 00:10:32.806 [2024-11-26T23:44:20.938Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:10:32.806 [2024-11-26T23:44:20.938Z] =================================================================================================================== 00:10:32.806 [2024-11-26T23:44:20.938Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:10:32.806 23:44:20 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:10:32.806 23:44:20 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:10:32.806 23:44:20 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 85728' 00:10:32.806 23:44:20 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@973 -- # kill 85728 00:10:32.806 [2024-11-26 23:44:20.729243] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:10:32.806 23:44:20 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@978 -- # wait 85728 00:10:32.806 [2024-11-26 23:44:20.760038] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:10:33.070 23:44:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@786 -- # return 0 00:10:33.070 ************************************ 00:10:33.070 END TEST raid_rebuild_test 00:10:33.070 ************************************ 00:10:33.070 00:10:33.070 real 0m13.467s 00:10:33.070 user 0m15.571s 00:10:33.070 sys 0m2.805s 00:10:33.070 23:44:20 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:10:33.070 23:44:20 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:10:33.070 23:44:21 bdev_raid -- bdev/bdev_raid.sh@979 -- # run_test raid_rebuild_test_sb raid_rebuild_test raid1 2 true false true 00:10:33.070 23:44:21 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:10:33.070 23:44:21 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:10:33.070 23:44:21 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:10:33.070 ************************************ 00:10:33.070 START TEST raid_rebuild_test_sb 00:10:33.070 ************************************ 00:10:33.070 23:44:21 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@1129 -- # raid_rebuild_test raid1 2 true false true 00:10:33.070 23:44:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:10:33.070 23:44:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=2 00:10:33.070 23:44:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@571 -- # local superblock=true 00:10:33.070 23:44:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:10:33.070 23:44:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@573 -- # local verify=true 00:10:33.070 23:44:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:10:33.070 23:44:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:10:33.070 23:44:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:10:33.070 23:44:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:10:33.070 23:44:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:10:33.070 23:44:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:10:33.070 23:44:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:10:33.070 23:44:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:10:33.071 23:44:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:10:33.071 23:44:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:10:33.071 23:44:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:10:33.071 23:44:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # local strip_size 00:10:33.071 23:44:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@577 -- # local create_arg 00:10:33.071 23:44:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:10:33.071 23:44:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@579 -- # local data_offset 00:10:33.071 23:44:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:10:33.071 23:44:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:10:33.071 23:44:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@592 -- # '[' true = true ']' 00:10:33.071 23:44:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@593 -- # create_arg+=' -s' 00:10:33.071 23:44:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@597 -- # raid_pid=86130 00:10:33.071 23:44:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:10:33.071 23:44:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@598 -- # waitforlisten 86130 00:10:33.071 23:44:21 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@835 -- # '[' -z 86130 ']' 00:10:33.071 23:44:21 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:33.071 23:44:21 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@840 -- # local max_retries=100 00:10:33.071 23:44:21 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:33.071 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:33.071 23:44:21 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@844 -- # xtrace_disable 00:10:33.071 23:44:21 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:33.071 [2024-11-26 23:44:21.124570] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:10:33.071 I/O size of 3145728 is greater than zero copy threshold (65536). 00:10:33.071 Zero copy mechanism will not be used. 00:10:33.071 [2024-11-26 23:44:21.124778] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid86130 ] 00:10:33.341 [2024-11-26 23:44:21.256982] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:10:33.341 [2024-11-26 23:44:21.285508] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:10:33.341 [2024-11-26 23:44:21.327901] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:33.341 [2024-11-26 23:44:21.327937] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:33.919 23:44:21 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:10:33.919 23:44:21 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@868 -- # return 0 00:10:33.919 23:44:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:10:33.919 23:44:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:10:33.919 23:44:21 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:33.919 23:44:21 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:33.919 BaseBdev1_malloc 00:10:33.919 23:44:21 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:33.919 23:44:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:10:33.919 23:44:21 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:33.919 23:44:21 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:33.919 [2024-11-26 23:44:21.971324] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:10:33.919 [2024-11-26 23:44:21.971411] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:33.919 [2024-11-26 23:44:21.971442] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:10:33.919 [2024-11-26 23:44:21.971460] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:33.919 [2024-11-26 23:44:21.973504] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:33.919 [2024-11-26 23:44:21.973540] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:10:33.919 BaseBdev1 00:10:33.919 23:44:21 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:33.919 23:44:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:10:33.919 23:44:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:10:33.919 23:44:21 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:33.919 23:44:21 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:33.919 BaseBdev2_malloc 00:10:33.919 23:44:21 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:33.919 23:44:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:10:33.920 23:44:21 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:33.920 23:44:21 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:33.920 [2024-11-26 23:44:21.999905] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:10:33.920 [2024-11-26 23:44:21.999961] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:33.920 [2024-11-26 23:44:21.999985] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:10:33.920 [2024-11-26 23:44:21.999994] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:33.920 [2024-11-26 23:44:22.002083] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:33.920 [2024-11-26 23:44:22.002122] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:10:33.920 BaseBdev2 00:10:33.920 23:44:22 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:33.920 23:44:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:10:33.920 23:44:22 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:33.920 23:44:22 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:33.920 spare_malloc 00:10:33.920 23:44:22 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:33.920 23:44:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:10:33.920 23:44:22 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:33.920 23:44:22 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:33.920 spare_delay 00:10:33.920 23:44:22 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:33.920 23:44:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:10:33.920 23:44:22 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:33.920 23:44:22 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:33.920 [2024-11-26 23:44:22.040245] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:10:33.920 [2024-11-26 23:44:22.040292] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:33.920 [2024-11-26 23:44:22.040310] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:10:33.920 [2024-11-26 23:44:22.040318] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:33.920 [2024-11-26 23:44:22.042380] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:33.920 [2024-11-26 23:44:22.042470] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:10:33.920 spare 00:10:33.920 23:44:22 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:33.920 23:44:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 00:10:33.920 23:44:22 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:33.920 23:44:22 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:34.179 [2024-11-26 23:44:22.052257] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:34.179 [2024-11-26 23:44:22.054057] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:34.179 [2024-11-26 23:44:22.054260] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:10:34.179 [2024-11-26 23:44:22.054276] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:10:34.179 [2024-11-26 23:44:22.054557] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:10:34.179 [2024-11-26 23:44:22.054697] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:10:34.179 [2024-11-26 23:44:22.054709] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:10:34.179 [2024-11-26 23:44:22.054824] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:34.179 23:44:22 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:34.179 23:44:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:10:34.179 23:44:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:34.179 23:44:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:34.179 23:44:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:34.179 23:44:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:34.179 23:44:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:10:34.179 23:44:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:34.179 23:44:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:34.179 23:44:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:34.179 23:44:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:34.179 23:44:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:34.179 23:44:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:34.179 23:44:22 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:34.179 23:44:22 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:34.179 23:44:22 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:34.179 23:44:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:34.179 "name": "raid_bdev1", 00:10:34.179 "uuid": "85fdf7dd-78f4-4f44-8be2-71ba94b80be3", 00:10:34.179 "strip_size_kb": 0, 00:10:34.179 "state": "online", 00:10:34.179 "raid_level": "raid1", 00:10:34.179 "superblock": true, 00:10:34.179 "num_base_bdevs": 2, 00:10:34.179 "num_base_bdevs_discovered": 2, 00:10:34.179 "num_base_bdevs_operational": 2, 00:10:34.179 "base_bdevs_list": [ 00:10:34.179 { 00:10:34.179 "name": "BaseBdev1", 00:10:34.179 "uuid": "d34d063c-e87e-52e5-8878-0b51a1339763", 00:10:34.179 "is_configured": true, 00:10:34.179 "data_offset": 2048, 00:10:34.179 "data_size": 63488 00:10:34.179 }, 00:10:34.179 { 00:10:34.179 "name": "BaseBdev2", 00:10:34.179 "uuid": "e1b7af0b-bd27-54d0-a26f-955520fc14d9", 00:10:34.179 "is_configured": true, 00:10:34.179 "data_offset": 2048, 00:10:34.179 "data_size": 63488 00:10:34.179 } 00:10:34.179 ] 00:10:34.179 }' 00:10:34.179 23:44:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:34.179 23:44:22 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:34.439 23:44:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:10:34.439 23:44:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:10:34.439 23:44:22 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:34.439 23:44:22 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:34.439 [2024-11-26 23:44:22.495756] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:34.439 23:44:22 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:34.439 23:44:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=63488 00:10:34.439 23:44:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:10:34.439 23:44:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:34.439 23:44:22 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:34.439 23:44:22 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:34.439 23:44:22 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:34.439 23:44:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # data_offset=2048 00:10:34.439 23:44:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:10:34.439 23:44:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:10:34.440 23:44:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:10:34.440 23:44:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:10:34.440 23:44:22 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:10:34.440 23:44:22 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:10:34.440 23:44:22 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # local bdev_list 00:10:34.440 23:44:22 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:10:34.440 23:44:22 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # local nbd_list 00:10:34.440 23:44:22 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@12 -- # local i 00:10:34.440 23:44:22 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:10:34.440 23:44:22 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:10:34.440 23:44:22 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:10:34.698 [2024-11-26 23:44:22.743141] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:10:34.698 /dev/nbd0 00:10:34.698 23:44:22 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:10:34.698 23:44:22 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:10:34.698 23:44:22 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:10:34.698 23:44:22 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@873 -- # local i 00:10:34.698 23:44:22 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:10:34.698 23:44:22 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:10:34.698 23:44:22 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:10:34.698 23:44:22 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@877 -- # break 00:10:34.698 23:44:22 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:10:34.698 23:44:22 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:10:34.698 23:44:22 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:10:34.698 1+0 records in 00:10:34.698 1+0 records out 00:10:34.698 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.00039709 s, 10.3 MB/s 00:10:34.698 23:44:22 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:10:34.698 23:44:22 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@890 -- # size=4096 00:10:34.698 23:44:22 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:10:34.698 23:44:22 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:10:34.698 23:44:22 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@893 -- # return 0 00:10:34.699 23:44:22 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:10:34.699 23:44:22 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:10:34.699 23:44:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@629 -- # '[' raid1 = raid5f ']' 00:10:34.699 23:44:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@633 -- # write_unit_size=1 00:10:34.699 23:44:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=512 count=63488 oflag=direct 00:10:38.907 63488+0 records in 00:10:38.907 63488+0 records out 00:10:38.907 32505856 bytes (33 MB, 31 MiB) copied, 3.74363 s, 8.7 MB/s 00:10:38.908 23:44:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:10:38.908 23:44:26 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:10:38.908 23:44:26 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:10:38.908 23:44:26 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # local nbd_list 00:10:38.908 23:44:26 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@51 -- # local i 00:10:38.908 23:44:26 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:10:38.908 23:44:26 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:10:38.908 23:44:26 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:10:38.908 [2024-11-26 23:44:26.772067] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:38.908 23:44:26 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:10:38.908 23:44:26 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:10:38.908 23:44:26 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:10:38.908 23:44:26 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:10:38.908 23:44:26 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:10:38.908 23:44:26 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:10:38.908 23:44:26 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:10:38.908 23:44:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:10:38.908 23:44:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:38.908 23:44:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:38.908 [2024-11-26 23:44:26.785081] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:10:38.908 23:44:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:38.908 23:44:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:10:38.908 23:44:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:38.908 23:44:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:38.908 23:44:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:38.908 23:44:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:38.908 23:44:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:10:38.908 23:44:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:38.908 23:44:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:38.908 23:44:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:38.908 23:44:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:38.908 23:44:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:38.908 23:44:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:38.908 23:44:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:38.908 23:44:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:38.908 23:44:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:38.908 23:44:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:38.908 "name": "raid_bdev1", 00:10:38.908 "uuid": "85fdf7dd-78f4-4f44-8be2-71ba94b80be3", 00:10:38.908 "strip_size_kb": 0, 00:10:38.908 "state": "online", 00:10:38.908 "raid_level": "raid1", 00:10:38.908 "superblock": true, 00:10:38.908 "num_base_bdevs": 2, 00:10:38.908 "num_base_bdevs_discovered": 1, 00:10:38.908 "num_base_bdevs_operational": 1, 00:10:38.908 "base_bdevs_list": [ 00:10:38.908 { 00:10:38.908 "name": null, 00:10:38.908 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:38.908 "is_configured": false, 00:10:38.908 "data_offset": 0, 00:10:38.908 "data_size": 63488 00:10:38.908 }, 00:10:38.908 { 00:10:38.908 "name": "BaseBdev2", 00:10:38.908 "uuid": "e1b7af0b-bd27-54d0-a26f-955520fc14d9", 00:10:38.908 "is_configured": true, 00:10:38.908 "data_offset": 2048, 00:10:38.908 "data_size": 63488 00:10:38.908 } 00:10:38.908 ] 00:10:38.908 }' 00:10:38.908 23:44:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:38.908 23:44:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:39.165 23:44:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:10:39.165 23:44:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:39.165 23:44:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:39.165 [2024-11-26 23:44:27.228330] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:10:39.165 [2024-11-26 23:44:27.233482] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000c3e280 00:10:39.165 23:44:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:39.165 23:44:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@647 -- # sleep 1 00:10:39.165 [2024-11-26 23:44:27.235459] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:10:40.535 23:44:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:10:40.535 23:44:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:10:40.535 23:44:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:10:40.535 23:44:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:10:40.535 23:44:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:10:40.535 23:44:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:40.535 23:44:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:40.535 23:44:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:40.535 23:44:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:40.535 23:44:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:40.535 23:44:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:10:40.535 "name": "raid_bdev1", 00:10:40.535 "uuid": "85fdf7dd-78f4-4f44-8be2-71ba94b80be3", 00:10:40.535 "strip_size_kb": 0, 00:10:40.535 "state": "online", 00:10:40.535 "raid_level": "raid1", 00:10:40.535 "superblock": true, 00:10:40.535 "num_base_bdevs": 2, 00:10:40.535 "num_base_bdevs_discovered": 2, 00:10:40.535 "num_base_bdevs_operational": 2, 00:10:40.535 "process": { 00:10:40.535 "type": "rebuild", 00:10:40.535 "target": "spare", 00:10:40.535 "progress": { 00:10:40.535 "blocks": 20480, 00:10:40.535 "percent": 32 00:10:40.535 } 00:10:40.535 }, 00:10:40.535 "base_bdevs_list": [ 00:10:40.535 { 00:10:40.535 "name": "spare", 00:10:40.535 "uuid": "7291dbb5-2d9e-54f2-bdad-c0959ea154f5", 00:10:40.535 "is_configured": true, 00:10:40.535 "data_offset": 2048, 00:10:40.535 "data_size": 63488 00:10:40.535 }, 00:10:40.535 { 00:10:40.535 "name": "BaseBdev2", 00:10:40.535 "uuid": "e1b7af0b-bd27-54d0-a26f-955520fc14d9", 00:10:40.535 "is_configured": true, 00:10:40.535 "data_offset": 2048, 00:10:40.535 "data_size": 63488 00:10:40.535 } 00:10:40.535 ] 00:10:40.535 }' 00:10:40.535 23:44:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:10:40.535 23:44:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:10:40.535 23:44:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:10:40.535 23:44:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:10:40.535 23:44:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:10:40.535 23:44:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:40.535 23:44:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:40.535 [2024-11-26 23:44:28.371755] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:10:40.535 [2024-11-26 23:44:28.440755] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:10:40.535 [2024-11-26 23:44:28.440823] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:40.535 [2024-11-26 23:44:28.440842] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:10:40.535 [2024-11-26 23:44:28.440850] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:10:40.535 23:44:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:40.535 23:44:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:10:40.535 23:44:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:40.535 23:44:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:40.535 23:44:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:40.535 23:44:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:40.535 23:44:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:10:40.535 23:44:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:40.535 23:44:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:40.535 23:44:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:40.535 23:44:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:40.535 23:44:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:40.535 23:44:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:40.535 23:44:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:40.535 23:44:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:40.535 23:44:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:40.535 23:44:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:40.535 "name": "raid_bdev1", 00:10:40.535 "uuid": "85fdf7dd-78f4-4f44-8be2-71ba94b80be3", 00:10:40.535 "strip_size_kb": 0, 00:10:40.535 "state": "online", 00:10:40.535 "raid_level": "raid1", 00:10:40.535 "superblock": true, 00:10:40.535 "num_base_bdevs": 2, 00:10:40.535 "num_base_bdevs_discovered": 1, 00:10:40.535 "num_base_bdevs_operational": 1, 00:10:40.535 "base_bdevs_list": [ 00:10:40.535 { 00:10:40.535 "name": null, 00:10:40.535 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:40.535 "is_configured": false, 00:10:40.535 "data_offset": 0, 00:10:40.535 "data_size": 63488 00:10:40.535 }, 00:10:40.535 { 00:10:40.535 "name": "BaseBdev2", 00:10:40.535 "uuid": "e1b7af0b-bd27-54d0-a26f-955520fc14d9", 00:10:40.535 "is_configured": true, 00:10:40.535 "data_offset": 2048, 00:10:40.535 "data_size": 63488 00:10:40.535 } 00:10:40.535 ] 00:10:40.535 }' 00:10:40.535 23:44:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:40.535 23:44:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:40.792 23:44:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:10:40.792 23:44:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:10:40.792 23:44:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:10:40.792 23:44:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:10:40.792 23:44:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:10:41.050 23:44:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:41.050 23:44:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:41.050 23:44:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:41.050 23:44:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:41.050 23:44:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:41.050 23:44:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:10:41.050 "name": "raid_bdev1", 00:10:41.050 "uuid": "85fdf7dd-78f4-4f44-8be2-71ba94b80be3", 00:10:41.050 "strip_size_kb": 0, 00:10:41.050 "state": "online", 00:10:41.050 "raid_level": "raid1", 00:10:41.050 "superblock": true, 00:10:41.050 "num_base_bdevs": 2, 00:10:41.050 "num_base_bdevs_discovered": 1, 00:10:41.050 "num_base_bdevs_operational": 1, 00:10:41.050 "base_bdevs_list": [ 00:10:41.050 { 00:10:41.050 "name": null, 00:10:41.050 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:41.050 "is_configured": false, 00:10:41.050 "data_offset": 0, 00:10:41.050 "data_size": 63488 00:10:41.050 }, 00:10:41.050 { 00:10:41.050 "name": "BaseBdev2", 00:10:41.050 "uuid": "e1b7af0b-bd27-54d0-a26f-955520fc14d9", 00:10:41.050 "is_configured": true, 00:10:41.050 "data_offset": 2048, 00:10:41.051 "data_size": 63488 00:10:41.051 } 00:10:41.051 ] 00:10:41.051 }' 00:10:41.051 23:44:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:10:41.051 23:44:29 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:10:41.051 23:44:29 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:10:41.051 23:44:29 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:10:41.051 23:44:29 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:10:41.051 23:44:29 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:41.051 23:44:29 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:41.051 [2024-11-26 23:44:29.052942] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:10:41.051 [2024-11-26 23:44:29.057997] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000c3e350 00:10:41.051 23:44:29 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:41.051 23:44:29 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@663 -- # sleep 1 00:10:41.051 [2024-11-26 23:44:29.059946] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:10:41.981 23:44:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:10:41.981 23:44:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:10:41.981 23:44:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:10:41.981 23:44:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:10:41.981 23:44:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:10:41.981 23:44:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:41.981 23:44:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:41.981 23:44:30 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:41.981 23:44:30 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:41.981 23:44:30 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:42.238 23:44:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:10:42.238 "name": "raid_bdev1", 00:10:42.238 "uuid": "85fdf7dd-78f4-4f44-8be2-71ba94b80be3", 00:10:42.238 "strip_size_kb": 0, 00:10:42.238 "state": "online", 00:10:42.238 "raid_level": "raid1", 00:10:42.238 "superblock": true, 00:10:42.238 "num_base_bdevs": 2, 00:10:42.238 "num_base_bdevs_discovered": 2, 00:10:42.238 "num_base_bdevs_operational": 2, 00:10:42.238 "process": { 00:10:42.238 "type": "rebuild", 00:10:42.238 "target": "spare", 00:10:42.238 "progress": { 00:10:42.238 "blocks": 20480, 00:10:42.238 "percent": 32 00:10:42.238 } 00:10:42.238 }, 00:10:42.238 "base_bdevs_list": [ 00:10:42.238 { 00:10:42.238 "name": "spare", 00:10:42.238 "uuid": "7291dbb5-2d9e-54f2-bdad-c0959ea154f5", 00:10:42.238 "is_configured": true, 00:10:42.238 "data_offset": 2048, 00:10:42.238 "data_size": 63488 00:10:42.238 }, 00:10:42.238 { 00:10:42.238 "name": "BaseBdev2", 00:10:42.238 "uuid": "e1b7af0b-bd27-54d0-a26f-955520fc14d9", 00:10:42.238 "is_configured": true, 00:10:42.238 "data_offset": 2048, 00:10:42.238 "data_size": 63488 00:10:42.238 } 00:10:42.238 ] 00:10:42.238 }' 00:10:42.238 23:44:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:10:42.238 23:44:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:10:42.238 23:44:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:10:42.238 23:44:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:10:42.238 23:44:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@666 -- # '[' true = true ']' 00:10:42.239 23:44:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@666 -- # '[' = false ']' 00:10:42.239 /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh: line 666: [: =: unary operator expected 00:10:42.239 23:44:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=2 00:10:42.239 23:44:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:10:42.239 23:44:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@693 -- # '[' 2 -gt 2 ']' 00:10:42.239 23:44:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@706 -- # local timeout=302 00:10:42.239 23:44:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:10:42.239 23:44:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:10:42.239 23:44:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:10:42.239 23:44:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:10:42.239 23:44:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:10:42.239 23:44:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:10:42.239 23:44:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:42.239 23:44:30 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:42.239 23:44:30 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:42.239 23:44:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:42.239 23:44:30 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:42.239 23:44:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:10:42.239 "name": "raid_bdev1", 00:10:42.239 "uuid": "85fdf7dd-78f4-4f44-8be2-71ba94b80be3", 00:10:42.239 "strip_size_kb": 0, 00:10:42.239 "state": "online", 00:10:42.239 "raid_level": "raid1", 00:10:42.239 "superblock": true, 00:10:42.239 "num_base_bdevs": 2, 00:10:42.239 "num_base_bdevs_discovered": 2, 00:10:42.239 "num_base_bdevs_operational": 2, 00:10:42.239 "process": { 00:10:42.239 "type": "rebuild", 00:10:42.239 "target": "spare", 00:10:42.239 "progress": { 00:10:42.239 "blocks": 22528, 00:10:42.239 "percent": 35 00:10:42.239 } 00:10:42.239 }, 00:10:42.239 "base_bdevs_list": [ 00:10:42.239 { 00:10:42.239 "name": "spare", 00:10:42.239 "uuid": "7291dbb5-2d9e-54f2-bdad-c0959ea154f5", 00:10:42.239 "is_configured": true, 00:10:42.239 "data_offset": 2048, 00:10:42.239 "data_size": 63488 00:10:42.239 }, 00:10:42.239 { 00:10:42.239 "name": "BaseBdev2", 00:10:42.239 "uuid": "e1b7af0b-bd27-54d0-a26f-955520fc14d9", 00:10:42.239 "is_configured": true, 00:10:42.239 "data_offset": 2048, 00:10:42.239 "data_size": 63488 00:10:42.239 } 00:10:42.239 ] 00:10:42.239 }' 00:10:42.239 23:44:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:10:42.239 23:44:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:10:42.239 23:44:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:10:42.239 23:44:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:10:42.239 23:44:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:10:43.610 23:44:31 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:10:43.610 23:44:31 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:10:43.610 23:44:31 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:10:43.610 23:44:31 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:10:43.610 23:44:31 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:10:43.611 23:44:31 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:10:43.611 23:44:31 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:43.611 23:44:31 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:43.611 23:44:31 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:43.611 23:44:31 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:43.611 23:44:31 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:43.611 23:44:31 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:10:43.611 "name": "raid_bdev1", 00:10:43.611 "uuid": "85fdf7dd-78f4-4f44-8be2-71ba94b80be3", 00:10:43.611 "strip_size_kb": 0, 00:10:43.611 "state": "online", 00:10:43.611 "raid_level": "raid1", 00:10:43.611 "superblock": true, 00:10:43.611 "num_base_bdevs": 2, 00:10:43.611 "num_base_bdevs_discovered": 2, 00:10:43.611 "num_base_bdevs_operational": 2, 00:10:43.611 "process": { 00:10:43.611 "type": "rebuild", 00:10:43.611 "target": "spare", 00:10:43.611 "progress": { 00:10:43.611 "blocks": 45056, 00:10:43.611 "percent": 70 00:10:43.611 } 00:10:43.611 }, 00:10:43.611 "base_bdevs_list": [ 00:10:43.611 { 00:10:43.611 "name": "spare", 00:10:43.611 "uuid": "7291dbb5-2d9e-54f2-bdad-c0959ea154f5", 00:10:43.611 "is_configured": true, 00:10:43.611 "data_offset": 2048, 00:10:43.611 "data_size": 63488 00:10:43.611 }, 00:10:43.611 { 00:10:43.611 "name": "BaseBdev2", 00:10:43.611 "uuid": "e1b7af0b-bd27-54d0-a26f-955520fc14d9", 00:10:43.611 "is_configured": true, 00:10:43.611 "data_offset": 2048, 00:10:43.611 "data_size": 63488 00:10:43.611 } 00:10:43.611 ] 00:10:43.611 }' 00:10:43.611 23:44:31 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:10:43.611 23:44:31 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:10:43.611 23:44:31 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:10:43.611 23:44:31 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:10:43.611 23:44:31 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:10:44.175 [2024-11-26 23:44:32.172460] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:10:44.175 [2024-11-26 23:44:32.172662] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:10:44.175 [2024-11-26 23:44:32.172798] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:44.433 23:44:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:10:44.433 23:44:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:10:44.433 23:44:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:10:44.433 23:44:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:10:44.433 23:44:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:10:44.433 23:44:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:10:44.433 23:44:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:44.433 23:44:32 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:44.433 23:44:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:44.433 23:44:32 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:44.433 23:44:32 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:44.731 23:44:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:10:44.731 "name": "raid_bdev1", 00:10:44.731 "uuid": "85fdf7dd-78f4-4f44-8be2-71ba94b80be3", 00:10:44.731 "strip_size_kb": 0, 00:10:44.731 "state": "online", 00:10:44.731 "raid_level": "raid1", 00:10:44.731 "superblock": true, 00:10:44.731 "num_base_bdevs": 2, 00:10:44.731 "num_base_bdevs_discovered": 2, 00:10:44.731 "num_base_bdevs_operational": 2, 00:10:44.731 "base_bdevs_list": [ 00:10:44.731 { 00:10:44.731 "name": "spare", 00:10:44.731 "uuid": "7291dbb5-2d9e-54f2-bdad-c0959ea154f5", 00:10:44.731 "is_configured": true, 00:10:44.731 "data_offset": 2048, 00:10:44.731 "data_size": 63488 00:10:44.731 }, 00:10:44.731 { 00:10:44.731 "name": "BaseBdev2", 00:10:44.731 "uuid": "e1b7af0b-bd27-54d0-a26f-955520fc14d9", 00:10:44.731 "is_configured": true, 00:10:44.731 "data_offset": 2048, 00:10:44.731 "data_size": 63488 00:10:44.731 } 00:10:44.731 ] 00:10:44.731 }' 00:10:44.731 23:44:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:10:44.731 23:44:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:10:44.731 23:44:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:10:44.731 23:44:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:10:44.731 23:44:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@709 -- # break 00:10:44.731 23:44:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:10:44.731 23:44:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:10:44.731 23:44:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:10:44.731 23:44:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:10:44.731 23:44:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:10:44.731 23:44:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:44.731 23:44:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:44.731 23:44:32 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:44.731 23:44:32 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:44.731 23:44:32 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:44.731 23:44:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:10:44.731 "name": "raid_bdev1", 00:10:44.731 "uuid": "85fdf7dd-78f4-4f44-8be2-71ba94b80be3", 00:10:44.731 "strip_size_kb": 0, 00:10:44.731 "state": "online", 00:10:44.731 "raid_level": "raid1", 00:10:44.731 "superblock": true, 00:10:44.731 "num_base_bdevs": 2, 00:10:44.731 "num_base_bdevs_discovered": 2, 00:10:44.731 "num_base_bdevs_operational": 2, 00:10:44.731 "base_bdevs_list": [ 00:10:44.731 { 00:10:44.731 "name": "spare", 00:10:44.731 "uuid": "7291dbb5-2d9e-54f2-bdad-c0959ea154f5", 00:10:44.731 "is_configured": true, 00:10:44.731 "data_offset": 2048, 00:10:44.731 "data_size": 63488 00:10:44.731 }, 00:10:44.731 { 00:10:44.731 "name": "BaseBdev2", 00:10:44.731 "uuid": "e1b7af0b-bd27-54d0-a26f-955520fc14d9", 00:10:44.731 "is_configured": true, 00:10:44.731 "data_offset": 2048, 00:10:44.731 "data_size": 63488 00:10:44.731 } 00:10:44.731 ] 00:10:44.731 }' 00:10:44.731 23:44:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:10:44.731 23:44:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:10:44.731 23:44:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:10:44.731 23:44:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:10:44.731 23:44:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:10:44.731 23:44:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:44.731 23:44:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:44.731 23:44:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:44.731 23:44:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:44.731 23:44:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:10:44.731 23:44:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:44.731 23:44:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:44.731 23:44:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:44.731 23:44:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:44.731 23:44:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:44.731 23:44:32 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:44.731 23:44:32 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:44.731 23:44:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:44.731 23:44:32 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:44.731 23:44:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:44.731 "name": "raid_bdev1", 00:10:44.731 "uuid": "85fdf7dd-78f4-4f44-8be2-71ba94b80be3", 00:10:44.732 "strip_size_kb": 0, 00:10:44.732 "state": "online", 00:10:44.732 "raid_level": "raid1", 00:10:44.732 "superblock": true, 00:10:44.732 "num_base_bdevs": 2, 00:10:44.732 "num_base_bdevs_discovered": 2, 00:10:44.732 "num_base_bdevs_operational": 2, 00:10:44.732 "base_bdevs_list": [ 00:10:44.732 { 00:10:44.732 "name": "spare", 00:10:44.732 "uuid": "7291dbb5-2d9e-54f2-bdad-c0959ea154f5", 00:10:44.732 "is_configured": true, 00:10:44.732 "data_offset": 2048, 00:10:44.732 "data_size": 63488 00:10:44.732 }, 00:10:44.732 { 00:10:44.732 "name": "BaseBdev2", 00:10:44.732 "uuid": "e1b7af0b-bd27-54d0-a26f-955520fc14d9", 00:10:44.732 "is_configured": true, 00:10:44.732 "data_offset": 2048, 00:10:44.732 "data_size": 63488 00:10:44.732 } 00:10:44.732 ] 00:10:44.732 }' 00:10:44.732 23:44:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:44.732 23:44:32 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:45.298 23:44:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:10:45.298 23:44:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:45.298 23:44:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:45.298 [2024-11-26 23:44:33.207966] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:10:45.298 [2024-11-26 23:44:33.208001] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:45.298 [2024-11-26 23:44:33.208096] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:45.298 [2024-11-26 23:44:33.208159] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:45.298 [2024-11-26 23:44:33.208171] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:10:45.298 23:44:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:45.298 23:44:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:45.298 23:44:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # jq length 00:10:45.298 23:44:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:45.298 23:44:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:45.298 23:44:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:45.298 23:44:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:10:45.298 23:44:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:10:45.298 23:44:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:10:45.298 23:44:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:10:45.298 23:44:33 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:10:45.298 23:44:33 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:10:45.298 23:44:33 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # local bdev_list 00:10:45.298 23:44:33 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:10:45.298 23:44:33 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # local nbd_list 00:10:45.298 23:44:33 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@12 -- # local i 00:10:45.298 23:44:33 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:10:45.298 23:44:33 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:10:45.298 23:44:33 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:10:45.557 /dev/nbd0 00:10:45.557 23:44:33 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:10:45.557 23:44:33 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:10:45.557 23:44:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:10:45.557 23:44:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@873 -- # local i 00:10:45.557 23:44:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:10:45.557 23:44:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:10:45.557 23:44:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:10:45.557 23:44:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@877 -- # break 00:10:45.557 23:44:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:10:45.557 23:44:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:10:45.557 23:44:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:10:45.557 1+0 records in 00:10:45.557 1+0 records out 00:10:45.557 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000222979 s, 18.4 MB/s 00:10:45.557 23:44:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:10:45.557 23:44:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@890 -- # size=4096 00:10:45.557 23:44:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:10:45.557 23:44:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:10:45.557 23:44:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@893 -- # return 0 00:10:45.557 23:44:33 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:10:45.557 23:44:33 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:10:45.557 23:44:33 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:10:45.816 /dev/nbd1 00:10:45.816 23:44:33 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:10:45.816 23:44:33 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:10:45.816 23:44:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:10:45.816 23:44:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@873 -- # local i 00:10:45.816 23:44:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:10:45.816 23:44:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:10:45.816 23:44:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:10:45.816 23:44:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@877 -- # break 00:10:45.816 23:44:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:10:45.816 23:44:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:10:45.816 23:44:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:10:45.816 1+0 records in 00:10:45.816 1+0 records out 00:10:45.816 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000403753 s, 10.1 MB/s 00:10:45.816 23:44:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:10:45.816 23:44:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@890 -- # size=4096 00:10:45.816 23:44:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:10:45.816 23:44:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:10:45.816 23:44:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@893 -- # return 0 00:10:45.816 23:44:33 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:10:45.816 23:44:33 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:10:45.816 23:44:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@738 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:10:45.816 23:44:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:10:45.816 23:44:33 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:10:45.816 23:44:33 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:10:45.816 23:44:33 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # local nbd_list 00:10:45.816 23:44:33 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@51 -- # local i 00:10:45.816 23:44:33 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:10:45.816 23:44:33 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:10:46.075 23:44:34 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:10:46.075 23:44:34 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:10:46.075 23:44:34 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:10:46.075 23:44:34 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:10:46.075 23:44:34 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:10:46.075 23:44:34 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:10:46.075 23:44:34 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:10:46.075 23:44:34 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:10:46.075 23:44:34 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:10:46.075 23:44:34 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:10:46.075 23:44:34 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:10:46.334 23:44:34 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:10:46.334 23:44:34 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:10:46.334 23:44:34 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:10:46.334 23:44:34 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:10:46.334 23:44:34 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:10:46.334 23:44:34 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:10:46.334 23:44:34 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:10:46.334 23:44:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@743 -- # '[' true = true ']' 00:10:46.334 23:44:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@745 -- # rpc_cmd bdev_passthru_delete spare 00:10:46.334 23:44:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:46.334 23:44:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:46.334 23:44:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:46.334 23:44:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@746 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:10:46.334 23:44:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:46.334 23:44:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:46.334 [2024-11-26 23:44:34.229871] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:10:46.334 [2024-11-26 23:44:34.229931] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:46.334 [2024-11-26 23:44:34.229970] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:10:46.334 [2024-11-26 23:44:34.229983] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:46.334 [2024-11-26 23:44:34.232103] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:46.334 [2024-11-26 23:44:34.232189] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:10:46.334 [2024-11-26 23:44:34.232276] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:10:46.334 [2024-11-26 23:44:34.232323] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:10:46.334 [2024-11-26 23:44:34.232449] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:46.334 spare 00:10:46.334 23:44:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:46.334 23:44:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@747 -- # rpc_cmd bdev_wait_for_examine 00:10:46.334 23:44:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:46.334 23:44:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:46.334 [2024-11-26 23:44:34.332338] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001580 00:10:46.334 [2024-11-26 23:44:34.332366] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:10:46.334 [2024-11-26 23:44:34.332597] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000cae960 00:10:46.334 [2024-11-26 23:44:34.332741] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001580 00:10:46.334 [2024-11-26 23:44:34.332758] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001580 00:10:46.334 [2024-11-26 23:44:34.332876] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:46.334 23:44:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:46.334 23:44:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@749 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:10:46.334 23:44:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:46.334 23:44:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:46.334 23:44:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:46.334 23:44:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:46.334 23:44:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:10:46.334 23:44:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:46.334 23:44:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:46.334 23:44:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:46.334 23:44:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:46.334 23:44:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:46.334 23:44:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:46.334 23:44:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:46.334 23:44:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:46.334 23:44:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:46.334 23:44:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:46.334 "name": "raid_bdev1", 00:10:46.334 "uuid": "85fdf7dd-78f4-4f44-8be2-71ba94b80be3", 00:10:46.334 "strip_size_kb": 0, 00:10:46.334 "state": "online", 00:10:46.334 "raid_level": "raid1", 00:10:46.334 "superblock": true, 00:10:46.334 "num_base_bdevs": 2, 00:10:46.334 "num_base_bdevs_discovered": 2, 00:10:46.334 "num_base_bdevs_operational": 2, 00:10:46.334 "base_bdevs_list": [ 00:10:46.334 { 00:10:46.334 "name": "spare", 00:10:46.334 "uuid": "7291dbb5-2d9e-54f2-bdad-c0959ea154f5", 00:10:46.334 "is_configured": true, 00:10:46.334 "data_offset": 2048, 00:10:46.334 "data_size": 63488 00:10:46.334 }, 00:10:46.334 { 00:10:46.335 "name": "BaseBdev2", 00:10:46.335 "uuid": "e1b7af0b-bd27-54d0-a26f-955520fc14d9", 00:10:46.335 "is_configured": true, 00:10:46.335 "data_offset": 2048, 00:10:46.335 "data_size": 63488 00:10:46.335 } 00:10:46.335 ] 00:10:46.335 }' 00:10:46.335 23:44:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:46.335 23:44:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:46.902 23:44:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@750 -- # verify_raid_bdev_process raid_bdev1 none none 00:10:46.902 23:44:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:10:46.902 23:44:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:10:46.902 23:44:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:10:46.902 23:44:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:10:46.902 23:44:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:46.902 23:44:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:46.902 23:44:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:46.902 23:44:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:46.902 23:44:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:46.902 23:44:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:10:46.902 "name": "raid_bdev1", 00:10:46.902 "uuid": "85fdf7dd-78f4-4f44-8be2-71ba94b80be3", 00:10:46.902 "strip_size_kb": 0, 00:10:46.902 "state": "online", 00:10:46.902 "raid_level": "raid1", 00:10:46.902 "superblock": true, 00:10:46.902 "num_base_bdevs": 2, 00:10:46.902 "num_base_bdevs_discovered": 2, 00:10:46.902 "num_base_bdevs_operational": 2, 00:10:46.902 "base_bdevs_list": [ 00:10:46.902 { 00:10:46.902 "name": "spare", 00:10:46.902 "uuid": "7291dbb5-2d9e-54f2-bdad-c0959ea154f5", 00:10:46.902 "is_configured": true, 00:10:46.902 "data_offset": 2048, 00:10:46.902 "data_size": 63488 00:10:46.902 }, 00:10:46.902 { 00:10:46.902 "name": "BaseBdev2", 00:10:46.902 "uuid": "e1b7af0b-bd27-54d0-a26f-955520fc14d9", 00:10:46.902 "is_configured": true, 00:10:46.902 "data_offset": 2048, 00:10:46.902 "data_size": 63488 00:10:46.902 } 00:10:46.902 ] 00:10:46.902 }' 00:10:46.902 23:44:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:10:46.902 23:44:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:10:46.902 23:44:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:10:46.902 23:44:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:10:46.902 23:44:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # jq -r '.[].base_bdevs_list[0].name' 00:10:46.902 23:44:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:46.902 23:44:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:46.902 23:44:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:46.902 23:44:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:46.902 23:44:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # [[ spare == \s\p\a\r\e ]] 00:10:46.902 23:44:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@754 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:10:46.902 23:44:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:46.902 23:44:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:46.902 [2024-11-26 23:44:35.000625] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:10:46.902 23:44:35 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:46.902 23:44:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@755 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:10:46.902 23:44:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:46.902 23:44:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:46.902 23:44:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:46.902 23:44:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:46.902 23:44:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:10:46.902 23:44:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:46.902 23:44:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:46.902 23:44:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:46.902 23:44:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:46.902 23:44:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:46.902 23:44:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:46.902 23:44:35 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:46.902 23:44:35 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:46.902 23:44:35 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:47.256 23:44:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:47.256 "name": "raid_bdev1", 00:10:47.256 "uuid": "85fdf7dd-78f4-4f44-8be2-71ba94b80be3", 00:10:47.256 "strip_size_kb": 0, 00:10:47.256 "state": "online", 00:10:47.256 "raid_level": "raid1", 00:10:47.256 "superblock": true, 00:10:47.256 "num_base_bdevs": 2, 00:10:47.256 "num_base_bdevs_discovered": 1, 00:10:47.256 "num_base_bdevs_operational": 1, 00:10:47.256 "base_bdevs_list": [ 00:10:47.256 { 00:10:47.256 "name": null, 00:10:47.256 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:47.256 "is_configured": false, 00:10:47.256 "data_offset": 0, 00:10:47.256 "data_size": 63488 00:10:47.256 }, 00:10:47.256 { 00:10:47.256 "name": "BaseBdev2", 00:10:47.256 "uuid": "e1b7af0b-bd27-54d0-a26f-955520fc14d9", 00:10:47.256 "is_configured": true, 00:10:47.256 "data_offset": 2048, 00:10:47.256 "data_size": 63488 00:10:47.256 } 00:10:47.256 ] 00:10:47.256 }' 00:10:47.256 23:44:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:47.256 23:44:35 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:47.538 23:44:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@756 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:10:47.538 23:44:35 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:47.538 23:44:35 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:47.538 [2024-11-26 23:44:35.372022] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:10:47.538 [2024-11-26 23:44:35.372265] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:10:47.538 [2024-11-26 23:44:35.372324] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:10:47.538 [2024-11-26 23:44:35.372451] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:10:47.538 [2024-11-26 23:44:35.377202] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000caea30 00:10:47.538 23:44:35 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:47.538 23:44:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@757 -- # sleep 1 00:10:47.538 [2024-11-26 23:44:35.379094] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:10:48.476 23:44:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@758 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:10:48.476 23:44:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:10:48.476 23:44:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:10:48.476 23:44:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:10:48.476 23:44:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:10:48.476 23:44:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:48.476 23:44:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:48.476 23:44:36 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:48.476 23:44:36 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:48.476 23:44:36 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:48.476 23:44:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:10:48.476 "name": "raid_bdev1", 00:10:48.476 "uuid": "85fdf7dd-78f4-4f44-8be2-71ba94b80be3", 00:10:48.476 "strip_size_kb": 0, 00:10:48.476 "state": "online", 00:10:48.476 "raid_level": "raid1", 00:10:48.476 "superblock": true, 00:10:48.476 "num_base_bdevs": 2, 00:10:48.476 "num_base_bdevs_discovered": 2, 00:10:48.476 "num_base_bdevs_operational": 2, 00:10:48.476 "process": { 00:10:48.476 "type": "rebuild", 00:10:48.476 "target": "spare", 00:10:48.476 "progress": { 00:10:48.476 "blocks": 20480, 00:10:48.476 "percent": 32 00:10:48.476 } 00:10:48.476 }, 00:10:48.476 "base_bdevs_list": [ 00:10:48.476 { 00:10:48.476 "name": "spare", 00:10:48.476 "uuid": "7291dbb5-2d9e-54f2-bdad-c0959ea154f5", 00:10:48.476 "is_configured": true, 00:10:48.476 "data_offset": 2048, 00:10:48.476 "data_size": 63488 00:10:48.476 }, 00:10:48.476 { 00:10:48.476 "name": "BaseBdev2", 00:10:48.476 "uuid": "e1b7af0b-bd27-54d0-a26f-955520fc14d9", 00:10:48.476 "is_configured": true, 00:10:48.476 "data_offset": 2048, 00:10:48.476 "data_size": 63488 00:10:48.476 } 00:10:48.476 ] 00:10:48.476 }' 00:10:48.476 23:44:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:10:48.476 23:44:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:10:48.476 23:44:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:10:48.476 23:44:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:10:48.476 23:44:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@761 -- # rpc_cmd bdev_passthru_delete spare 00:10:48.476 23:44:36 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:48.476 23:44:36 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:48.476 [2024-11-26 23:44:36.539147] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:10:48.476 [2024-11-26 23:44:36.583013] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:10:48.476 [2024-11-26 23:44:36.583117] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:48.476 [2024-11-26 23:44:36.583159] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:10:48.476 [2024-11-26 23:44:36.583180] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:10:48.476 23:44:36 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:48.476 23:44:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@762 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:10:48.476 23:44:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:48.476 23:44:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:48.476 23:44:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:48.476 23:44:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:48.476 23:44:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:10:48.476 23:44:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:48.476 23:44:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:48.476 23:44:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:48.476 23:44:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:48.476 23:44:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:48.476 23:44:36 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:48.476 23:44:36 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:48.476 23:44:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:48.736 23:44:36 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:48.736 23:44:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:48.736 "name": "raid_bdev1", 00:10:48.736 "uuid": "85fdf7dd-78f4-4f44-8be2-71ba94b80be3", 00:10:48.736 "strip_size_kb": 0, 00:10:48.736 "state": "online", 00:10:48.736 "raid_level": "raid1", 00:10:48.736 "superblock": true, 00:10:48.736 "num_base_bdevs": 2, 00:10:48.736 "num_base_bdevs_discovered": 1, 00:10:48.736 "num_base_bdevs_operational": 1, 00:10:48.736 "base_bdevs_list": [ 00:10:48.736 { 00:10:48.736 "name": null, 00:10:48.736 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:48.736 "is_configured": false, 00:10:48.736 "data_offset": 0, 00:10:48.736 "data_size": 63488 00:10:48.736 }, 00:10:48.736 { 00:10:48.736 "name": "BaseBdev2", 00:10:48.736 "uuid": "e1b7af0b-bd27-54d0-a26f-955520fc14d9", 00:10:48.736 "is_configured": true, 00:10:48.736 "data_offset": 2048, 00:10:48.736 "data_size": 63488 00:10:48.736 } 00:10:48.736 ] 00:10:48.736 }' 00:10:48.736 23:44:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:48.736 23:44:36 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:48.996 23:44:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@763 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:10:48.996 23:44:37 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:48.996 23:44:37 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:48.996 [2024-11-26 23:44:37.043095] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:10:48.996 [2024-11-26 23:44:37.043166] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:48.996 [2024-11-26 23:44:37.043191] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009f80 00:10:48.996 [2024-11-26 23:44:37.043200] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:48.996 [2024-11-26 23:44:37.043624] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:48.996 [2024-11-26 23:44:37.043648] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:10:48.996 [2024-11-26 23:44:37.043751] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:10:48.996 [2024-11-26 23:44:37.043764] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:10:48.996 [2024-11-26 23:44:37.043789] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:10:48.996 [2024-11-26 23:44:37.043810] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:10:48.996 spare 00:10:48.996 [2024-11-26 23:44:37.048520] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000caeb00 00:10:48.996 23:44:37 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:48.996 23:44:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@764 -- # sleep 1 00:10:48.996 [2024-11-26 23:44:37.050378] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:10:49.931 23:44:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@765 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:10:49.931 23:44:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:10:49.931 23:44:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:10:49.931 23:44:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:10:49.931 23:44:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:10:50.191 23:44:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:50.191 23:44:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:50.191 23:44:38 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:50.191 23:44:38 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:50.191 23:44:38 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:50.191 23:44:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:10:50.191 "name": "raid_bdev1", 00:10:50.191 "uuid": "85fdf7dd-78f4-4f44-8be2-71ba94b80be3", 00:10:50.191 "strip_size_kb": 0, 00:10:50.191 "state": "online", 00:10:50.191 "raid_level": "raid1", 00:10:50.191 "superblock": true, 00:10:50.191 "num_base_bdevs": 2, 00:10:50.191 "num_base_bdevs_discovered": 2, 00:10:50.191 "num_base_bdevs_operational": 2, 00:10:50.191 "process": { 00:10:50.191 "type": "rebuild", 00:10:50.191 "target": "spare", 00:10:50.191 "progress": { 00:10:50.191 "blocks": 20480, 00:10:50.191 "percent": 32 00:10:50.191 } 00:10:50.191 }, 00:10:50.191 "base_bdevs_list": [ 00:10:50.191 { 00:10:50.191 "name": "spare", 00:10:50.191 "uuid": "7291dbb5-2d9e-54f2-bdad-c0959ea154f5", 00:10:50.191 "is_configured": true, 00:10:50.191 "data_offset": 2048, 00:10:50.191 "data_size": 63488 00:10:50.191 }, 00:10:50.191 { 00:10:50.191 "name": "BaseBdev2", 00:10:50.191 "uuid": "e1b7af0b-bd27-54d0-a26f-955520fc14d9", 00:10:50.191 "is_configured": true, 00:10:50.191 "data_offset": 2048, 00:10:50.191 "data_size": 63488 00:10:50.191 } 00:10:50.191 ] 00:10:50.191 }' 00:10:50.191 23:44:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:10:50.191 23:44:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:10:50.191 23:44:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:10:50.191 23:44:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:10:50.191 23:44:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@768 -- # rpc_cmd bdev_passthru_delete spare 00:10:50.191 23:44:38 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:50.191 23:44:38 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:50.191 [2024-11-26 23:44:38.194541] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:10:50.191 [2024-11-26 23:44:38.254372] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:10:50.191 [2024-11-26 23:44:38.254496] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:50.191 [2024-11-26 23:44:38.254529] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:10:50.191 [2024-11-26 23:44:38.254552] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:10:50.191 23:44:38 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:50.191 23:44:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@769 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:10:50.191 23:44:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:50.191 23:44:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:50.191 23:44:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:50.191 23:44:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:50.191 23:44:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:10:50.191 23:44:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:50.191 23:44:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:50.191 23:44:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:50.191 23:44:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:50.191 23:44:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:50.191 23:44:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:50.191 23:44:38 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:50.191 23:44:38 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:50.191 23:44:38 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:50.191 23:44:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:50.191 "name": "raid_bdev1", 00:10:50.191 "uuid": "85fdf7dd-78f4-4f44-8be2-71ba94b80be3", 00:10:50.191 "strip_size_kb": 0, 00:10:50.191 "state": "online", 00:10:50.191 "raid_level": "raid1", 00:10:50.191 "superblock": true, 00:10:50.191 "num_base_bdevs": 2, 00:10:50.191 "num_base_bdevs_discovered": 1, 00:10:50.191 "num_base_bdevs_operational": 1, 00:10:50.191 "base_bdevs_list": [ 00:10:50.191 { 00:10:50.191 "name": null, 00:10:50.191 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:50.191 "is_configured": false, 00:10:50.191 "data_offset": 0, 00:10:50.191 "data_size": 63488 00:10:50.191 }, 00:10:50.191 { 00:10:50.191 "name": "BaseBdev2", 00:10:50.191 "uuid": "e1b7af0b-bd27-54d0-a26f-955520fc14d9", 00:10:50.191 "is_configured": true, 00:10:50.191 "data_offset": 2048, 00:10:50.191 "data_size": 63488 00:10:50.191 } 00:10:50.191 ] 00:10:50.191 }' 00:10:50.191 23:44:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:50.191 23:44:38 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:50.776 23:44:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@770 -- # verify_raid_bdev_process raid_bdev1 none none 00:10:50.776 23:44:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:10:50.776 23:44:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:10:50.776 23:44:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:10:50.776 23:44:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:10:50.776 23:44:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:50.776 23:44:38 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:50.776 23:44:38 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:50.776 23:44:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:50.776 23:44:38 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:50.776 23:44:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:10:50.776 "name": "raid_bdev1", 00:10:50.776 "uuid": "85fdf7dd-78f4-4f44-8be2-71ba94b80be3", 00:10:50.776 "strip_size_kb": 0, 00:10:50.776 "state": "online", 00:10:50.776 "raid_level": "raid1", 00:10:50.776 "superblock": true, 00:10:50.776 "num_base_bdevs": 2, 00:10:50.776 "num_base_bdevs_discovered": 1, 00:10:50.776 "num_base_bdevs_operational": 1, 00:10:50.776 "base_bdevs_list": [ 00:10:50.776 { 00:10:50.776 "name": null, 00:10:50.776 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:50.776 "is_configured": false, 00:10:50.776 "data_offset": 0, 00:10:50.776 "data_size": 63488 00:10:50.776 }, 00:10:50.776 { 00:10:50.776 "name": "BaseBdev2", 00:10:50.776 "uuid": "e1b7af0b-bd27-54d0-a26f-955520fc14d9", 00:10:50.776 "is_configured": true, 00:10:50.776 "data_offset": 2048, 00:10:50.776 "data_size": 63488 00:10:50.776 } 00:10:50.776 ] 00:10:50.776 }' 00:10:50.776 23:44:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:10:50.776 23:44:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:10:50.776 23:44:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:10:50.776 23:44:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:10:50.776 23:44:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@773 -- # rpc_cmd bdev_passthru_delete BaseBdev1 00:10:50.776 23:44:38 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:50.776 23:44:38 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:50.776 23:44:38 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:50.776 23:44:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@774 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:10:50.776 23:44:38 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:50.776 23:44:38 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:50.776 [2024-11-26 23:44:38.878289] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:10:50.776 [2024-11-26 23:44:38.878405] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:50.776 [2024-11-26 23:44:38.878456] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a580 00:10:50.776 [2024-11-26 23:44:38.878489] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:50.776 [2024-11-26 23:44:38.878886] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:50.776 [2024-11-26 23:44:38.878946] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:10:50.776 [2024-11-26 23:44:38.879041] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:10:50.776 [2024-11-26 23:44:38.879085] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:10:50.777 [2024-11-26 23:44:38.879134] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:10:50.777 [2024-11-26 23:44:38.879196] bdev_raid.c:3894:raid_bdev_examine_done: *ERROR*: Failed to examine bdev BaseBdev1: Invalid argument 00:10:50.777 BaseBdev1 00:10:50.777 23:44:38 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:50.777 23:44:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@775 -- # sleep 1 00:10:52.151 23:44:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@776 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:10:52.151 23:44:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:52.151 23:44:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:52.151 23:44:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:52.151 23:44:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:52.151 23:44:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:10:52.151 23:44:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:52.151 23:44:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:52.151 23:44:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:52.151 23:44:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:52.151 23:44:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:52.151 23:44:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:52.151 23:44:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:52.151 23:44:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:52.151 23:44:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:52.151 23:44:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:52.151 "name": "raid_bdev1", 00:10:52.151 "uuid": "85fdf7dd-78f4-4f44-8be2-71ba94b80be3", 00:10:52.151 "strip_size_kb": 0, 00:10:52.151 "state": "online", 00:10:52.151 "raid_level": "raid1", 00:10:52.151 "superblock": true, 00:10:52.151 "num_base_bdevs": 2, 00:10:52.151 "num_base_bdevs_discovered": 1, 00:10:52.151 "num_base_bdevs_operational": 1, 00:10:52.151 "base_bdevs_list": [ 00:10:52.151 { 00:10:52.151 "name": null, 00:10:52.151 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:52.151 "is_configured": false, 00:10:52.151 "data_offset": 0, 00:10:52.151 "data_size": 63488 00:10:52.151 }, 00:10:52.151 { 00:10:52.151 "name": "BaseBdev2", 00:10:52.151 "uuid": "e1b7af0b-bd27-54d0-a26f-955520fc14d9", 00:10:52.151 "is_configured": true, 00:10:52.151 "data_offset": 2048, 00:10:52.151 "data_size": 63488 00:10:52.151 } 00:10:52.151 ] 00:10:52.151 }' 00:10:52.151 23:44:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:52.151 23:44:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:52.412 23:44:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@777 -- # verify_raid_bdev_process raid_bdev1 none none 00:10:52.412 23:44:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:10:52.412 23:44:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:10:52.412 23:44:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:10:52.412 23:44:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:10:52.412 23:44:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:52.412 23:44:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:52.412 23:44:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:52.412 23:44:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:52.412 23:44:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:52.412 23:44:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:10:52.412 "name": "raid_bdev1", 00:10:52.412 "uuid": "85fdf7dd-78f4-4f44-8be2-71ba94b80be3", 00:10:52.412 "strip_size_kb": 0, 00:10:52.412 "state": "online", 00:10:52.412 "raid_level": "raid1", 00:10:52.412 "superblock": true, 00:10:52.412 "num_base_bdevs": 2, 00:10:52.412 "num_base_bdevs_discovered": 1, 00:10:52.412 "num_base_bdevs_operational": 1, 00:10:52.412 "base_bdevs_list": [ 00:10:52.412 { 00:10:52.412 "name": null, 00:10:52.412 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:52.412 "is_configured": false, 00:10:52.412 "data_offset": 0, 00:10:52.412 "data_size": 63488 00:10:52.412 }, 00:10:52.412 { 00:10:52.412 "name": "BaseBdev2", 00:10:52.412 "uuid": "e1b7af0b-bd27-54d0-a26f-955520fc14d9", 00:10:52.412 "is_configured": true, 00:10:52.412 "data_offset": 2048, 00:10:52.412 "data_size": 63488 00:10:52.412 } 00:10:52.412 ] 00:10:52.412 }' 00:10:52.412 23:44:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:10:52.412 23:44:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:10:52.412 23:44:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:10:52.412 23:44:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:10:52.412 23:44:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@778 -- # NOT rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:10:52.412 23:44:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@652 -- # local es=0 00:10:52.412 23:44:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:10:52.412 23:44:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:10:52.412 23:44:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:10:52.412 23:44:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:10:52.412 23:44:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:10:52.412 23:44:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:10:52.412 23:44:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:52.412 23:44:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:52.412 [2024-11-26 23:44:40.507691] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:52.412 [2024-11-26 23:44:40.507910] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:10:52.412 [2024-11-26 23:44:40.507977] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:10:52.412 request: 00:10:52.412 { 00:10:52.412 "base_bdev": "BaseBdev1", 00:10:52.412 "raid_bdev": "raid_bdev1", 00:10:52.412 "method": "bdev_raid_add_base_bdev", 00:10:52.412 "req_id": 1 00:10:52.412 } 00:10:52.412 Got JSON-RPC error response 00:10:52.412 response: 00:10:52.412 { 00:10:52.412 "code": -22, 00:10:52.412 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:10:52.412 } 00:10:52.412 23:44:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:10:52.412 23:44:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@655 -- # es=1 00:10:52.412 23:44:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:10:52.412 23:44:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:10:52.412 23:44:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:10:52.412 23:44:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@779 -- # sleep 1 00:10:53.802 23:44:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@780 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:10:53.802 23:44:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:53.802 23:44:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:53.802 23:44:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:53.802 23:44:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:53.802 23:44:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:10:53.802 23:44:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:53.802 23:44:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:53.802 23:44:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:53.802 23:44:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:53.802 23:44:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:53.802 23:44:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:53.802 23:44:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:53.802 23:44:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:53.802 23:44:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:53.802 23:44:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:53.802 "name": "raid_bdev1", 00:10:53.802 "uuid": "85fdf7dd-78f4-4f44-8be2-71ba94b80be3", 00:10:53.802 "strip_size_kb": 0, 00:10:53.802 "state": "online", 00:10:53.802 "raid_level": "raid1", 00:10:53.802 "superblock": true, 00:10:53.803 "num_base_bdevs": 2, 00:10:53.803 "num_base_bdevs_discovered": 1, 00:10:53.803 "num_base_bdevs_operational": 1, 00:10:53.803 "base_bdevs_list": [ 00:10:53.803 { 00:10:53.803 "name": null, 00:10:53.803 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:53.803 "is_configured": false, 00:10:53.803 "data_offset": 0, 00:10:53.803 "data_size": 63488 00:10:53.803 }, 00:10:53.803 { 00:10:53.803 "name": "BaseBdev2", 00:10:53.803 "uuid": "e1b7af0b-bd27-54d0-a26f-955520fc14d9", 00:10:53.803 "is_configured": true, 00:10:53.803 "data_offset": 2048, 00:10:53.803 "data_size": 63488 00:10:53.803 } 00:10:53.803 ] 00:10:53.803 }' 00:10:53.803 23:44:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:53.803 23:44:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:54.074 23:44:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@781 -- # verify_raid_bdev_process raid_bdev1 none none 00:10:54.074 23:44:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:10:54.074 23:44:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:10:54.074 23:44:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:10:54.074 23:44:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:10:54.074 23:44:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:54.074 23:44:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:54.074 23:44:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:54.074 23:44:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:54.074 23:44:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:54.074 23:44:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:10:54.074 "name": "raid_bdev1", 00:10:54.074 "uuid": "85fdf7dd-78f4-4f44-8be2-71ba94b80be3", 00:10:54.074 "strip_size_kb": 0, 00:10:54.074 "state": "online", 00:10:54.074 "raid_level": "raid1", 00:10:54.074 "superblock": true, 00:10:54.074 "num_base_bdevs": 2, 00:10:54.074 "num_base_bdevs_discovered": 1, 00:10:54.074 "num_base_bdevs_operational": 1, 00:10:54.074 "base_bdevs_list": [ 00:10:54.074 { 00:10:54.074 "name": null, 00:10:54.074 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:54.074 "is_configured": false, 00:10:54.074 "data_offset": 0, 00:10:54.074 "data_size": 63488 00:10:54.074 }, 00:10:54.074 { 00:10:54.074 "name": "BaseBdev2", 00:10:54.074 "uuid": "e1b7af0b-bd27-54d0-a26f-955520fc14d9", 00:10:54.074 "is_configured": true, 00:10:54.074 "data_offset": 2048, 00:10:54.074 "data_size": 63488 00:10:54.074 } 00:10:54.074 ] 00:10:54.074 }' 00:10:54.074 23:44:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:10:54.074 23:44:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:10:54.074 23:44:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:10:54.074 23:44:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:10:54.074 23:44:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@784 -- # killprocess 86130 00:10:54.074 23:44:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@954 -- # '[' -z 86130 ']' 00:10:54.074 23:44:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@958 -- # kill -0 86130 00:10:54.074 23:44:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@959 -- # uname 00:10:54.074 23:44:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:10:54.074 23:44:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 86130 00:10:54.074 23:44:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:10:54.074 killing process with pid 86130 00:10:54.074 Received shutdown signal, test time was about 60.000000 seconds 00:10:54.074 00:10:54.074 Latency(us) 00:10:54.074 [2024-11-26T23:44:42.206Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:10:54.074 [2024-11-26T23:44:42.206Z] =================================================================================================================== 00:10:54.074 [2024-11-26T23:44:42.206Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:10:54.074 23:44:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:10:54.074 23:44:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@972 -- # echo 'killing process with pid 86130' 00:10:54.074 23:44:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@973 -- # kill 86130 00:10:54.074 [2024-11-26 23:44:42.113935] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:10:54.074 23:44:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@978 -- # wait 86130 00:10:54.074 [2024-11-26 23:44:42.114061] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:54.074 [2024-11-26 23:44:42.114113] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:54.074 [2024-11-26 23:44:42.114122] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state offline 00:10:54.074 [2024-11-26 23:44:42.145797] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:10:54.333 23:44:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@786 -- # return 0 00:10:54.333 00:10:54.333 real 0m21.323s 00:10:54.333 user 0m26.622s 00:10:54.333 sys 0m3.431s 00:10:54.333 23:44:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@1130 -- # xtrace_disable 00:10:54.333 23:44:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:54.333 ************************************ 00:10:54.333 END TEST raid_rebuild_test_sb 00:10:54.333 ************************************ 00:10:54.333 23:44:42 bdev_raid -- bdev/bdev_raid.sh@980 -- # run_test raid_rebuild_test_io raid_rebuild_test raid1 2 false true true 00:10:54.333 23:44:42 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:10:54.333 23:44:42 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:10:54.333 23:44:42 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:10:54.333 ************************************ 00:10:54.333 START TEST raid_rebuild_test_io 00:10:54.333 ************************************ 00:10:54.333 23:44:42 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@1129 -- # raid_rebuild_test raid1 2 false true true 00:10:54.333 23:44:42 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:10:54.333 23:44:42 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=2 00:10:54.333 23:44:42 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@571 -- # local superblock=false 00:10:54.333 23:44:42 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@572 -- # local background_io=true 00:10:54.333 23:44:42 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@573 -- # local verify=true 00:10:54.333 23:44:42 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:10:54.333 23:44:42 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:10:54.333 23:44:42 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:10:54.333 23:44:42 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:10:54.333 23:44:42 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:10:54.333 23:44:42 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:10:54.333 23:44:42 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:10:54.333 23:44:42 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:10:54.333 23:44:42 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:10:54.333 23:44:42 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:10:54.333 23:44:42 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:10:54.333 23:44:42 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@576 -- # local strip_size 00:10:54.333 23:44:42 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@577 -- # local create_arg 00:10:54.333 23:44:42 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:10:54.333 23:44:42 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@579 -- # local data_offset 00:10:54.333 23:44:42 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:10:54.333 23:44:42 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:10:54.333 23:44:42 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@592 -- # '[' false = true ']' 00:10:54.333 23:44:42 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@597 -- # raid_pid=86839 00:10:54.333 23:44:42 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:10:54.333 23:44:42 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@598 -- # waitforlisten 86839 00:10:54.333 23:44:42 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@835 -- # '[' -z 86839 ']' 00:10:54.333 23:44:42 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:54.333 23:44:42 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@840 -- # local max_retries=100 00:10:54.333 23:44:42 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:54.333 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:54.333 23:44:42 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@844 -- # xtrace_disable 00:10:54.333 23:44:42 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:10:54.592 [2024-11-26 23:44:42.511944] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:10:54.592 [2024-11-26 23:44:42.512160] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.ealI/O size of 3145728 is greater than zero copy threshold (65536). 00:10:54.593 Zero copy mechanism will not be used. 00:10:54.593 :6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid86839 ] 00:10:54.593 [2024-11-26 23:44:42.662513] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:10:54.593 [2024-11-26 23:44:42.688417] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:10:54.852 [2024-11-26 23:44:42.730465] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:54.852 [2024-11-26 23:44:42.730581] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:55.422 23:44:43 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:10:55.422 23:44:43 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@868 -- # return 0 00:10:55.422 23:44:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:10:55.422 23:44:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:10:55.422 23:44:43 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:55.422 23:44:43 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:10:55.422 BaseBdev1_malloc 00:10:55.422 23:44:43 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:55.422 23:44:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:10:55.422 23:44:43 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:55.422 23:44:43 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:10:55.422 [2024-11-26 23:44:43.353782] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:10:55.422 [2024-11-26 23:44:43.353840] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:55.422 [2024-11-26 23:44:43.353883] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:10:55.422 [2024-11-26 23:44:43.353895] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:55.422 [2024-11-26 23:44:43.355945] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:55.422 [2024-11-26 23:44:43.356019] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:10:55.422 BaseBdev1 00:10:55.422 23:44:43 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:55.422 23:44:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:10:55.422 23:44:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:10:55.422 23:44:43 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:55.422 23:44:43 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:10:55.422 BaseBdev2_malloc 00:10:55.422 23:44:43 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:55.422 23:44:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:10:55.422 23:44:43 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:55.422 23:44:43 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:10:55.422 [2024-11-26 23:44:43.382202] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:10:55.422 [2024-11-26 23:44:43.382314] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:55.422 [2024-11-26 23:44:43.382358] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:10:55.422 [2024-11-26 23:44:43.382368] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:55.422 [2024-11-26 23:44:43.384338] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:55.422 [2024-11-26 23:44:43.384390] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:10:55.422 BaseBdev2 00:10:55.422 23:44:43 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:55.422 23:44:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:10:55.422 23:44:43 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:55.422 23:44:43 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:10:55.422 spare_malloc 00:10:55.422 23:44:43 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:55.422 23:44:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:10:55.422 23:44:43 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:55.422 23:44:43 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:10:55.422 spare_delay 00:10:55.422 23:44:43 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:55.422 23:44:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:10:55.422 23:44:43 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:55.422 23:44:43 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:10:55.422 [2024-11-26 23:44:43.422700] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:10:55.422 [2024-11-26 23:44:43.422752] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:55.422 [2024-11-26 23:44:43.422774] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:10:55.422 [2024-11-26 23:44:43.422783] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:55.422 [2024-11-26 23:44:43.424953] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:55.422 [2024-11-26 23:44:43.425026] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:10:55.422 spare 00:10:55.422 23:44:43 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:55.422 23:44:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 00:10:55.422 23:44:43 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:55.422 23:44:43 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:10:55.422 [2024-11-26 23:44:43.434714] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:55.422 [2024-11-26 23:44:43.436565] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:55.422 [2024-11-26 23:44:43.436658] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:10:55.422 [2024-11-26 23:44:43.436668] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:10:55.423 [2024-11-26 23:44:43.436937] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:10:55.423 [2024-11-26 23:44:43.437065] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:10:55.423 [2024-11-26 23:44:43.437080] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:10:55.423 [2024-11-26 23:44:43.437218] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:55.423 23:44:43 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:55.423 23:44:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:10:55.423 23:44:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:55.423 23:44:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:55.423 23:44:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:55.423 23:44:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:55.423 23:44:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:10:55.423 23:44:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:55.423 23:44:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:55.423 23:44:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:55.423 23:44:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:55.423 23:44:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:55.423 23:44:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:55.423 23:44:43 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:55.423 23:44:43 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:10:55.423 23:44:43 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:55.423 23:44:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:55.423 "name": "raid_bdev1", 00:10:55.423 "uuid": "62e3811f-beef-425d-89c5-a6ceabbb6e90", 00:10:55.423 "strip_size_kb": 0, 00:10:55.423 "state": "online", 00:10:55.423 "raid_level": "raid1", 00:10:55.423 "superblock": false, 00:10:55.423 "num_base_bdevs": 2, 00:10:55.423 "num_base_bdevs_discovered": 2, 00:10:55.423 "num_base_bdevs_operational": 2, 00:10:55.423 "base_bdevs_list": [ 00:10:55.423 { 00:10:55.423 "name": "BaseBdev1", 00:10:55.423 "uuid": "ba460ab6-400f-5b7c-b4dd-3a51790b0521", 00:10:55.423 "is_configured": true, 00:10:55.423 "data_offset": 0, 00:10:55.423 "data_size": 65536 00:10:55.423 }, 00:10:55.423 { 00:10:55.423 "name": "BaseBdev2", 00:10:55.423 "uuid": "9d2f47dc-ddee-571d-9740-6643e13d090d", 00:10:55.423 "is_configured": true, 00:10:55.423 "data_offset": 0, 00:10:55.423 "data_size": 65536 00:10:55.423 } 00:10:55.423 ] 00:10:55.423 }' 00:10:55.423 23:44:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:55.423 23:44:43 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:10:55.990 23:44:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:10:55.990 23:44:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:10:55.990 23:44:43 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:55.990 23:44:43 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:10:55.990 [2024-11-26 23:44:43.854270] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:55.990 23:44:43 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:55.990 23:44:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=65536 00:10:55.990 23:44:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:10:55.990 23:44:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:55.990 23:44:43 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:55.990 23:44:43 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:10:55.990 23:44:43 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:55.990 23:44:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@619 -- # data_offset=0 00:10:55.990 23:44:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@621 -- # '[' true = true ']' 00:10:55.990 23:44:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:10:55.990 23:44:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@623 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:10:55.990 23:44:43 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:55.990 23:44:43 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:10:55.990 [2024-11-26 23:44:43.917883] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:10:55.990 23:44:43 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:55.990 23:44:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:10:55.990 23:44:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:55.990 23:44:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:55.990 23:44:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:55.990 23:44:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:55.990 23:44:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:10:55.990 23:44:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:55.990 23:44:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:55.990 23:44:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:55.990 23:44:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:55.990 23:44:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:55.990 23:44:43 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:55.990 23:44:43 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:10:55.990 23:44:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:55.990 23:44:43 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:55.990 23:44:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:55.990 "name": "raid_bdev1", 00:10:55.990 "uuid": "62e3811f-beef-425d-89c5-a6ceabbb6e90", 00:10:55.990 "strip_size_kb": 0, 00:10:55.990 "state": "online", 00:10:55.990 "raid_level": "raid1", 00:10:55.990 "superblock": false, 00:10:55.990 "num_base_bdevs": 2, 00:10:55.990 "num_base_bdevs_discovered": 1, 00:10:55.990 "num_base_bdevs_operational": 1, 00:10:55.990 "base_bdevs_list": [ 00:10:55.990 { 00:10:55.990 "name": null, 00:10:55.990 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:55.990 "is_configured": false, 00:10:55.990 "data_offset": 0, 00:10:55.990 "data_size": 65536 00:10:55.990 }, 00:10:55.990 { 00:10:55.990 "name": "BaseBdev2", 00:10:55.990 "uuid": "9d2f47dc-ddee-571d-9740-6643e13d090d", 00:10:55.990 "is_configured": true, 00:10:55.990 "data_offset": 0, 00:10:55.990 "data_size": 65536 00:10:55.990 } 00:10:55.990 ] 00:10:55.990 }' 00:10:55.990 23:44:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:55.990 23:44:43 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:10:55.990 [2024-11-26 23:44:44.040670] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000026d0 00:10:55.990 I/O size of 3145728 is greater than zero copy threshold (65536). 00:10:55.990 Zero copy mechanism will not be used. 00:10:55.990 Running I/O for 60 seconds... 00:10:56.250 23:44:44 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:10:56.250 23:44:44 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:56.250 23:44:44 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:10:56.250 [2024-11-26 23:44:44.296186] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:10:56.250 23:44:44 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:56.250 23:44:44 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@647 -- # sleep 1 00:10:56.250 [2024-11-26 23:44:44.338717] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000027a0 00:10:56.250 [2024-11-26 23:44:44.340683] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:10:56.508 [2024-11-26 23:44:44.459014] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:10:56.508 [2024-11-26 23:44:44.459611] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:10:56.767 [2024-11-26 23:44:44.673860] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:10:56.767 [2024-11-26 23:44:44.674279] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:10:57.025 153.00 IOPS, 459.00 MiB/s [2024-11-26T23:44:45.157Z] [2024-11-26 23:44:45.133383] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:10:57.025 [2024-11-26 23:44:45.133746] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:10:57.282 23:44:45 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:10:57.282 23:44:45 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:10:57.282 23:44:45 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:10:57.282 23:44:45 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:10:57.282 23:44:45 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:10:57.282 23:44:45 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:57.282 23:44:45 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:57.282 23:44:45 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:57.282 23:44:45 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:10:57.282 23:44:45 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:57.282 23:44:45 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:10:57.282 "name": "raid_bdev1", 00:10:57.282 "uuid": "62e3811f-beef-425d-89c5-a6ceabbb6e90", 00:10:57.282 "strip_size_kb": 0, 00:10:57.282 "state": "online", 00:10:57.282 "raid_level": "raid1", 00:10:57.282 "superblock": false, 00:10:57.282 "num_base_bdevs": 2, 00:10:57.282 "num_base_bdevs_discovered": 2, 00:10:57.282 "num_base_bdevs_operational": 2, 00:10:57.282 "process": { 00:10:57.282 "type": "rebuild", 00:10:57.282 "target": "spare", 00:10:57.282 "progress": { 00:10:57.282 "blocks": 10240, 00:10:57.282 "percent": 15 00:10:57.282 } 00:10:57.282 }, 00:10:57.282 "base_bdevs_list": [ 00:10:57.282 { 00:10:57.282 "name": "spare", 00:10:57.282 "uuid": "7fa4a95e-b00f-56c3-821d-49b907ee069a", 00:10:57.282 "is_configured": true, 00:10:57.282 "data_offset": 0, 00:10:57.282 "data_size": 65536 00:10:57.282 }, 00:10:57.282 { 00:10:57.282 "name": "BaseBdev2", 00:10:57.282 "uuid": "9d2f47dc-ddee-571d-9740-6643e13d090d", 00:10:57.282 "is_configured": true, 00:10:57.282 "data_offset": 0, 00:10:57.282 "data_size": 65536 00:10:57.282 } 00:10:57.282 ] 00:10:57.282 }' 00:10:57.283 23:44:45 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:10:57.540 23:44:45 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:10:57.540 23:44:45 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:10:57.540 23:44:45 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:10:57.540 23:44:45 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:10:57.540 23:44:45 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:57.540 23:44:45 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:10:57.540 [2024-11-26 23:44:45.469240] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:10:57.540 [2024-11-26 23:44:45.469335] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 14336 offset_begin: 12288 offset_end: 18432 00:10:57.540 [2024-11-26 23:44:45.569075] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:10:57.540 [2024-11-26 23:44:45.571168] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:57.540 [2024-11-26 23:44:45.571211] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:10:57.540 [2024-11-26 23:44:45.571228] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:10:57.540 [2024-11-26 23:44:45.595469] bdev_raid.c:1974:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 0 raid_ch: 0x60d0000026d0 00:10:57.540 23:44:45 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:57.540 23:44:45 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:10:57.540 23:44:45 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:57.540 23:44:45 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:57.540 23:44:45 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:57.540 23:44:45 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:57.540 23:44:45 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:10:57.540 23:44:45 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:57.540 23:44:45 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:57.540 23:44:45 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:57.540 23:44:45 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:57.540 23:44:45 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:57.540 23:44:45 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:57.540 23:44:45 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:57.541 23:44:45 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:10:57.541 23:44:45 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:57.541 23:44:45 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:57.541 "name": "raid_bdev1", 00:10:57.541 "uuid": "62e3811f-beef-425d-89c5-a6ceabbb6e90", 00:10:57.541 "strip_size_kb": 0, 00:10:57.541 "state": "online", 00:10:57.541 "raid_level": "raid1", 00:10:57.541 "superblock": false, 00:10:57.541 "num_base_bdevs": 2, 00:10:57.541 "num_base_bdevs_discovered": 1, 00:10:57.541 "num_base_bdevs_operational": 1, 00:10:57.541 "base_bdevs_list": [ 00:10:57.541 { 00:10:57.541 "name": null, 00:10:57.541 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:57.541 "is_configured": false, 00:10:57.541 "data_offset": 0, 00:10:57.541 "data_size": 65536 00:10:57.541 }, 00:10:57.541 { 00:10:57.541 "name": "BaseBdev2", 00:10:57.541 "uuid": "9d2f47dc-ddee-571d-9740-6643e13d090d", 00:10:57.541 "is_configured": true, 00:10:57.541 "data_offset": 0, 00:10:57.541 "data_size": 65536 00:10:57.541 } 00:10:57.541 ] 00:10:57.541 }' 00:10:57.541 23:44:45 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:57.541 23:44:45 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:10:58.107 23:44:45 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:10:58.107 23:44:45 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:10:58.107 23:44:45 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:10:58.107 23:44:45 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:10:58.107 23:44:45 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:10:58.107 23:44:45 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:58.107 23:44:45 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:58.107 23:44:45 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:58.107 23:44:45 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:10:58.107 23:44:46 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:58.107 23:44:46 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:10:58.107 "name": "raid_bdev1", 00:10:58.107 "uuid": "62e3811f-beef-425d-89c5-a6ceabbb6e90", 00:10:58.107 "strip_size_kb": 0, 00:10:58.107 "state": "online", 00:10:58.107 "raid_level": "raid1", 00:10:58.107 "superblock": false, 00:10:58.107 "num_base_bdevs": 2, 00:10:58.107 "num_base_bdevs_discovered": 1, 00:10:58.107 "num_base_bdevs_operational": 1, 00:10:58.107 "base_bdevs_list": [ 00:10:58.107 { 00:10:58.107 "name": null, 00:10:58.107 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:58.107 "is_configured": false, 00:10:58.107 "data_offset": 0, 00:10:58.107 "data_size": 65536 00:10:58.107 }, 00:10:58.107 { 00:10:58.107 "name": "BaseBdev2", 00:10:58.107 "uuid": "9d2f47dc-ddee-571d-9740-6643e13d090d", 00:10:58.107 "is_configured": true, 00:10:58.107 "data_offset": 0, 00:10:58.107 "data_size": 65536 00:10:58.107 } 00:10:58.107 ] 00:10:58.107 }' 00:10:58.107 23:44:46 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:10:58.107 176.00 IOPS, 528.00 MiB/s [2024-11-26T23:44:46.239Z] 23:44:46 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:10:58.107 23:44:46 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:10:58.107 23:44:46 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:10:58.107 23:44:46 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:10:58.107 23:44:46 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:58.107 23:44:46 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:10:58.107 [2024-11-26 23:44:46.127261] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:10:58.107 23:44:46 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:58.107 23:44:46 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@663 -- # sleep 1 00:10:58.107 [2024-11-26 23:44:46.159698] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002870 00:10:58.107 [2024-11-26 23:44:46.161569] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:10:58.366 [2024-11-26 23:44:46.277775] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:10:58.625 [2024-11-26 23:44:46.508295] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:10:58.625 [2024-11-26 23:44:46.508604] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:10:58.625 [2024-11-26 23:44:46.748334] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:10:58.884 [2024-11-26 23:44:46.962539] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:10:59.143 161.00 IOPS, 483.00 MiB/s [2024-11-26T23:44:47.275Z] 23:44:47 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:10:59.143 23:44:47 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:10:59.143 23:44:47 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:10:59.143 23:44:47 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:10:59.143 23:44:47 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:10:59.143 23:44:47 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:59.143 23:44:47 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:59.143 23:44:47 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:59.143 23:44:47 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:10:59.143 23:44:47 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:59.143 23:44:47 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:10:59.143 "name": "raid_bdev1", 00:10:59.143 "uuid": "62e3811f-beef-425d-89c5-a6ceabbb6e90", 00:10:59.143 "strip_size_kb": 0, 00:10:59.143 "state": "online", 00:10:59.143 "raid_level": "raid1", 00:10:59.143 "superblock": false, 00:10:59.143 "num_base_bdevs": 2, 00:10:59.143 "num_base_bdevs_discovered": 2, 00:10:59.143 "num_base_bdevs_operational": 2, 00:10:59.143 "process": { 00:10:59.143 "type": "rebuild", 00:10:59.143 "target": "spare", 00:10:59.143 "progress": { 00:10:59.143 "blocks": 12288, 00:10:59.143 "percent": 18 00:10:59.143 } 00:10:59.143 }, 00:10:59.143 "base_bdevs_list": [ 00:10:59.143 { 00:10:59.143 "name": "spare", 00:10:59.143 "uuid": "7fa4a95e-b00f-56c3-821d-49b907ee069a", 00:10:59.143 "is_configured": true, 00:10:59.143 "data_offset": 0, 00:10:59.143 "data_size": 65536 00:10:59.143 }, 00:10:59.143 { 00:10:59.143 "name": "BaseBdev2", 00:10:59.143 "uuid": "9d2f47dc-ddee-571d-9740-6643e13d090d", 00:10:59.143 "is_configured": true, 00:10:59.143 "data_offset": 0, 00:10:59.143 "data_size": 65536 00:10:59.143 } 00:10:59.143 ] 00:10:59.143 }' 00:10:59.143 23:44:47 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:10:59.143 23:44:47 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:10:59.143 23:44:47 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:10:59.402 [2024-11-26 23:44:47.304051] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:10:59.402 23:44:47 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:10:59.402 23:44:47 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@666 -- # '[' false = true ']' 00:10:59.402 23:44:47 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=2 00:10:59.402 23:44:47 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:10:59.402 23:44:47 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@693 -- # '[' 2 -gt 2 ']' 00:10:59.402 23:44:47 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@706 -- # local timeout=319 00:10:59.402 23:44:47 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:10:59.402 23:44:47 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:10:59.402 23:44:47 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:10:59.402 23:44:47 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:10:59.402 23:44:47 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:10:59.402 23:44:47 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:10:59.402 23:44:47 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:59.402 23:44:47 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:59.402 23:44:47 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:10:59.402 23:44:47 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:59.402 23:44:47 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:59.402 23:44:47 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:10:59.402 "name": "raid_bdev1", 00:10:59.402 "uuid": "62e3811f-beef-425d-89c5-a6ceabbb6e90", 00:10:59.402 "strip_size_kb": 0, 00:10:59.402 "state": "online", 00:10:59.402 "raid_level": "raid1", 00:10:59.402 "superblock": false, 00:10:59.402 "num_base_bdevs": 2, 00:10:59.402 "num_base_bdevs_discovered": 2, 00:10:59.402 "num_base_bdevs_operational": 2, 00:10:59.402 "process": { 00:10:59.402 "type": "rebuild", 00:10:59.402 "target": "spare", 00:10:59.402 "progress": { 00:10:59.402 "blocks": 16384, 00:10:59.402 "percent": 25 00:10:59.402 } 00:10:59.402 }, 00:10:59.402 "base_bdevs_list": [ 00:10:59.402 { 00:10:59.402 "name": "spare", 00:10:59.402 "uuid": "7fa4a95e-b00f-56c3-821d-49b907ee069a", 00:10:59.402 "is_configured": true, 00:10:59.402 "data_offset": 0, 00:10:59.402 "data_size": 65536 00:10:59.402 }, 00:10:59.402 { 00:10:59.402 "name": "BaseBdev2", 00:10:59.402 "uuid": "9d2f47dc-ddee-571d-9740-6643e13d090d", 00:10:59.402 "is_configured": true, 00:10:59.402 "data_offset": 0, 00:10:59.402 "data_size": 65536 00:10:59.402 } 00:10:59.402 ] 00:10:59.402 }' 00:10:59.402 23:44:47 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:10:59.402 23:44:47 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:10:59.402 23:44:47 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:10:59.402 23:44:47 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:10:59.402 23:44:47 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:10:59.661 [2024-11-26 23:44:47.622181] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 22528 offset_begin: 18432 offset_end: 24576 00:10:59.920 [2024-11-26 23:44:47.839987] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 26624 offset_begin: 24576 offset_end: 30720 00:10:59.920 [2024-11-26 23:44:47.840532] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 26624 offset_begin: 24576 offset_end: 30720 00:11:00.186 139.00 IOPS, 417.00 MiB/s [2024-11-26T23:44:48.318Z] [2024-11-26 23:44:48.255980] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 32768 offset_begin: 30720 offset_end: 36864 00:11:00.448 23:44:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:11:00.448 23:44:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:00.448 23:44:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:00.448 23:44:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:00.448 23:44:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:00.448 23:44:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:00.448 23:44:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:00.448 23:44:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:00.448 23:44:48 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:00.448 23:44:48 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:00.448 23:44:48 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:00.448 23:44:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:00.448 "name": "raid_bdev1", 00:11:00.448 "uuid": "62e3811f-beef-425d-89c5-a6ceabbb6e90", 00:11:00.448 "strip_size_kb": 0, 00:11:00.448 "state": "online", 00:11:00.448 "raid_level": "raid1", 00:11:00.448 "superblock": false, 00:11:00.448 "num_base_bdevs": 2, 00:11:00.448 "num_base_bdevs_discovered": 2, 00:11:00.448 "num_base_bdevs_operational": 2, 00:11:00.448 "process": { 00:11:00.448 "type": "rebuild", 00:11:00.448 "target": "spare", 00:11:00.448 "progress": { 00:11:00.448 "blocks": 34816, 00:11:00.448 "percent": 53 00:11:00.448 } 00:11:00.448 }, 00:11:00.448 "base_bdevs_list": [ 00:11:00.448 { 00:11:00.448 "name": "spare", 00:11:00.448 "uuid": "7fa4a95e-b00f-56c3-821d-49b907ee069a", 00:11:00.448 "is_configured": true, 00:11:00.448 "data_offset": 0, 00:11:00.448 "data_size": 65536 00:11:00.448 }, 00:11:00.448 { 00:11:00.448 "name": "BaseBdev2", 00:11:00.448 "uuid": "9d2f47dc-ddee-571d-9740-6643e13d090d", 00:11:00.448 "is_configured": true, 00:11:00.448 "data_offset": 0, 00:11:00.448 "data_size": 65536 00:11:00.448 } 00:11:00.448 ] 00:11:00.448 }' 00:11:00.448 23:44:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:00.448 23:44:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:00.448 23:44:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:00.707 23:44:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:00.707 23:44:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:11:01.542 120.80 IOPS, 362.40 MiB/s [2024-11-26T23:44:49.674Z] [2024-11-26 23:44:49.552807] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 57344 offset_begin: 55296 offset_end: 61440 00:11:01.542 23:44:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:11:01.542 23:44:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:01.542 23:44:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:01.542 23:44:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:01.542 23:44:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:01.542 23:44:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:01.542 23:44:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:01.542 23:44:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:01.542 23:44:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:01.542 23:44:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:01.542 23:44:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:01.542 23:44:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:01.542 "name": "raid_bdev1", 00:11:01.542 "uuid": "62e3811f-beef-425d-89c5-a6ceabbb6e90", 00:11:01.542 "strip_size_kb": 0, 00:11:01.542 "state": "online", 00:11:01.542 "raid_level": "raid1", 00:11:01.542 "superblock": false, 00:11:01.542 "num_base_bdevs": 2, 00:11:01.542 "num_base_bdevs_discovered": 2, 00:11:01.542 "num_base_bdevs_operational": 2, 00:11:01.542 "process": { 00:11:01.542 "type": "rebuild", 00:11:01.542 "target": "spare", 00:11:01.542 "progress": { 00:11:01.542 "blocks": 57344, 00:11:01.542 "percent": 87 00:11:01.542 } 00:11:01.542 }, 00:11:01.542 "base_bdevs_list": [ 00:11:01.542 { 00:11:01.542 "name": "spare", 00:11:01.542 "uuid": "7fa4a95e-b00f-56c3-821d-49b907ee069a", 00:11:01.542 "is_configured": true, 00:11:01.542 "data_offset": 0, 00:11:01.542 "data_size": 65536 00:11:01.542 }, 00:11:01.542 { 00:11:01.542 "name": "BaseBdev2", 00:11:01.542 "uuid": "9d2f47dc-ddee-571d-9740-6643e13d090d", 00:11:01.542 "is_configured": true, 00:11:01.542 "data_offset": 0, 00:11:01.542 "data_size": 65536 00:11:01.542 } 00:11:01.542 ] 00:11:01.542 }' 00:11:01.542 23:44:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:01.801 23:44:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:01.801 23:44:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:01.801 23:44:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:01.801 23:44:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:11:02.060 106.33 IOPS, 319.00 MiB/s [2024-11-26T23:44:50.192Z] [2024-11-26 23:44:50.086083] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:11:02.320 [2024-11-26 23:44:50.191339] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:11:02.320 [2024-11-26 23:44:50.193435] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:02.890 23:44:50 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:11:02.890 23:44:50 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:02.890 23:44:50 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:02.890 23:44:50 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:02.890 23:44:50 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:02.890 23:44:50 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:02.890 23:44:50 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:02.890 23:44:50 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:02.890 23:44:50 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:02.890 23:44:50 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:02.890 23:44:50 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:02.890 23:44:50 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:02.890 "name": "raid_bdev1", 00:11:02.890 "uuid": "62e3811f-beef-425d-89c5-a6ceabbb6e90", 00:11:02.890 "strip_size_kb": 0, 00:11:02.890 "state": "online", 00:11:02.890 "raid_level": "raid1", 00:11:02.890 "superblock": false, 00:11:02.890 "num_base_bdevs": 2, 00:11:02.890 "num_base_bdevs_discovered": 2, 00:11:02.890 "num_base_bdevs_operational": 2, 00:11:02.890 "base_bdevs_list": [ 00:11:02.890 { 00:11:02.890 "name": "spare", 00:11:02.890 "uuid": "7fa4a95e-b00f-56c3-821d-49b907ee069a", 00:11:02.890 "is_configured": true, 00:11:02.890 "data_offset": 0, 00:11:02.890 "data_size": 65536 00:11:02.890 }, 00:11:02.890 { 00:11:02.890 "name": "BaseBdev2", 00:11:02.890 "uuid": "9d2f47dc-ddee-571d-9740-6643e13d090d", 00:11:02.890 "is_configured": true, 00:11:02.890 "data_offset": 0, 00:11:02.890 "data_size": 65536 00:11:02.890 } 00:11:02.890 ] 00:11:02.890 }' 00:11:02.890 23:44:50 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:02.890 23:44:50 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:11:02.890 23:44:50 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:02.890 23:44:50 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:11:02.890 23:44:50 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@709 -- # break 00:11:02.890 23:44:50 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:11:02.890 23:44:50 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:02.890 23:44:50 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:11:02.890 23:44:50 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:11:02.890 23:44:50 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:02.890 23:44:50 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:02.890 23:44:50 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:02.890 23:44:50 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:02.890 23:44:50 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:02.890 23:44:50 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:02.890 23:44:50 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:02.890 "name": "raid_bdev1", 00:11:02.890 "uuid": "62e3811f-beef-425d-89c5-a6ceabbb6e90", 00:11:02.890 "strip_size_kb": 0, 00:11:02.890 "state": "online", 00:11:02.890 "raid_level": "raid1", 00:11:02.890 "superblock": false, 00:11:02.890 "num_base_bdevs": 2, 00:11:02.890 "num_base_bdevs_discovered": 2, 00:11:02.890 "num_base_bdevs_operational": 2, 00:11:02.890 "base_bdevs_list": [ 00:11:02.890 { 00:11:02.890 "name": "spare", 00:11:02.890 "uuid": "7fa4a95e-b00f-56c3-821d-49b907ee069a", 00:11:02.890 "is_configured": true, 00:11:02.890 "data_offset": 0, 00:11:02.890 "data_size": 65536 00:11:02.890 }, 00:11:02.890 { 00:11:02.890 "name": "BaseBdev2", 00:11:02.890 "uuid": "9d2f47dc-ddee-571d-9740-6643e13d090d", 00:11:02.890 "is_configured": true, 00:11:02.890 "data_offset": 0, 00:11:02.890 "data_size": 65536 00:11:02.890 } 00:11:02.890 ] 00:11:02.890 }' 00:11:02.890 23:44:50 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:02.890 23:44:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:11:02.890 23:44:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:03.149 95.71 IOPS, 287.14 MiB/s [2024-11-26T23:44:51.281Z] 23:44:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:11:03.149 23:44:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:11:03.149 23:44:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:03.149 23:44:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:03.149 23:44:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:03.149 23:44:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:03.149 23:44:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:11:03.149 23:44:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:03.149 23:44:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:03.149 23:44:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:03.149 23:44:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:03.149 23:44:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:03.149 23:44:51 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:03.149 23:44:51 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:03.149 23:44:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:03.149 23:44:51 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:03.149 23:44:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:03.149 "name": "raid_bdev1", 00:11:03.149 "uuid": "62e3811f-beef-425d-89c5-a6ceabbb6e90", 00:11:03.149 "strip_size_kb": 0, 00:11:03.149 "state": "online", 00:11:03.149 "raid_level": "raid1", 00:11:03.149 "superblock": false, 00:11:03.149 "num_base_bdevs": 2, 00:11:03.149 "num_base_bdevs_discovered": 2, 00:11:03.149 "num_base_bdevs_operational": 2, 00:11:03.149 "base_bdevs_list": [ 00:11:03.149 { 00:11:03.149 "name": "spare", 00:11:03.149 "uuid": "7fa4a95e-b00f-56c3-821d-49b907ee069a", 00:11:03.149 "is_configured": true, 00:11:03.149 "data_offset": 0, 00:11:03.149 "data_size": 65536 00:11:03.149 }, 00:11:03.149 { 00:11:03.149 "name": "BaseBdev2", 00:11:03.149 "uuid": "9d2f47dc-ddee-571d-9740-6643e13d090d", 00:11:03.149 "is_configured": true, 00:11:03.149 "data_offset": 0, 00:11:03.149 "data_size": 65536 00:11:03.149 } 00:11:03.149 ] 00:11:03.149 }' 00:11:03.149 23:44:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:03.149 23:44:51 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:03.423 23:44:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:11:03.423 23:44:51 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:03.423 23:44:51 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:03.423 [2024-11-26 23:44:51.539939] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:11:03.423 [2024-11-26 23:44:51.539972] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:11:03.685 00:11:03.685 Latency(us) 00:11:03.685 [2024-11-26T23:44:51.817Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:11:03.685 Job: raid_bdev1 (Core Mask 0x1, workload: randrw, percentage: 50, depth: 2, IO size: 3145728) 00:11:03.685 raid_bdev1 : 7.56 91.68 275.03 0.00 0.00 15614.48 280.82 108062.85 00:11:03.685 [2024-11-26T23:44:51.817Z] =================================================================================================================== 00:11:03.685 [2024-11-26T23:44:51.817Z] Total : 91.68 275.03 0.00 0.00 15614.48 280.82 108062.85 00:11:03.685 { 00:11:03.685 "results": [ 00:11:03.685 { 00:11:03.685 "job": "raid_bdev1", 00:11:03.685 "core_mask": "0x1", 00:11:03.685 "workload": "randrw", 00:11:03.685 "percentage": 50, 00:11:03.685 "status": "finished", 00:11:03.685 "queue_depth": 2, 00:11:03.685 "io_size": 3145728, 00:11:03.685 "runtime": 7.559204, 00:11:03.685 "iops": 91.6763193584933, 00:11:03.685 "mibps": 275.0289580754799, 00:11:03.685 "io_failed": 0, 00:11:03.685 "io_timeout": 0, 00:11:03.685 "avg_latency_us": 15614.478166568995, 00:11:03.685 "min_latency_us": 280.8174672489083, 00:11:03.685 "max_latency_us": 108062.85414847161 00:11:03.685 } 00:11:03.685 ], 00:11:03.685 "core_count": 1 00:11:03.685 } 00:11:03.685 [2024-11-26 23:44:51.591419] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:11:03.685 [2024-11-26 23:44:51.591467] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:03.685 [2024-11-26 23:44:51.591537] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:11:03.685 [2024-11-26 23:44:51.591559] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:11:03.685 23:44:51 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:03.685 23:44:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:03.685 23:44:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@720 -- # jq length 00:11:03.685 23:44:51 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:03.685 23:44:51 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:03.685 23:44:51 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:03.685 23:44:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:11:03.685 23:44:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:11:03.685 23:44:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@723 -- # '[' true = true ']' 00:11:03.685 23:44:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@725 -- # nbd_start_disks /var/tmp/spdk.sock spare /dev/nbd0 00:11:03.685 23:44:51 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:11:03.685 23:44:51 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # bdev_list=('spare') 00:11:03.685 23:44:51 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:11:03.685 23:44:51 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:11:03.685 23:44:51 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:11:03.685 23:44:51 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@12 -- # local i 00:11:03.685 23:44:51 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:11:03.685 23:44:51 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:11:03.685 23:44:51 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd0 00:11:03.946 /dev/nbd0 00:11:03.946 23:44:51 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:11:03.946 23:44:51 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:11:03.946 23:44:51 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:11:03.946 23:44:51 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@873 -- # local i 00:11:03.946 23:44:51 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:11:03.946 23:44:51 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:11:03.946 23:44:51 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:11:03.946 23:44:51 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@877 -- # break 00:11:03.946 23:44:51 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:11:03.946 23:44:51 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:11:03.946 23:44:51 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:11:03.946 1+0 records in 00:11:03.946 1+0 records out 00:11:03.946 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000377632 s, 10.8 MB/s 00:11:03.946 23:44:51 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:03.946 23:44:51 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@890 -- # size=4096 00:11:03.946 23:44:51 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:03.946 23:44:51 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:11:03.946 23:44:51 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@893 -- # return 0 00:11:03.946 23:44:51 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:11:03.946 23:44:51 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:11:03.946 23:44:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@726 -- # for bdev in "${base_bdevs[@]:1}" 00:11:03.946 23:44:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@727 -- # '[' -z BaseBdev2 ']' 00:11:03.946 23:44:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@730 -- # nbd_start_disks /var/tmp/spdk.sock BaseBdev2 /dev/nbd1 00:11:03.946 23:44:51 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:11:03.946 23:44:51 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev2') 00:11:03.946 23:44:51 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:11:03.946 23:44:51 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd1') 00:11:03.946 23:44:51 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:11:03.946 23:44:51 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@12 -- # local i 00:11:03.946 23:44:51 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:11:03.946 23:44:51 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:11:03.946 23:44:51 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev2 /dev/nbd1 00:11:04.219 /dev/nbd1 00:11:04.219 23:44:52 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:11:04.219 23:44:52 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:11:04.219 23:44:52 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:11:04.219 23:44:52 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@873 -- # local i 00:11:04.219 23:44:52 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:11:04.219 23:44:52 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:11:04.219 23:44:52 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:11:04.219 23:44:52 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@877 -- # break 00:11:04.219 23:44:52 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:11:04.219 23:44:52 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:11:04.219 23:44:52 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:11:04.219 1+0 records in 00:11:04.219 1+0 records out 00:11:04.219 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000418545 s, 9.8 MB/s 00:11:04.219 23:44:52 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:04.219 23:44:52 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@890 -- # size=4096 00:11:04.219 23:44:52 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:04.219 23:44:52 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:11:04.219 23:44:52 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@893 -- # return 0 00:11:04.219 23:44:52 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:11:04.219 23:44:52 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:11:04.219 23:44:52 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@731 -- # cmp -i 0 /dev/nbd0 /dev/nbd1 00:11:04.219 23:44:52 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@732 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd1 00:11:04.219 23:44:52 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:11:04.219 23:44:52 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd1') 00:11:04.219 23:44:52 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:11:04.219 23:44:52 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@51 -- # local i 00:11:04.219 23:44:52 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:11:04.219 23:44:52 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:11:04.499 23:44:52 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:11:04.499 23:44:52 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:11:04.499 23:44:52 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:11:04.499 23:44:52 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:11:04.499 23:44:52 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:11:04.499 23:44:52 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:11:04.499 23:44:52 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@41 -- # break 00:11:04.499 23:44:52 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@45 -- # return 0 00:11:04.499 23:44:52 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@734 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:11:04.499 23:44:52 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:11:04.499 23:44:52 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:11:04.499 23:44:52 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:11:04.499 23:44:52 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@51 -- # local i 00:11:04.499 23:44:52 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:11:04.499 23:44:52 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:11:04.499 23:44:52 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:11:04.759 23:44:52 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:11:04.759 23:44:52 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:11:04.759 23:44:52 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:11:04.759 23:44:52 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:11:04.759 23:44:52 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:11:04.759 23:44:52 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@41 -- # break 00:11:04.759 23:44:52 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@45 -- # return 0 00:11:04.759 23:44:52 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@743 -- # '[' false = true ']' 00:11:04.759 23:44:52 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@784 -- # killprocess 86839 00:11:04.759 23:44:52 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@954 -- # '[' -z 86839 ']' 00:11:04.759 23:44:52 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@958 -- # kill -0 86839 00:11:04.759 23:44:52 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@959 -- # uname 00:11:04.759 23:44:52 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:11:04.759 23:44:52 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 86839 00:11:04.759 23:44:52 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:11:04.759 23:44:52 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:11:04.759 killing process with pid 86839 00:11:04.759 23:44:52 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@972 -- # echo 'killing process with pid 86839' 00:11:04.759 23:44:52 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@973 -- # kill 86839 00:11:04.759 Received shutdown signal, test time was about 8.665037 seconds 00:11:04.759 00:11:04.759 Latency(us) 00:11:04.759 [2024-11-26T23:44:52.891Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:11:04.759 [2024-11-26T23:44:52.891Z] =================================================================================================================== 00:11:04.759 [2024-11-26T23:44:52.891Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:11:04.759 [2024-11-26 23:44:52.691535] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:11:04.759 23:44:52 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@978 -- # wait 86839 00:11:04.759 [2024-11-26 23:44:52.740992] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:11:05.019 23:44:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@786 -- # return 0 00:11:05.019 00:11:05.019 real 0m10.645s 00:11:05.019 user 0m13.704s 00:11:05.019 sys 0m1.305s 00:11:05.019 ************************************ 00:11:05.019 END TEST raid_rebuild_test_io 00:11:05.019 ************************************ 00:11:05.019 23:44:53 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@1130 -- # xtrace_disable 00:11:05.019 23:44:53 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:05.019 23:44:53 bdev_raid -- bdev/bdev_raid.sh@981 -- # run_test raid_rebuild_test_sb_io raid_rebuild_test raid1 2 true true true 00:11:05.020 23:44:53 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:11:05.020 23:44:53 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:11:05.020 23:44:53 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:11:05.020 ************************************ 00:11:05.020 START TEST raid_rebuild_test_sb_io 00:11:05.020 ************************************ 00:11:05.020 23:44:53 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@1129 -- # raid_rebuild_test raid1 2 true true true 00:11:05.020 23:44:53 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:11:05.020 23:44:53 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=2 00:11:05.020 23:44:53 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@571 -- # local superblock=true 00:11:05.020 23:44:53 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@572 -- # local background_io=true 00:11:05.020 23:44:53 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@573 -- # local verify=true 00:11:05.020 23:44:53 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:11:05.020 23:44:53 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:11:05.020 23:44:53 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:11:05.020 23:44:53 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:11:05.020 23:44:53 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:11:05.020 23:44:53 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:11:05.020 23:44:53 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:11:05.020 23:44:53 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:11:05.020 23:44:53 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:11:05.020 23:44:53 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:11:05.020 23:44:53 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:11:05.020 23:44:53 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@576 -- # local strip_size 00:11:05.020 23:44:53 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@577 -- # local create_arg 00:11:05.020 23:44:53 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:11:05.020 23:44:53 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@579 -- # local data_offset 00:11:05.020 23:44:53 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:11:05.020 23:44:53 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:11:05.020 23:44:53 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@592 -- # '[' true = true ']' 00:11:05.020 23:44:53 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@593 -- # create_arg+=' -s' 00:11:05.020 23:44:53 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@597 -- # raid_pid=87201 00:11:05.020 23:44:53 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@598 -- # waitforlisten 87201 00:11:05.020 23:44:53 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:11:05.020 23:44:53 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@835 -- # '[' -z 87201 ']' 00:11:05.020 23:44:53 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:11:05.281 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:11:05.281 23:44:53 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@840 -- # local max_retries=100 00:11:05.281 23:44:53 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:11:05.281 23:44:53 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@844 -- # xtrace_disable 00:11:05.281 23:44:53 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:05.281 [2024-11-26 23:44:53.235734] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:11:05.281 I/O size of 3145728 is greater than zero copy threshold (65536). 00:11:05.281 Zero copy mechanism will not be used. 00:11:05.281 [2024-11-26 23:44:53.235975] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid87201 ] 00:11:05.281 [2024-11-26 23:44:53.390701] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:11:05.541 [2024-11-26 23:44:53.432647] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:11:05.541 [2024-11-26 23:44:53.508777] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:05.541 [2024-11-26 23:44:53.508818] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:06.109 23:44:54 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:11:06.109 23:44:54 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@868 -- # return 0 00:11:06.109 23:44:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:11:06.109 23:44:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:11:06.109 23:44:54 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:06.109 23:44:54 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:06.109 BaseBdev1_malloc 00:11:06.109 23:44:54 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:06.109 23:44:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:11:06.109 23:44:54 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:06.109 23:44:54 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:06.109 [2024-11-26 23:44:54.065818] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:11:06.109 [2024-11-26 23:44:54.065897] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:06.109 [2024-11-26 23:44:54.065929] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:11:06.109 [2024-11-26 23:44:54.065943] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:06.109 [2024-11-26 23:44:54.068443] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:06.109 [2024-11-26 23:44:54.068479] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:11:06.109 BaseBdev1 00:11:06.109 23:44:54 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:06.109 23:44:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:11:06.109 23:44:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:11:06.109 23:44:54 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:06.109 23:44:54 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:06.109 BaseBdev2_malloc 00:11:06.109 23:44:54 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:06.110 23:44:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:11:06.110 23:44:54 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:06.110 23:44:54 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:06.110 [2024-11-26 23:44:54.100798] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:11:06.110 [2024-11-26 23:44:54.100862] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:06.110 [2024-11-26 23:44:54.100891] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:11:06.110 [2024-11-26 23:44:54.100900] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:06.110 [2024-11-26 23:44:54.103463] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:06.110 [2024-11-26 23:44:54.103552] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:11:06.110 BaseBdev2 00:11:06.110 23:44:54 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:06.110 23:44:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:11:06.110 23:44:54 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:06.110 23:44:54 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:06.110 spare_malloc 00:11:06.110 23:44:54 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:06.110 23:44:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:11:06.110 23:44:54 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:06.110 23:44:54 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:06.110 spare_delay 00:11:06.110 23:44:54 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:06.110 23:44:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:11:06.110 23:44:54 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:06.110 23:44:54 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:06.110 [2024-11-26 23:44:54.147767] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:11:06.110 [2024-11-26 23:44:54.147826] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:06.110 [2024-11-26 23:44:54.147849] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:11:06.110 [2024-11-26 23:44:54.147859] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:06.110 [2024-11-26 23:44:54.150396] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:06.110 [2024-11-26 23:44:54.150481] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:11:06.110 spare 00:11:06.110 23:44:54 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:06.110 23:44:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 00:11:06.110 23:44:54 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:06.110 23:44:54 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:06.110 [2024-11-26 23:44:54.159804] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:11:06.110 [2024-11-26 23:44:54.161982] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:11:06.110 [2024-11-26 23:44:54.162209] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:11:06.110 [2024-11-26 23:44:54.162228] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:11:06.110 [2024-11-26 23:44:54.162546] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:11:06.110 [2024-11-26 23:44:54.162710] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:11:06.110 [2024-11-26 23:44:54.162724] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:11:06.110 [2024-11-26 23:44:54.162859] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:06.110 23:44:54 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:06.110 23:44:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:11:06.110 23:44:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:06.110 23:44:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:06.110 23:44:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:06.110 23:44:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:06.110 23:44:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:11:06.110 23:44:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:06.110 23:44:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:06.110 23:44:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:06.110 23:44:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:06.110 23:44:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:06.110 23:44:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:06.110 23:44:54 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:06.110 23:44:54 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:06.110 23:44:54 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:06.110 23:44:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:06.110 "name": "raid_bdev1", 00:11:06.110 "uuid": "dfec84f0-3912-4aa3-b332-ee4e99f2a7bd", 00:11:06.110 "strip_size_kb": 0, 00:11:06.110 "state": "online", 00:11:06.110 "raid_level": "raid1", 00:11:06.110 "superblock": true, 00:11:06.110 "num_base_bdevs": 2, 00:11:06.110 "num_base_bdevs_discovered": 2, 00:11:06.110 "num_base_bdevs_operational": 2, 00:11:06.110 "base_bdevs_list": [ 00:11:06.110 { 00:11:06.110 "name": "BaseBdev1", 00:11:06.110 "uuid": "fbd89d9b-ecfa-581e-8b7e-eb784bf1d0e6", 00:11:06.110 "is_configured": true, 00:11:06.110 "data_offset": 2048, 00:11:06.110 "data_size": 63488 00:11:06.110 }, 00:11:06.110 { 00:11:06.110 "name": "BaseBdev2", 00:11:06.110 "uuid": "da9693d2-7593-58b9-8d53-16093d346229", 00:11:06.110 "is_configured": true, 00:11:06.110 "data_offset": 2048, 00:11:06.110 "data_size": 63488 00:11:06.110 } 00:11:06.110 ] 00:11:06.110 }' 00:11:06.110 23:44:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:06.110 23:44:54 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:06.687 23:44:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:11:06.687 23:44:54 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:06.687 23:44:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:11:06.687 23:44:54 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:06.687 [2024-11-26 23:44:54.587440] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:11:06.687 23:44:54 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:06.687 23:44:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=63488 00:11:06.688 23:44:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:06.688 23:44:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:11:06.688 23:44:54 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:06.688 23:44:54 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:06.688 23:44:54 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:06.688 23:44:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@619 -- # data_offset=2048 00:11:06.688 23:44:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@621 -- # '[' true = true ']' 00:11:06.688 23:44:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@623 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:11:06.688 23:44:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:11:06.688 23:44:54 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:06.688 23:44:54 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:06.688 [2024-11-26 23:44:54.678944] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:11:06.688 23:44:54 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:06.688 23:44:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:11:06.688 23:44:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:06.688 23:44:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:06.688 23:44:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:06.688 23:44:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:06.688 23:44:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:11:06.688 23:44:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:06.689 23:44:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:06.689 23:44:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:06.689 23:44:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:06.689 23:44:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:06.689 23:44:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:06.689 23:44:54 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:06.689 23:44:54 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:06.689 23:44:54 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:06.689 23:44:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:06.689 "name": "raid_bdev1", 00:11:06.689 "uuid": "dfec84f0-3912-4aa3-b332-ee4e99f2a7bd", 00:11:06.689 "strip_size_kb": 0, 00:11:06.689 "state": "online", 00:11:06.689 "raid_level": "raid1", 00:11:06.689 "superblock": true, 00:11:06.689 "num_base_bdevs": 2, 00:11:06.689 "num_base_bdevs_discovered": 1, 00:11:06.689 "num_base_bdevs_operational": 1, 00:11:06.689 "base_bdevs_list": [ 00:11:06.689 { 00:11:06.689 "name": null, 00:11:06.689 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:06.689 "is_configured": false, 00:11:06.689 "data_offset": 0, 00:11:06.689 "data_size": 63488 00:11:06.689 }, 00:11:06.689 { 00:11:06.689 "name": "BaseBdev2", 00:11:06.689 "uuid": "da9693d2-7593-58b9-8d53-16093d346229", 00:11:06.689 "is_configured": true, 00:11:06.689 "data_offset": 2048, 00:11:06.689 "data_size": 63488 00:11:06.689 } 00:11:06.689 ] 00:11:06.689 }' 00:11:06.689 23:44:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:06.689 23:44:54 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:06.689 [2024-11-26 23:44:54.764663] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000026d0 00:11:06.689 I/O size of 3145728 is greater than zero copy threshold (65536). 00:11:06.689 Zero copy mechanism will not be used. 00:11:06.689 Running I/O for 60 seconds... 00:11:07.265 23:44:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:11:07.265 23:44:55 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:07.265 23:44:55 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:07.265 [2024-11-26 23:44:55.122432] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:11:07.265 23:44:55 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:07.265 23:44:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@647 -- # sleep 1 00:11:07.265 [2024-11-26 23:44:55.181008] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000027a0 00:11:07.265 [2024-11-26 23:44:55.182919] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:11:07.265 [2024-11-26 23:44:55.293523] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:11:07.265 [2024-11-26 23:44:55.293988] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:11:07.523 [2024-11-26 23:44:55.537071] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:11:07.781 237.00 IOPS, 711.00 MiB/s [2024-11-26T23:44:55.913Z] [2024-11-26 23:44:55.782724] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:11:07.781 [2024-11-26 23:44:55.783094] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:11:08.039 [2024-11-26 23:44:55.995253] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:11:08.039 [2024-11-26 23:44:55.995464] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:11:08.039 23:44:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:08.039 23:44:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:08.039 23:44:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:08.039 23:44:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:08.039 23:44:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:08.039 23:44:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:08.039 23:44:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:08.039 23:44:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:08.039 23:44:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:08.298 23:44:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:08.298 23:44:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:08.298 "name": "raid_bdev1", 00:11:08.298 "uuid": "dfec84f0-3912-4aa3-b332-ee4e99f2a7bd", 00:11:08.298 "strip_size_kb": 0, 00:11:08.298 "state": "online", 00:11:08.298 "raid_level": "raid1", 00:11:08.298 "superblock": true, 00:11:08.298 "num_base_bdevs": 2, 00:11:08.298 "num_base_bdevs_discovered": 2, 00:11:08.298 "num_base_bdevs_operational": 2, 00:11:08.298 "process": { 00:11:08.298 "type": "rebuild", 00:11:08.298 "target": "spare", 00:11:08.298 "progress": { 00:11:08.298 "blocks": 10240, 00:11:08.298 "percent": 16 00:11:08.298 } 00:11:08.298 }, 00:11:08.298 "base_bdevs_list": [ 00:11:08.298 { 00:11:08.298 "name": "spare", 00:11:08.298 "uuid": "1d4c4384-173d-51ef-ae81-2eb69239e686", 00:11:08.298 "is_configured": true, 00:11:08.298 "data_offset": 2048, 00:11:08.298 "data_size": 63488 00:11:08.298 }, 00:11:08.298 { 00:11:08.298 "name": "BaseBdev2", 00:11:08.298 "uuid": "da9693d2-7593-58b9-8d53-16093d346229", 00:11:08.298 "is_configured": true, 00:11:08.298 "data_offset": 2048, 00:11:08.298 "data_size": 63488 00:11:08.298 } 00:11:08.298 ] 00:11:08.299 }' 00:11:08.299 23:44:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:08.299 23:44:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:08.299 23:44:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:08.299 23:44:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:08.299 23:44:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:11:08.299 23:44:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:08.299 23:44:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:08.299 [2024-11-26 23:44:56.316567] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:11:08.299 [2024-11-26 23:44:56.328489] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 14336 offset_begin: 12288 offset_end: 18432 00:11:08.299 [2024-11-26 23:44:56.328890] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 14336 offset_begin: 12288 offset_end: 18432 00:11:08.557 [2024-11-26 23:44:56.429669] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:11:08.557 [2024-11-26 23:44:56.431399] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:08.557 [2024-11-26 23:44:56.431473] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:11:08.557 [2024-11-26 23:44:56.431491] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:11:08.557 [2024-11-26 23:44:56.448577] bdev_raid.c:1974:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 0 raid_ch: 0x60d0000026d0 00:11:08.557 23:44:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:08.557 23:44:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:11:08.557 23:44:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:08.557 23:44:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:08.557 23:44:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:08.557 23:44:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:08.557 23:44:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:11:08.557 23:44:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:08.557 23:44:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:08.557 23:44:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:08.557 23:44:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:08.557 23:44:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:08.557 23:44:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:08.557 23:44:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:08.557 23:44:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:08.557 23:44:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:08.557 23:44:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:08.557 "name": "raid_bdev1", 00:11:08.557 "uuid": "dfec84f0-3912-4aa3-b332-ee4e99f2a7bd", 00:11:08.557 "strip_size_kb": 0, 00:11:08.557 "state": "online", 00:11:08.557 "raid_level": "raid1", 00:11:08.557 "superblock": true, 00:11:08.557 "num_base_bdevs": 2, 00:11:08.557 "num_base_bdevs_discovered": 1, 00:11:08.557 "num_base_bdevs_operational": 1, 00:11:08.557 "base_bdevs_list": [ 00:11:08.557 { 00:11:08.557 "name": null, 00:11:08.557 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:08.557 "is_configured": false, 00:11:08.557 "data_offset": 0, 00:11:08.557 "data_size": 63488 00:11:08.557 }, 00:11:08.557 { 00:11:08.557 "name": "BaseBdev2", 00:11:08.557 "uuid": "da9693d2-7593-58b9-8d53-16093d346229", 00:11:08.557 "is_configured": true, 00:11:08.557 "data_offset": 2048, 00:11:08.557 "data_size": 63488 00:11:08.557 } 00:11:08.557 ] 00:11:08.557 }' 00:11:08.557 23:44:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:08.557 23:44:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:08.815 184.50 IOPS, 553.50 MiB/s [2024-11-26T23:44:56.947Z] 23:44:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:11:08.815 23:44:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:08.815 23:44:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:11:08.815 23:44:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:11:08.815 23:44:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:08.815 23:44:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:08.815 23:44:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:08.815 23:44:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:08.815 23:44:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:08.815 23:44:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:09.074 23:44:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:09.074 "name": "raid_bdev1", 00:11:09.074 "uuid": "dfec84f0-3912-4aa3-b332-ee4e99f2a7bd", 00:11:09.074 "strip_size_kb": 0, 00:11:09.074 "state": "online", 00:11:09.074 "raid_level": "raid1", 00:11:09.074 "superblock": true, 00:11:09.074 "num_base_bdevs": 2, 00:11:09.074 "num_base_bdevs_discovered": 1, 00:11:09.074 "num_base_bdevs_operational": 1, 00:11:09.074 "base_bdevs_list": [ 00:11:09.074 { 00:11:09.074 "name": null, 00:11:09.074 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:09.074 "is_configured": false, 00:11:09.074 "data_offset": 0, 00:11:09.074 "data_size": 63488 00:11:09.074 }, 00:11:09.074 { 00:11:09.074 "name": "BaseBdev2", 00:11:09.074 "uuid": "da9693d2-7593-58b9-8d53-16093d346229", 00:11:09.074 "is_configured": true, 00:11:09.074 "data_offset": 2048, 00:11:09.074 "data_size": 63488 00:11:09.074 } 00:11:09.074 ] 00:11:09.074 }' 00:11:09.074 23:44:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:09.074 23:44:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:11:09.074 23:44:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:09.074 23:44:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:11:09.074 23:44:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:11:09.074 23:44:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:09.074 23:44:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:09.074 [2024-11-26 23:44:57.006677] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:11:09.074 23:44:57 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:09.074 23:44:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@663 -- # sleep 1 00:11:09.074 [2024-11-26 23:44:57.039590] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002870 00:11:09.074 [2024-11-26 23:44:57.041472] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:11:09.074 [2024-11-26 23:44:57.154132] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:11:09.074 [2024-11-26 23:44:57.154612] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:11:09.333 [2024-11-26 23:44:57.272815] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:11:09.333 [2024-11-26 23:44:57.273093] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:11:09.592 [2024-11-26 23:44:57.703365] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:11:09.592 [2024-11-26 23:44:57.703729] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:11:09.850 180.00 IOPS, 540.00 MiB/s [2024-11-26T23:44:57.982Z] [2024-11-26 23:44:57.931023] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 14336 offset_begin: 12288 offset_end: 18432 00:11:09.850 [2024-11-26 23:44:57.931368] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 14336 offset_begin: 12288 offset_end: 18432 00:11:10.108 23:44:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:10.108 23:44:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:10.108 23:44:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:10.108 23:44:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:10.108 23:44:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:10.108 23:44:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:10.108 23:44:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:10.108 23:44:58 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:10.108 23:44:58 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:10.108 23:44:58 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:10.108 23:44:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:10.108 "name": "raid_bdev1", 00:11:10.108 "uuid": "dfec84f0-3912-4aa3-b332-ee4e99f2a7bd", 00:11:10.108 "strip_size_kb": 0, 00:11:10.108 "state": "online", 00:11:10.108 "raid_level": "raid1", 00:11:10.108 "superblock": true, 00:11:10.108 "num_base_bdevs": 2, 00:11:10.108 "num_base_bdevs_discovered": 2, 00:11:10.108 "num_base_bdevs_operational": 2, 00:11:10.108 "process": { 00:11:10.108 "type": "rebuild", 00:11:10.108 "target": "spare", 00:11:10.108 "progress": { 00:11:10.108 "blocks": 14336, 00:11:10.108 "percent": 22 00:11:10.108 } 00:11:10.108 }, 00:11:10.108 "base_bdevs_list": [ 00:11:10.108 { 00:11:10.108 "name": "spare", 00:11:10.108 "uuid": "1d4c4384-173d-51ef-ae81-2eb69239e686", 00:11:10.108 "is_configured": true, 00:11:10.108 "data_offset": 2048, 00:11:10.108 "data_size": 63488 00:11:10.108 }, 00:11:10.108 { 00:11:10.108 "name": "BaseBdev2", 00:11:10.108 "uuid": "da9693d2-7593-58b9-8d53-16093d346229", 00:11:10.108 "is_configured": true, 00:11:10.108 "data_offset": 2048, 00:11:10.108 "data_size": 63488 00:11:10.108 } 00:11:10.108 ] 00:11:10.108 }' 00:11:10.108 23:44:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:10.108 23:44:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:10.108 23:44:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:10.108 [2024-11-26 23:44:58.150307] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:11:10.108 [2024-11-26 23:44:58.150662] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:11:10.108 23:44:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:10.108 23:44:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@666 -- # '[' true = true ']' 00:11:10.108 23:44:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@666 -- # '[' = false ']' 00:11:10.108 /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh: line 666: [: =: unary operator expected 00:11:10.108 23:44:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=2 00:11:10.109 23:44:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:11:10.109 23:44:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@693 -- # '[' 2 -gt 2 ']' 00:11:10.109 23:44:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@706 -- # local timeout=330 00:11:10.109 23:44:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:11:10.109 23:44:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:10.109 23:44:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:10.109 23:44:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:10.109 23:44:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:10.109 23:44:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:10.109 23:44:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:10.109 23:44:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:10.109 23:44:58 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:10.109 23:44:58 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:10.109 23:44:58 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:10.368 23:44:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:10.368 "name": "raid_bdev1", 00:11:10.368 "uuid": "dfec84f0-3912-4aa3-b332-ee4e99f2a7bd", 00:11:10.368 "strip_size_kb": 0, 00:11:10.368 "state": "online", 00:11:10.368 "raid_level": "raid1", 00:11:10.368 "superblock": true, 00:11:10.368 "num_base_bdevs": 2, 00:11:10.368 "num_base_bdevs_discovered": 2, 00:11:10.368 "num_base_bdevs_operational": 2, 00:11:10.368 "process": { 00:11:10.368 "type": "rebuild", 00:11:10.368 "target": "spare", 00:11:10.368 "progress": { 00:11:10.368 "blocks": 16384, 00:11:10.368 "percent": 25 00:11:10.368 } 00:11:10.368 }, 00:11:10.368 "base_bdevs_list": [ 00:11:10.368 { 00:11:10.368 "name": "spare", 00:11:10.368 "uuid": "1d4c4384-173d-51ef-ae81-2eb69239e686", 00:11:10.368 "is_configured": true, 00:11:10.368 "data_offset": 2048, 00:11:10.368 "data_size": 63488 00:11:10.368 }, 00:11:10.368 { 00:11:10.368 "name": "BaseBdev2", 00:11:10.368 "uuid": "da9693d2-7593-58b9-8d53-16093d346229", 00:11:10.368 "is_configured": true, 00:11:10.368 "data_offset": 2048, 00:11:10.368 "data_size": 63488 00:11:10.368 } 00:11:10.368 ] 00:11:10.368 }' 00:11:10.368 23:44:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:10.368 23:44:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:10.368 23:44:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:10.368 23:44:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:10.368 23:44:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:11:10.627 [2024-11-26 23:44:58.573798] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 22528 offset_begin: 18432 offset_end: 24576 00:11:11.145 149.00 IOPS, 447.00 MiB/s [2024-11-26T23:44:59.277Z] [2024-11-26 23:44:59.056362] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 28672 offset_begin: 24576 offset_end: 30720 00:11:11.402 23:44:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:11:11.402 23:44:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:11.402 23:44:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:11.402 23:44:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:11.402 23:44:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:11.402 23:44:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:11.402 23:44:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:11.402 23:44:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:11.402 23:44:59 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:11.402 23:44:59 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:11.402 23:44:59 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:11.402 23:44:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:11.402 "name": "raid_bdev1", 00:11:11.402 "uuid": "dfec84f0-3912-4aa3-b332-ee4e99f2a7bd", 00:11:11.402 "strip_size_kb": 0, 00:11:11.402 "state": "online", 00:11:11.402 "raid_level": "raid1", 00:11:11.402 "superblock": true, 00:11:11.402 "num_base_bdevs": 2, 00:11:11.402 "num_base_bdevs_discovered": 2, 00:11:11.402 "num_base_bdevs_operational": 2, 00:11:11.402 "process": { 00:11:11.402 "type": "rebuild", 00:11:11.402 "target": "spare", 00:11:11.402 "progress": { 00:11:11.402 "blocks": 30720, 00:11:11.402 "percent": 48 00:11:11.402 } 00:11:11.402 }, 00:11:11.402 "base_bdevs_list": [ 00:11:11.402 { 00:11:11.402 "name": "spare", 00:11:11.402 "uuid": "1d4c4384-173d-51ef-ae81-2eb69239e686", 00:11:11.402 "is_configured": true, 00:11:11.402 "data_offset": 2048, 00:11:11.402 "data_size": 63488 00:11:11.402 }, 00:11:11.402 { 00:11:11.402 "name": "BaseBdev2", 00:11:11.402 "uuid": "da9693d2-7593-58b9-8d53-16093d346229", 00:11:11.402 "is_configured": true, 00:11:11.402 "data_offset": 2048, 00:11:11.402 "data_size": 63488 00:11:11.402 } 00:11:11.402 ] 00:11:11.402 }' 00:11:11.402 23:44:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:11.402 23:44:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:11.402 23:44:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:11.402 23:44:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:11.402 23:44:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:11:11.661 [2024-11-26 23:44:59.693618] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 38912 offset_begin: 36864 offset_end: 43008 00:11:11.919 129.00 IOPS, 387.00 MiB/s [2024-11-26T23:45:00.051Z] [2024-11-26 23:44:59.918859] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 40960 offset_begin: 36864 offset_end: 43008 00:11:12.519 23:45:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:11:12.519 23:45:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:12.519 23:45:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:12.519 23:45:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:12.519 23:45:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:12.519 23:45:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:12.519 23:45:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:12.519 23:45:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:12.519 23:45:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:12.519 23:45:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:12.519 23:45:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:12.519 23:45:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:12.519 "name": "raid_bdev1", 00:11:12.519 "uuid": "dfec84f0-3912-4aa3-b332-ee4e99f2a7bd", 00:11:12.520 "strip_size_kb": 0, 00:11:12.520 "state": "online", 00:11:12.520 "raid_level": "raid1", 00:11:12.520 "superblock": true, 00:11:12.520 "num_base_bdevs": 2, 00:11:12.520 "num_base_bdevs_discovered": 2, 00:11:12.520 "num_base_bdevs_operational": 2, 00:11:12.520 "process": { 00:11:12.520 "type": "rebuild", 00:11:12.520 "target": "spare", 00:11:12.520 "progress": { 00:11:12.520 "blocks": 49152, 00:11:12.520 "percent": 77 00:11:12.520 } 00:11:12.520 }, 00:11:12.520 "base_bdevs_list": [ 00:11:12.520 { 00:11:12.520 "name": "spare", 00:11:12.520 "uuid": "1d4c4384-173d-51ef-ae81-2eb69239e686", 00:11:12.520 "is_configured": true, 00:11:12.520 "data_offset": 2048, 00:11:12.520 "data_size": 63488 00:11:12.520 }, 00:11:12.520 { 00:11:12.520 "name": "BaseBdev2", 00:11:12.520 "uuid": "da9693d2-7593-58b9-8d53-16093d346229", 00:11:12.520 "is_configured": true, 00:11:12.520 "data_offset": 2048, 00:11:12.520 "data_size": 63488 00:11:12.520 } 00:11:12.520 ] 00:11:12.520 }' 00:11:12.520 23:45:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:12.520 23:45:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:12.520 23:45:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:12.520 23:45:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:12.520 23:45:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:11:12.520 [2024-11-26 23:45:00.581154] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 53248 offset_begin: 49152 offset_end: 55296 00:11:13.072 112.33 IOPS, 337.00 MiB/s [2024-11-26T23:45:01.204Z] [2024-11-26 23:45:00.997578] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 59392 offset_begin: 55296 offset_end: 61440 00:11:13.072 [2024-11-26 23:45:00.997924] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 59392 offset_begin: 55296 offset_end: 61440 00:11:13.332 [2024-11-26 23:45:01.230943] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:11:13.332 [2024-11-26 23:45:01.335802] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:11:13.332 [2024-11-26 23:45:01.337845] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:13.592 23:45:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:11:13.592 23:45:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:13.592 23:45:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:13.592 23:45:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:13.592 23:45:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:13.592 23:45:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:13.592 23:45:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:13.592 23:45:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:13.592 23:45:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:13.592 23:45:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:13.592 23:45:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:13.592 23:45:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:13.592 "name": "raid_bdev1", 00:11:13.592 "uuid": "dfec84f0-3912-4aa3-b332-ee4e99f2a7bd", 00:11:13.592 "strip_size_kb": 0, 00:11:13.592 "state": "online", 00:11:13.592 "raid_level": "raid1", 00:11:13.592 "superblock": true, 00:11:13.592 "num_base_bdevs": 2, 00:11:13.592 "num_base_bdevs_discovered": 2, 00:11:13.592 "num_base_bdevs_operational": 2, 00:11:13.592 "base_bdevs_list": [ 00:11:13.592 { 00:11:13.592 "name": "spare", 00:11:13.592 "uuid": "1d4c4384-173d-51ef-ae81-2eb69239e686", 00:11:13.592 "is_configured": true, 00:11:13.592 "data_offset": 2048, 00:11:13.592 "data_size": 63488 00:11:13.592 }, 00:11:13.592 { 00:11:13.592 "name": "BaseBdev2", 00:11:13.592 "uuid": "da9693d2-7593-58b9-8d53-16093d346229", 00:11:13.592 "is_configured": true, 00:11:13.592 "data_offset": 2048, 00:11:13.592 "data_size": 63488 00:11:13.592 } 00:11:13.592 ] 00:11:13.592 }' 00:11:13.592 23:45:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:13.592 23:45:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:11:13.592 23:45:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:13.592 23:45:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:11:13.592 23:45:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@709 -- # break 00:11:13.592 23:45:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:11:13.592 23:45:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:13.592 23:45:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:11:13.592 23:45:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:11:13.592 23:45:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:13.592 23:45:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:13.592 23:45:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:13.592 23:45:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:13.592 23:45:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:13.851 23:45:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:13.851 101.29 IOPS, 303.86 MiB/s [2024-11-26T23:45:01.983Z] 23:45:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:13.851 "name": "raid_bdev1", 00:11:13.851 "uuid": "dfec84f0-3912-4aa3-b332-ee4e99f2a7bd", 00:11:13.851 "strip_size_kb": 0, 00:11:13.851 "state": "online", 00:11:13.851 "raid_level": "raid1", 00:11:13.851 "superblock": true, 00:11:13.851 "num_base_bdevs": 2, 00:11:13.851 "num_base_bdevs_discovered": 2, 00:11:13.851 "num_base_bdevs_operational": 2, 00:11:13.851 "base_bdevs_list": [ 00:11:13.851 { 00:11:13.851 "name": "spare", 00:11:13.851 "uuid": "1d4c4384-173d-51ef-ae81-2eb69239e686", 00:11:13.851 "is_configured": true, 00:11:13.851 "data_offset": 2048, 00:11:13.851 "data_size": 63488 00:11:13.851 }, 00:11:13.851 { 00:11:13.851 "name": "BaseBdev2", 00:11:13.851 "uuid": "da9693d2-7593-58b9-8d53-16093d346229", 00:11:13.851 "is_configured": true, 00:11:13.851 "data_offset": 2048, 00:11:13.851 "data_size": 63488 00:11:13.851 } 00:11:13.851 ] 00:11:13.851 }' 00:11:13.851 23:45:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:13.851 23:45:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:11:13.851 23:45:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:13.852 23:45:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:11:13.852 23:45:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:11:13.852 23:45:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:13.852 23:45:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:13.852 23:45:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:13.852 23:45:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:13.852 23:45:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:11:13.852 23:45:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:13.852 23:45:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:13.852 23:45:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:13.852 23:45:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:13.852 23:45:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:13.852 23:45:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:13.852 23:45:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:13.852 23:45:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:13.852 23:45:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:13.852 23:45:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:13.852 "name": "raid_bdev1", 00:11:13.852 "uuid": "dfec84f0-3912-4aa3-b332-ee4e99f2a7bd", 00:11:13.852 "strip_size_kb": 0, 00:11:13.852 "state": "online", 00:11:13.852 "raid_level": "raid1", 00:11:13.852 "superblock": true, 00:11:13.852 "num_base_bdevs": 2, 00:11:13.852 "num_base_bdevs_discovered": 2, 00:11:13.852 "num_base_bdevs_operational": 2, 00:11:13.852 "base_bdevs_list": [ 00:11:13.852 { 00:11:13.852 "name": "spare", 00:11:13.852 "uuid": "1d4c4384-173d-51ef-ae81-2eb69239e686", 00:11:13.852 "is_configured": true, 00:11:13.852 "data_offset": 2048, 00:11:13.852 "data_size": 63488 00:11:13.852 }, 00:11:13.852 { 00:11:13.852 "name": "BaseBdev2", 00:11:13.852 "uuid": "da9693d2-7593-58b9-8d53-16093d346229", 00:11:13.852 "is_configured": true, 00:11:13.852 "data_offset": 2048, 00:11:13.852 "data_size": 63488 00:11:13.852 } 00:11:13.852 ] 00:11:13.852 }' 00:11:13.852 23:45:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:13.852 23:45:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:14.111 23:45:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:11:14.111 23:45:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:14.111 23:45:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:14.111 [2024-11-26 23:45:02.233930] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:11:14.111 [2024-11-26 23:45:02.233967] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:11:14.372 00:11:14.372 Latency(us) 00:11:14.372 [2024-11-26T23:45:02.504Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:11:14.372 Job: raid_bdev1 (Core Mask 0x1, workload: randrw, percentage: 50, depth: 2, IO size: 3145728) 00:11:14.372 raid_bdev1 : 7.56 96.99 290.98 0.00 0.00 13615.31 279.03 112183.90 00:11:14.372 [2024-11-26T23:45:02.504Z] =================================================================================================================== 00:11:14.372 [2024-11-26T23:45:02.504Z] Total : 96.99 290.98 0.00 0.00 13615.31 279.03 112183.90 00:11:14.372 [2024-11-26 23:45:02.312920] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:11:14.372 [2024-11-26 23:45:02.312969] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:14.372 [2024-11-26 23:45:02.313037] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:11:14.372 [2024-11-26 23:45:02.313059] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:11:14.372 { 00:11:14.372 "results": [ 00:11:14.372 { 00:11:14.372 "job": "raid_bdev1", 00:11:14.372 "core_mask": "0x1", 00:11:14.372 "workload": "randrw", 00:11:14.372 "percentage": 50, 00:11:14.372 "status": "finished", 00:11:14.372 "queue_depth": 2, 00:11:14.372 "io_size": 3145728, 00:11:14.372 "runtime": 7.557144, 00:11:14.372 "iops": 96.99431425416797, 00:11:14.372 "mibps": 290.9829427625039, 00:11:14.372 "io_failed": 0, 00:11:14.372 "io_timeout": 0, 00:11:14.372 "avg_latency_us": 13615.31463567203, 00:11:14.372 "min_latency_us": 279.0288209606987, 00:11:14.372 "max_latency_us": 112183.89519650655 00:11:14.372 } 00:11:14.372 ], 00:11:14.372 "core_count": 1 00:11:14.372 } 00:11:14.372 23:45:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:14.372 23:45:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:14.372 23:45:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:14.372 23:45:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@720 -- # jq length 00:11:14.372 23:45:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:14.372 23:45:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:14.372 23:45:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:11:14.372 23:45:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:11:14.372 23:45:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@723 -- # '[' true = true ']' 00:11:14.372 23:45:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@725 -- # nbd_start_disks /var/tmp/spdk.sock spare /dev/nbd0 00:11:14.372 23:45:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:11:14.372 23:45:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # bdev_list=('spare') 00:11:14.372 23:45:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:11:14.372 23:45:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:11:14.372 23:45:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:11:14.372 23:45:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@12 -- # local i 00:11:14.372 23:45:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:11:14.372 23:45:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:11:14.372 23:45:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd0 00:11:14.632 /dev/nbd0 00:11:14.632 23:45:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:11:14.632 23:45:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:11:14.632 23:45:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:11:14.632 23:45:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@873 -- # local i 00:11:14.632 23:45:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:11:14.632 23:45:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:11:14.632 23:45:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:11:14.632 23:45:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@877 -- # break 00:11:14.632 23:45:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:11:14.632 23:45:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:11:14.633 23:45:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:11:14.633 1+0 records in 00:11:14.633 1+0 records out 00:11:14.633 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000500998 s, 8.2 MB/s 00:11:14.633 23:45:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:14.633 23:45:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@890 -- # size=4096 00:11:14.633 23:45:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:14.633 23:45:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:11:14.633 23:45:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@893 -- # return 0 00:11:14.633 23:45:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:11:14.633 23:45:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:11:14.633 23:45:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@726 -- # for bdev in "${base_bdevs[@]:1}" 00:11:14.633 23:45:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@727 -- # '[' -z BaseBdev2 ']' 00:11:14.633 23:45:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@730 -- # nbd_start_disks /var/tmp/spdk.sock BaseBdev2 /dev/nbd1 00:11:14.633 23:45:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:11:14.633 23:45:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev2') 00:11:14.633 23:45:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:11:14.633 23:45:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd1') 00:11:14.633 23:45:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:11:14.633 23:45:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@12 -- # local i 00:11:14.633 23:45:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:11:14.633 23:45:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:11:14.633 23:45:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev2 /dev/nbd1 00:11:14.893 /dev/nbd1 00:11:14.893 23:45:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:11:14.893 23:45:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:11:14.893 23:45:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:11:14.893 23:45:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@873 -- # local i 00:11:14.893 23:45:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:11:14.893 23:45:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:11:14.893 23:45:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:11:14.893 23:45:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@877 -- # break 00:11:14.893 23:45:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:11:14.893 23:45:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:11:14.893 23:45:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:11:14.893 1+0 records in 00:11:14.893 1+0 records out 00:11:14.893 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000280304 s, 14.6 MB/s 00:11:14.893 23:45:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:14.893 23:45:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@890 -- # size=4096 00:11:14.893 23:45:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:14.893 23:45:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:11:14.893 23:45:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@893 -- # return 0 00:11:14.894 23:45:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:11:14.894 23:45:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:11:14.894 23:45:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@731 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:11:14.894 23:45:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@732 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd1 00:11:14.894 23:45:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:11:14.894 23:45:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd1') 00:11:14.894 23:45:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:11:14.894 23:45:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@51 -- # local i 00:11:14.894 23:45:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:11:14.894 23:45:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:11:15.154 23:45:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:11:15.154 23:45:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:11:15.154 23:45:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:11:15.154 23:45:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:11:15.154 23:45:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:11:15.154 23:45:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:11:15.154 23:45:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@41 -- # break 00:11:15.154 23:45:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@45 -- # return 0 00:11:15.154 23:45:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@734 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:11:15.154 23:45:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:11:15.154 23:45:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:11:15.154 23:45:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:11:15.154 23:45:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@51 -- # local i 00:11:15.154 23:45:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:11:15.154 23:45:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:11:15.414 23:45:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:11:15.414 23:45:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:11:15.414 23:45:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:11:15.414 23:45:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:11:15.414 23:45:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:11:15.414 23:45:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:11:15.414 23:45:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@41 -- # break 00:11:15.414 23:45:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@45 -- # return 0 00:11:15.414 23:45:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@743 -- # '[' true = true ']' 00:11:15.414 23:45:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@745 -- # rpc_cmd bdev_passthru_delete spare 00:11:15.414 23:45:03 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:15.414 23:45:03 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:15.414 23:45:03 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:15.414 23:45:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@746 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:11:15.414 23:45:03 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:15.414 23:45:03 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:15.414 [2024-11-26 23:45:03.377717] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:11:15.414 [2024-11-26 23:45:03.377779] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:15.414 [2024-11-26 23:45:03.377800] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009980 00:11:15.414 [2024-11-26 23:45:03.377810] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:15.414 [2024-11-26 23:45:03.380057] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:15.414 [2024-11-26 23:45:03.380134] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:11:15.414 [2024-11-26 23:45:03.380258] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:11:15.414 [2024-11-26 23:45:03.380340] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:11:15.414 [2024-11-26 23:45:03.380504] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:11:15.414 spare 00:11:15.414 23:45:03 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:15.414 23:45:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@747 -- # rpc_cmd bdev_wait_for_examine 00:11:15.414 23:45:03 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:15.414 23:45:03 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:15.414 [2024-11-26 23:45:03.480438] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001580 00:11:15.414 [2024-11-26 23:45:03.480473] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:11:15.414 [2024-11-26 23:45:03.480716] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000027720 00:11:15.414 [2024-11-26 23:45:03.480835] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001580 00:11:15.414 [2024-11-26 23:45:03.480847] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001580 00:11:15.414 [2024-11-26 23:45:03.480973] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:15.414 23:45:03 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:15.414 23:45:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@749 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:11:15.414 23:45:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:15.414 23:45:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:15.415 23:45:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:15.415 23:45:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:15.415 23:45:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:11:15.415 23:45:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:15.415 23:45:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:15.415 23:45:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:15.415 23:45:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:15.415 23:45:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:15.415 23:45:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:15.415 23:45:03 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:15.415 23:45:03 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:15.415 23:45:03 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:15.415 23:45:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:15.415 "name": "raid_bdev1", 00:11:15.415 "uuid": "dfec84f0-3912-4aa3-b332-ee4e99f2a7bd", 00:11:15.415 "strip_size_kb": 0, 00:11:15.415 "state": "online", 00:11:15.415 "raid_level": "raid1", 00:11:15.415 "superblock": true, 00:11:15.415 "num_base_bdevs": 2, 00:11:15.415 "num_base_bdevs_discovered": 2, 00:11:15.415 "num_base_bdevs_operational": 2, 00:11:15.415 "base_bdevs_list": [ 00:11:15.415 { 00:11:15.415 "name": "spare", 00:11:15.415 "uuid": "1d4c4384-173d-51ef-ae81-2eb69239e686", 00:11:15.415 "is_configured": true, 00:11:15.415 "data_offset": 2048, 00:11:15.415 "data_size": 63488 00:11:15.415 }, 00:11:15.415 { 00:11:15.415 "name": "BaseBdev2", 00:11:15.415 "uuid": "da9693d2-7593-58b9-8d53-16093d346229", 00:11:15.415 "is_configured": true, 00:11:15.415 "data_offset": 2048, 00:11:15.415 "data_size": 63488 00:11:15.415 } 00:11:15.415 ] 00:11:15.415 }' 00:11:15.415 23:45:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:15.415 23:45:03 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:15.984 23:45:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@750 -- # verify_raid_bdev_process raid_bdev1 none none 00:11:15.984 23:45:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:15.984 23:45:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:11:15.984 23:45:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:11:15.984 23:45:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:15.984 23:45:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:15.984 23:45:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:15.984 23:45:03 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:15.984 23:45:03 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:15.984 23:45:03 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:15.984 23:45:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:15.984 "name": "raid_bdev1", 00:11:15.984 "uuid": "dfec84f0-3912-4aa3-b332-ee4e99f2a7bd", 00:11:15.984 "strip_size_kb": 0, 00:11:15.984 "state": "online", 00:11:15.984 "raid_level": "raid1", 00:11:15.984 "superblock": true, 00:11:15.984 "num_base_bdevs": 2, 00:11:15.984 "num_base_bdevs_discovered": 2, 00:11:15.984 "num_base_bdevs_operational": 2, 00:11:15.984 "base_bdevs_list": [ 00:11:15.984 { 00:11:15.984 "name": "spare", 00:11:15.984 "uuid": "1d4c4384-173d-51ef-ae81-2eb69239e686", 00:11:15.984 "is_configured": true, 00:11:15.984 "data_offset": 2048, 00:11:15.984 "data_size": 63488 00:11:15.984 }, 00:11:15.984 { 00:11:15.984 "name": "BaseBdev2", 00:11:15.984 "uuid": "da9693d2-7593-58b9-8d53-16093d346229", 00:11:15.984 "is_configured": true, 00:11:15.984 "data_offset": 2048, 00:11:15.984 "data_size": 63488 00:11:15.984 } 00:11:15.984 ] 00:11:15.984 }' 00:11:15.984 23:45:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:15.984 23:45:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:11:15.984 23:45:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:15.984 23:45:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:11:15.984 23:45:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@751 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:15.984 23:45:04 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:15.984 23:45:04 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:15.984 23:45:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@751 -- # jq -r '.[].base_bdevs_list[0].name' 00:11:15.984 23:45:04 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:16.244 23:45:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@751 -- # [[ spare == \s\p\a\r\e ]] 00:11:16.244 23:45:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@754 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:11:16.244 23:45:04 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:16.244 23:45:04 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:16.244 [2024-11-26 23:45:04.140498] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:11:16.244 23:45:04 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:16.244 23:45:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@755 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:11:16.244 23:45:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:16.244 23:45:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:16.244 23:45:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:16.244 23:45:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:16.244 23:45:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:11:16.244 23:45:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:16.244 23:45:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:16.244 23:45:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:16.244 23:45:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:16.244 23:45:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:16.244 23:45:04 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:16.244 23:45:04 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:16.244 23:45:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:16.244 23:45:04 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:16.244 23:45:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:16.244 "name": "raid_bdev1", 00:11:16.244 "uuid": "dfec84f0-3912-4aa3-b332-ee4e99f2a7bd", 00:11:16.244 "strip_size_kb": 0, 00:11:16.244 "state": "online", 00:11:16.244 "raid_level": "raid1", 00:11:16.244 "superblock": true, 00:11:16.244 "num_base_bdevs": 2, 00:11:16.244 "num_base_bdevs_discovered": 1, 00:11:16.244 "num_base_bdevs_operational": 1, 00:11:16.244 "base_bdevs_list": [ 00:11:16.244 { 00:11:16.244 "name": null, 00:11:16.244 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:16.244 "is_configured": false, 00:11:16.244 "data_offset": 0, 00:11:16.244 "data_size": 63488 00:11:16.244 }, 00:11:16.244 { 00:11:16.244 "name": "BaseBdev2", 00:11:16.244 "uuid": "da9693d2-7593-58b9-8d53-16093d346229", 00:11:16.244 "is_configured": true, 00:11:16.244 "data_offset": 2048, 00:11:16.244 "data_size": 63488 00:11:16.244 } 00:11:16.244 ] 00:11:16.244 }' 00:11:16.244 23:45:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:16.244 23:45:04 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:16.504 23:45:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@756 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:11:16.504 23:45:04 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:16.504 23:45:04 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:16.504 [2024-11-26 23:45:04.607782] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:11:16.504 [2024-11-26 23:45:04.608058] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:11:16.504 [2024-11-26 23:45:04.608116] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:11:16.504 [2024-11-26 23:45:04.608187] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:11:16.504 [2024-11-26 23:45:04.613296] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000277f0 00:11:16.504 23:45:04 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:16.504 23:45:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@757 -- # sleep 1 00:11:16.504 [2024-11-26 23:45:04.615182] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:11:17.884 23:45:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@758 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:17.884 23:45:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:17.884 23:45:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:17.884 23:45:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:17.885 23:45:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:17.885 23:45:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:17.885 23:45:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:17.885 23:45:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:17.885 23:45:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:17.885 23:45:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:17.885 23:45:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:17.885 "name": "raid_bdev1", 00:11:17.885 "uuid": "dfec84f0-3912-4aa3-b332-ee4e99f2a7bd", 00:11:17.885 "strip_size_kb": 0, 00:11:17.885 "state": "online", 00:11:17.885 "raid_level": "raid1", 00:11:17.885 "superblock": true, 00:11:17.885 "num_base_bdevs": 2, 00:11:17.885 "num_base_bdevs_discovered": 2, 00:11:17.885 "num_base_bdevs_operational": 2, 00:11:17.885 "process": { 00:11:17.885 "type": "rebuild", 00:11:17.885 "target": "spare", 00:11:17.885 "progress": { 00:11:17.885 "blocks": 20480, 00:11:17.885 "percent": 32 00:11:17.885 } 00:11:17.885 }, 00:11:17.885 "base_bdevs_list": [ 00:11:17.885 { 00:11:17.885 "name": "spare", 00:11:17.885 "uuid": "1d4c4384-173d-51ef-ae81-2eb69239e686", 00:11:17.885 "is_configured": true, 00:11:17.885 "data_offset": 2048, 00:11:17.885 "data_size": 63488 00:11:17.885 }, 00:11:17.885 { 00:11:17.885 "name": "BaseBdev2", 00:11:17.885 "uuid": "da9693d2-7593-58b9-8d53-16093d346229", 00:11:17.885 "is_configured": true, 00:11:17.885 "data_offset": 2048, 00:11:17.885 "data_size": 63488 00:11:17.885 } 00:11:17.885 ] 00:11:17.885 }' 00:11:17.885 23:45:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:17.885 23:45:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:17.885 23:45:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:17.885 23:45:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:17.885 23:45:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@761 -- # rpc_cmd bdev_passthru_delete spare 00:11:17.885 23:45:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:17.885 23:45:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:17.885 [2024-11-26 23:45:05.779474] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:11:17.885 [2024-11-26 23:45:05.819325] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:11:17.885 [2024-11-26 23:45:05.819463] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:17.885 [2024-11-26 23:45:05.819503] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:11:17.885 [2024-11-26 23:45:05.819525] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:11:17.885 23:45:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:17.885 23:45:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@762 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:11:17.885 23:45:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:17.885 23:45:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:17.885 23:45:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:17.885 23:45:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:17.885 23:45:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:11:17.885 23:45:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:17.885 23:45:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:17.885 23:45:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:17.885 23:45:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:17.885 23:45:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:17.885 23:45:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:17.885 23:45:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:17.885 23:45:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:17.885 23:45:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:17.885 23:45:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:17.885 "name": "raid_bdev1", 00:11:17.885 "uuid": "dfec84f0-3912-4aa3-b332-ee4e99f2a7bd", 00:11:17.885 "strip_size_kb": 0, 00:11:17.885 "state": "online", 00:11:17.885 "raid_level": "raid1", 00:11:17.885 "superblock": true, 00:11:17.885 "num_base_bdevs": 2, 00:11:17.885 "num_base_bdevs_discovered": 1, 00:11:17.885 "num_base_bdevs_operational": 1, 00:11:17.885 "base_bdevs_list": [ 00:11:17.885 { 00:11:17.885 "name": null, 00:11:17.885 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:17.885 "is_configured": false, 00:11:17.885 "data_offset": 0, 00:11:17.885 "data_size": 63488 00:11:17.885 }, 00:11:17.885 { 00:11:17.885 "name": "BaseBdev2", 00:11:17.885 "uuid": "da9693d2-7593-58b9-8d53-16093d346229", 00:11:17.885 "is_configured": true, 00:11:17.885 "data_offset": 2048, 00:11:17.885 "data_size": 63488 00:11:17.885 } 00:11:17.885 ] 00:11:17.885 }' 00:11:17.885 23:45:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:17.885 23:45:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:18.143 23:45:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@763 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:11:18.143 23:45:06 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:18.143 23:45:06 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:18.143 [2024-11-26 23:45:06.271672] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:11:18.143 [2024-11-26 23:45:06.271791] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:18.143 [2024-11-26 23:45:06.271830] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:11:18.143 [2024-11-26 23:45:06.271859] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:18.143 [2024-11-26 23:45:06.272304] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:18.143 [2024-11-26 23:45:06.272373] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:11:18.143 [2024-11-26 23:45:06.272490] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:11:18.143 [2024-11-26 23:45:06.272530] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:11:18.143 [2024-11-26 23:45:06.272572] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:11:18.143 [2024-11-26 23:45:06.272651] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:11:18.402 [2024-11-26 23:45:06.277821] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000278c0 00:11:18.402 spare 00:11:18.402 23:45:06 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:18.402 23:45:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@764 -- # sleep 1 00:11:18.402 [2024-11-26 23:45:06.279761] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:11:19.341 23:45:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@765 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:19.341 23:45:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:19.341 23:45:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:19.341 23:45:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:19.341 23:45:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:19.341 23:45:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:19.341 23:45:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:19.341 23:45:07 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:19.341 23:45:07 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:19.341 23:45:07 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:19.341 23:45:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:19.341 "name": "raid_bdev1", 00:11:19.341 "uuid": "dfec84f0-3912-4aa3-b332-ee4e99f2a7bd", 00:11:19.341 "strip_size_kb": 0, 00:11:19.341 "state": "online", 00:11:19.341 "raid_level": "raid1", 00:11:19.341 "superblock": true, 00:11:19.341 "num_base_bdevs": 2, 00:11:19.341 "num_base_bdevs_discovered": 2, 00:11:19.341 "num_base_bdevs_operational": 2, 00:11:19.341 "process": { 00:11:19.341 "type": "rebuild", 00:11:19.341 "target": "spare", 00:11:19.341 "progress": { 00:11:19.341 "blocks": 20480, 00:11:19.341 "percent": 32 00:11:19.341 } 00:11:19.341 }, 00:11:19.341 "base_bdevs_list": [ 00:11:19.341 { 00:11:19.341 "name": "spare", 00:11:19.341 "uuid": "1d4c4384-173d-51ef-ae81-2eb69239e686", 00:11:19.341 "is_configured": true, 00:11:19.341 "data_offset": 2048, 00:11:19.341 "data_size": 63488 00:11:19.341 }, 00:11:19.341 { 00:11:19.341 "name": "BaseBdev2", 00:11:19.341 "uuid": "da9693d2-7593-58b9-8d53-16093d346229", 00:11:19.341 "is_configured": true, 00:11:19.341 "data_offset": 2048, 00:11:19.341 "data_size": 63488 00:11:19.341 } 00:11:19.341 ] 00:11:19.341 }' 00:11:19.341 23:45:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:19.341 23:45:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:19.341 23:45:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:19.341 23:45:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:19.341 23:45:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@768 -- # rpc_cmd bdev_passthru_delete spare 00:11:19.341 23:45:07 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:19.341 23:45:07 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:19.341 [2024-11-26 23:45:07.444030] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:11:19.601 [2024-11-26 23:45:07.483899] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:11:19.601 [2024-11-26 23:45:07.484019] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:19.601 [2024-11-26 23:45:07.484056] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:11:19.601 [2024-11-26 23:45:07.484079] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:11:19.601 23:45:07 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:19.601 23:45:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@769 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:11:19.601 23:45:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:19.601 23:45:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:19.601 23:45:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:19.601 23:45:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:19.601 23:45:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:11:19.601 23:45:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:19.601 23:45:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:19.601 23:45:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:19.601 23:45:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:19.601 23:45:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:19.601 23:45:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:19.601 23:45:07 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:19.601 23:45:07 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:19.601 23:45:07 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:19.601 23:45:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:19.601 "name": "raid_bdev1", 00:11:19.601 "uuid": "dfec84f0-3912-4aa3-b332-ee4e99f2a7bd", 00:11:19.601 "strip_size_kb": 0, 00:11:19.601 "state": "online", 00:11:19.601 "raid_level": "raid1", 00:11:19.601 "superblock": true, 00:11:19.601 "num_base_bdevs": 2, 00:11:19.601 "num_base_bdevs_discovered": 1, 00:11:19.601 "num_base_bdevs_operational": 1, 00:11:19.601 "base_bdevs_list": [ 00:11:19.601 { 00:11:19.601 "name": null, 00:11:19.601 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:19.601 "is_configured": false, 00:11:19.601 "data_offset": 0, 00:11:19.601 "data_size": 63488 00:11:19.601 }, 00:11:19.601 { 00:11:19.601 "name": "BaseBdev2", 00:11:19.601 "uuid": "da9693d2-7593-58b9-8d53-16093d346229", 00:11:19.601 "is_configured": true, 00:11:19.601 "data_offset": 2048, 00:11:19.601 "data_size": 63488 00:11:19.601 } 00:11:19.601 ] 00:11:19.601 }' 00:11:19.601 23:45:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:19.601 23:45:07 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:19.860 23:45:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@770 -- # verify_raid_bdev_process raid_bdev1 none none 00:11:19.860 23:45:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:19.860 23:45:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:11:19.860 23:45:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:11:19.860 23:45:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:19.860 23:45:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:19.860 23:45:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:19.860 23:45:07 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:19.860 23:45:07 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:20.120 23:45:07 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:20.120 23:45:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:20.120 "name": "raid_bdev1", 00:11:20.120 "uuid": "dfec84f0-3912-4aa3-b332-ee4e99f2a7bd", 00:11:20.120 "strip_size_kb": 0, 00:11:20.120 "state": "online", 00:11:20.120 "raid_level": "raid1", 00:11:20.120 "superblock": true, 00:11:20.120 "num_base_bdevs": 2, 00:11:20.120 "num_base_bdevs_discovered": 1, 00:11:20.120 "num_base_bdevs_operational": 1, 00:11:20.120 "base_bdevs_list": [ 00:11:20.120 { 00:11:20.120 "name": null, 00:11:20.120 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:20.120 "is_configured": false, 00:11:20.120 "data_offset": 0, 00:11:20.120 "data_size": 63488 00:11:20.120 }, 00:11:20.120 { 00:11:20.120 "name": "BaseBdev2", 00:11:20.120 "uuid": "da9693d2-7593-58b9-8d53-16093d346229", 00:11:20.120 "is_configured": true, 00:11:20.120 "data_offset": 2048, 00:11:20.120 "data_size": 63488 00:11:20.120 } 00:11:20.120 ] 00:11:20.120 }' 00:11:20.120 23:45:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:20.120 23:45:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:11:20.120 23:45:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:20.120 23:45:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:11:20.120 23:45:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@773 -- # rpc_cmd bdev_passthru_delete BaseBdev1 00:11:20.120 23:45:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:20.120 23:45:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:20.120 23:45:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:20.120 23:45:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@774 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:11:20.120 23:45:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:20.120 23:45:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:20.120 [2024-11-26 23:45:08.116063] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:11:20.120 [2024-11-26 23:45:08.116149] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:20.120 [2024-11-26 23:45:08.116171] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a880 00:11:20.120 [2024-11-26 23:45:08.116184] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:20.120 [2024-11-26 23:45:08.116671] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:20.120 [2024-11-26 23:45:08.116699] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:11:20.120 [2024-11-26 23:45:08.116781] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:11:20.120 [2024-11-26 23:45:08.116813] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:11:20.120 [2024-11-26 23:45:08.116822] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:11:20.120 [2024-11-26 23:45:08.116836] bdev_raid.c:3894:raid_bdev_examine_done: *ERROR*: Failed to examine bdev BaseBdev1: Invalid argument 00:11:20.120 BaseBdev1 00:11:20.120 23:45:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:20.120 23:45:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@775 -- # sleep 1 00:11:21.058 23:45:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@776 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:11:21.058 23:45:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:21.058 23:45:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:21.058 23:45:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:21.058 23:45:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:21.058 23:45:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:11:21.058 23:45:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:21.058 23:45:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:21.058 23:45:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:21.058 23:45:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:21.058 23:45:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:21.058 23:45:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:21.058 23:45:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:21.058 23:45:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:21.058 23:45:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:21.058 23:45:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:21.058 "name": "raid_bdev1", 00:11:21.058 "uuid": "dfec84f0-3912-4aa3-b332-ee4e99f2a7bd", 00:11:21.058 "strip_size_kb": 0, 00:11:21.058 "state": "online", 00:11:21.058 "raid_level": "raid1", 00:11:21.058 "superblock": true, 00:11:21.058 "num_base_bdevs": 2, 00:11:21.058 "num_base_bdevs_discovered": 1, 00:11:21.058 "num_base_bdevs_operational": 1, 00:11:21.058 "base_bdevs_list": [ 00:11:21.058 { 00:11:21.058 "name": null, 00:11:21.058 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:21.058 "is_configured": false, 00:11:21.058 "data_offset": 0, 00:11:21.058 "data_size": 63488 00:11:21.058 }, 00:11:21.058 { 00:11:21.058 "name": "BaseBdev2", 00:11:21.058 "uuid": "da9693d2-7593-58b9-8d53-16093d346229", 00:11:21.058 "is_configured": true, 00:11:21.058 "data_offset": 2048, 00:11:21.058 "data_size": 63488 00:11:21.058 } 00:11:21.058 ] 00:11:21.058 }' 00:11:21.058 23:45:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:21.058 23:45:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:21.625 23:45:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@777 -- # verify_raid_bdev_process raid_bdev1 none none 00:11:21.625 23:45:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:21.625 23:45:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:11:21.625 23:45:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:11:21.625 23:45:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:21.625 23:45:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:21.625 23:45:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:21.625 23:45:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:21.625 23:45:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:21.625 23:45:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:21.625 23:45:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:21.625 "name": "raid_bdev1", 00:11:21.625 "uuid": "dfec84f0-3912-4aa3-b332-ee4e99f2a7bd", 00:11:21.625 "strip_size_kb": 0, 00:11:21.625 "state": "online", 00:11:21.625 "raid_level": "raid1", 00:11:21.625 "superblock": true, 00:11:21.625 "num_base_bdevs": 2, 00:11:21.625 "num_base_bdevs_discovered": 1, 00:11:21.625 "num_base_bdevs_operational": 1, 00:11:21.625 "base_bdevs_list": [ 00:11:21.625 { 00:11:21.625 "name": null, 00:11:21.625 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:21.625 "is_configured": false, 00:11:21.625 "data_offset": 0, 00:11:21.625 "data_size": 63488 00:11:21.625 }, 00:11:21.625 { 00:11:21.625 "name": "BaseBdev2", 00:11:21.625 "uuid": "da9693d2-7593-58b9-8d53-16093d346229", 00:11:21.625 "is_configured": true, 00:11:21.625 "data_offset": 2048, 00:11:21.625 "data_size": 63488 00:11:21.625 } 00:11:21.625 ] 00:11:21.625 }' 00:11:21.625 23:45:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:21.625 23:45:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:11:21.625 23:45:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:21.625 23:45:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:11:21.625 23:45:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@778 -- # NOT rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:11:21.625 23:45:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@652 -- # local es=0 00:11:21.625 23:45:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:11:21.625 23:45:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:11:21.625 23:45:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:11:21.625 23:45:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:11:21.625 23:45:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:11:21.625 23:45:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:11:21.625 23:45:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:21.625 23:45:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:21.625 [2024-11-26 23:45:09.741707] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:11:21.626 [2024-11-26 23:45:09.741941] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:11:21.626 [2024-11-26 23:45:09.742002] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:11:21.626 request: 00:11:21.626 { 00:11:21.626 "base_bdev": "BaseBdev1", 00:11:21.626 "raid_bdev": "raid_bdev1", 00:11:21.626 "method": "bdev_raid_add_base_bdev", 00:11:21.626 "req_id": 1 00:11:21.626 } 00:11:21.626 Got JSON-RPC error response 00:11:21.626 response: 00:11:21.626 { 00:11:21.626 "code": -22, 00:11:21.626 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:11:21.626 } 00:11:21.626 23:45:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:11:21.626 23:45:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@655 -- # es=1 00:11:21.626 23:45:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:11:21.626 23:45:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:11:21.626 23:45:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:11:21.626 23:45:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@779 -- # sleep 1 00:11:23.007 23:45:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@780 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:11:23.007 23:45:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:23.007 23:45:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:23.007 23:45:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:23.007 23:45:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:23.007 23:45:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:11:23.007 23:45:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:23.007 23:45:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:23.007 23:45:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:23.007 23:45:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:23.007 23:45:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:23.007 23:45:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:23.007 23:45:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:23.007 23:45:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:23.007 23:45:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:23.007 23:45:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:23.007 "name": "raid_bdev1", 00:11:23.007 "uuid": "dfec84f0-3912-4aa3-b332-ee4e99f2a7bd", 00:11:23.007 "strip_size_kb": 0, 00:11:23.007 "state": "online", 00:11:23.007 "raid_level": "raid1", 00:11:23.007 "superblock": true, 00:11:23.007 "num_base_bdevs": 2, 00:11:23.007 "num_base_bdevs_discovered": 1, 00:11:23.007 "num_base_bdevs_operational": 1, 00:11:23.007 "base_bdevs_list": [ 00:11:23.007 { 00:11:23.007 "name": null, 00:11:23.007 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:23.007 "is_configured": false, 00:11:23.007 "data_offset": 0, 00:11:23.007 "data_size": 63488 00:11:23.007 }, 00:11:23.007 { 00:11:23.007 "name": "BaseBdev2", 00:11:23.007 "uuid": "da9693d2-7593-58b9-8d53-16093d346229", 00:11:23.007 "is_configured": true, 00:11:23.007 "data_offset": 2048, 00:11:23.007 "data_size": 63488 00:11:23.007 } 00:11:23.007 ] 00:11:23.007 }' 00:11:23.007 23:45:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:23.007 23:45:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:23.267 23:45:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@781 -- # verify_raid_bdev_process raid_bdev1 none none 00:11:23.267 23:45:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:23.267 23:45:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:11:23.267 23:45:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:11:23.267 23:45:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:23.267 23:45:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:23.267 23:45:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:23.267 23:45:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:23.267 23:45:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:23.267 23:45:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:23.267 23:45:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:23.267 "name": "raid_bdev1", 00:11:23.267 "uuid": "dfec84f0-3912-4aa3-b332-ee4e99f2a7bd", 00:11:23.267 "strip_size_kb": 0, 00:11:23.267 "state": "online", 00:11:23.267 "raid_level": "raid1", 00:11:23.267 "superblock": true, 00:11:23.267 "num_base_bdevs": 2, 00:11:23.267 "num_base_bdevs_discovered": 1, 00:11:23.267 "num_base_bdevs_operational": 1, 00:11:23.267 "base_bdevs_list": [ 00:11:23.267 { 00:11:23.267 "name": null, 00:11:23.267 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:23.267 "is_configured": false, 00:11:23.267 "data_offset": 0, 00:11:23.267 "data_size": 63488 00:11:23.267 }, 00:11:23.267 { 00:11:23.267 "name": "BaseBdev2", 00:11:23.267 "uuid": "da9693d2-7593-58b9-8d53-16093d346229", 00:11:23.267 "is_configured": true, 00:11:23.267 "data_offset": 2048, 00:11:23.267 "data_size": 63488 00:11:23.267 } 00:11:23.267 ] 00:11:23.267 }' 00:11:23.267 23:45:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:23.267 23:45:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:11:23.267 23:45:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:23.267 23:45:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:11:23.267 23:45:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@784 -- # killprocess 87201 00:11:23.267 23:45:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@954 -- # '[' -z 87201 ']' 00:11:23.267 23:45:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@958 -- # kill -0 87201 00:11:23.267 23:45:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@959 -- # uname 00:11:23.267 23:45:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:11:23.267 23:45:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 87201 00:11:23.267 killing process with pid 87201 00:11:23.267 Received shutdown signal, test time was about 16.652228 seconds 00:11:23.267 00:11:23.267 Latency(us) 00:11:23.267 [2024-11-26T23:45:11.399Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:11:23.267 [2024-11-26T23:45:11.399Z] =================================================================================================================== 00:11:23.267 [2024-11-26T23:45:11.399Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:11:23.267 23:45:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:11:23.267 23:45:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:11:23.267 23:45:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@972 -- # echo 'killing process with pid 87201' 00:11:23.267 23:45:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@973 -- # kill 87201 00:11:23.267 [2024-11-26 23:45:11.387313] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:11:23.267 [2024-11-26 23:45:11.387469] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:11:23.267 [2024-11-26 23:45:11.387533] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to fr 23:45:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@978 -- # wait 87201 00:11:23.267 ee all in destruct 00:11:23.267 [2024-11-26 23:45:11.387547] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state offline 00:11:23.527 [2024-11-26 23:45:11.413704] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:11:23.527 ************************************ 00:11:23.528 END TEST raid_rebuild_test_sb_io 00:11:23.528 ************************************ 00:11:23.528 23:45:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@786 -- # return 0 00:11:23.528 00:11:23.528 real 0m18.486s 00:11:23.528 user 0m24.481s 00:11:23.528 sys 0m2.200s 00:11:23.528 23:45:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@1130 -- # xtrace_disable 00:11:23.528 23:45:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:23.788 23:45:11 bdev_raid -- bdev/bdev_raid.sh@977 -- # for n in 2 4 00:11:23.788 23:45:11 bdev_raid -- bdev/bdev_raid.sh@978 -- # run_test raid_rebuild_test raid_rebuild_test raid1 4 false false true 00:11:23.788 23:45:11 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:11:23.788 23:45:11 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:11:23.788 23:45:11 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:11:23.788 ************************************ 00:11:23.788 START TEST raid_rebuild_test 00:11:23.788 ************************************ 00:11:23.788 23:45:11 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@1129 -- # raid_rebuild_test raid1 4 false false true 00:11:23.788 23:45:11 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:11:23.788 23:45:11 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=4 00:11:23.788 23:45:11 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@571 -- # local superblock=false 00:11:23.788 23:45:11 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:11:23.788 23:45:11 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@573 -- # local verify=true 00:11:23.788 23:45:11 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:11:23.788 23:45:11 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:11:23.788 23:45:11 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:11:23.788 23:45:11 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:11:23.788 23:45:11 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:11:23.788 23:45:11 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:11:23.788 23:45:11 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:11:23.788 23:45:11 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:11:23.788 23:45:11 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev3 00:11:23.788 23:45:11 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:11:23.788 23:45:11 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:11:23.788 23:45:11 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev4 00:11:23.788 23:45:11 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:11:23.788 23:45:11 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:11:23.788 23:45:11 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:11:23.788 23:45:11 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:11:23.788 23:45:11 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:11:23.788 23:45:11 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@576 -- # local strip_size 00:11:23.788 23:45:11 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@577 -- # local create_arg 00:11:23.788 23:45:11 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:11:23.788 23:45:11 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@579 -- # local data_offset 00:11:23.788 23:45:11 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:11:23.788 23:45:11 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:11:23.788 23:45:11 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@592 -- # '[' false = true ']' 00:11:23.788 23:45:11 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@597 -- # raid_pid=87876 00:11:23.788 23:45:11 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:11:23.788 23:45:11 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@598 -- # waitforlisten 87876 00:11:23.788 23:45:11 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@835 -- # '[' -z 87876 ']' 00:11:23.788 23:45:11 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:11:23.788 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:11:23.788 23:45:11 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:11:23.788 23:45:11 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:11:23.788 23:45:11 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:11:23.788 23:45:11 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:23.788 I/O size of 3145728 is greater than zero copy threshold (65536). 00:11:23.788 Zero copy mechanism will not be used. 00:11:23.788 [2024-11-26 23:45:11.789818] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:11:23.788 [2024-11-26 23:45:11.789979] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid87876 ] 00:11:24.049 [2024-11-26 23:45:11.946086] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:11:24.049 [2024-11-26 23:45:11.971975] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:11:24.049 [2024-11-26 23:45:12.016088] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:24.049 [2024-11-26 23:45:12.016204] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:24.619 23:45:12 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:11:24.619 23:45:12 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@868 -- # return 0 00:11:24.619 23:45:12 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:11:24.619 23:45:12 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:11:24.619 23:45:12 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:24.619 23:45:12 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:24.619 BaseBdev1_malloc 00:11:24.619 23:45:12 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:24.619 23:45:12 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:11:24.619 23:45:12 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:24.619 23:45:12 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:24.619 [2024-11-26 23:45:12.651874] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:11:24.619 [2024-11-26 23:45:12.651941] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:24.619 [2024-11-26 23:45:12.651988] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:11:24.619 [2024-11-26 23:45:12.652007] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:24.619 [2024-11-26 23:45:12.654059] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:24.619 [2024-11-26 23:45:12.654095] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:11:24.619 BaseBdev1 00:11:24.619 23:45:12 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:24.619 23:45:12 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:11:24.619 23:45:12 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:11:24.619 23:45:12 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:24.619 23:45:12 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:24.619 BaseBdev2_malloc 00:11:24.619 23:45:12 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:24.619 23:45:12 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:11:24.619 23:45:12 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:24.619 23:45:12 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:24.619 [2024-11-26 23:45:12.680122] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:11:24.619 [2024-11-26 23:45:12.680177] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:24.619 [2024-11-26 23:45:12.680214] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:11:24.619 [2024-11-26 23:45:12.680222] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:24.619 [2024-11-26 23:45:12.682238] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:24.619 [2024-11-26 23:45:12.682279] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:11:24.619 BaseBdev2 00:11:24.619 23:45:12 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:24.619 23:45:12 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:11:24.619 23:45:12 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:11:24.619 23:45:12 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:24.619 23:45:12 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:24.619 BaseBdev3_malloc 00:11:24.619 23:45:12 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:24.619 23:45:12 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev3_malloc -p BaseBdev3 00:11:24.619 23:45:12 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:24.619 23:45:12 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:24.620 [2024-11-26 23:45:12.708319] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev3_malloc 00:11:24.620 [2024-11-26 23:45:12.708476] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:24.620 [2024-11-26 23:45:12.708518] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:11:24.620 [2024-11-26 23:45:12.708527] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:24.620 [2024-11-26 23:45:12.710561] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:24.620 [2024-11-26 23:45:12.710596] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:11:24.620 BaseBdev3 00:11:24.620 23:45:12 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:24.620 23:45:12 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:11:24.620 23:45:12 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:11:24.620 23:45:12 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:24.620 23:45:12 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:24.879 BaseBdev4_malloc 00:11:24.879 23:45:12 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:24.879 23:45:12 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev4_malloc -p BaseBdev4 00:11:24.879 23:45:12 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:24.879 23:45:12 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:24.879 [2024-11-26 23:45:12.756797] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev4_malloc 00:11:24.879 [2024-11-26 23:45:12.756901] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:24.879 [2024-11-26 23:45:12.756949] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:11:24.879 [2024-11-26 23:45:12.756970] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:24.879 [2024-11-26 23:45:12.760616] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:24.879 [2024-11-26 23:45:12.760666] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:11:24.879 BaseBdev4 00:11:24.879 23:45:12 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:24.879 23:45:12 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:11:24.879 23:45:12 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:24.879 23:45:12 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:24.879 spare_malloc 00:11:24.879 23:45:12 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:24.879 23:45:12 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:11:24.879 23:45:12 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:24.879 23:45:12 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:24.879 spare_delay 00:11:24.879 23:45:12 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:24.879 23:45:12 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:11:24.879 23:45:12 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:24.879 23:45:12 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:24.879 [2024-11-26 23:45:12.797686] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:11:24.879 [2024-11-26 23:45:12.797733] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:24.879 [2024-11-26 23:45:12.797766] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009c80 00:11:24.879 [2024-11-26 23:45:12.797775] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:24.879 [2024-11-26 23:45:12.799901] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:24.879 [2024-11-26 23:45:12.799970] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:11:24.879 spare 00:11:24.879 23:45:12 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:24.879 23:45:12 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 00:11:24.879 23:45:12 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:24.879 23:45:12 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:24.879 [2024-11-26 23:45:12.809741] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:11:24.879 [2024-11-26 23:45:12.811495] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:11:24.879 [2024-11-26 23:45:12.811555] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:11:24.879 [2024-11-26 23:45:12.811602] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:11:24.879 [2024-11-26 23:45:12.811678] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:11:24.879 [2024-11-26 23:45:12.811687] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:11:24.879 [2024-11-26 23:45:12.811927] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:11:24.879 [2024-11-26 23:45:12.812067] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:11:24.879 [2024-11-26 23:45:12.812078] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:11:24.879 [2024-11-26 23:45:12.812190] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:24.879 23:45:12 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:24.879 23:45:12 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:11:24.879 23:45:12 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:24.879 23:45:12 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:24.879 23:45:12 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:24.879 23:45:12 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:24.879 23:45:12 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:24.879 23:45:12 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:24.879 23:45:12 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:24.879 23:45:12 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:24.879 23:45:12 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:24.879 23:45:12 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:24.879 23:45:12 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:24.879 23:45:12 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:24.879 23:45:12 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:24.879 23:45:12 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:24.879 23:45:12 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:24.879 "name": "raid_bdev1", 00:11:24.879 "uuid": "49c74909-b545-42a5-8fa7-cf975caea55f", 00:11:24.879 "strip_size_kb": 0, 00:11:24.879 "state": "online", 00:11:24.879 "raid_level": "raid1", 00:11:24.879 "superblock": false, 00:11:24.879 "num_base_bdevs": 4, 00:11:24.879 "num_base_bdevs_discovered": 4, 00:11:24.879 "num_base_bdevs_operational": 4, 00:11:24.879 "base_bdevs_list": [ 00:11:24.879 { 00:11:24.879 "name": "BaseBdev1", 00:11:24.879 "uuid": "66b920ae-f082-525b-adff-9a5812a3d705", 00:11:24.879 "is_configured": true, 00:11:24.879 "data_offset": 0, 00:11:24.879 "data_size": 65536 00:11:24.879 }, 00:11:24.879 { 00:11:24.879 "name": "BaseBdev2", 00:11:24.879 "uuid": "da29134b-b34b-5dc6-9377-4447cd1702f1", 00:11:24.879 "is_configured": true, 00:11:24.879 "data_offset": 0, 00:11:24.879 "data_size": 65536 00:11:24.879 }, 00:11:24.879 { 00:11:24.880 "name": "BaseBdev3", 00:11:24.880 "uuid": "6a57c1b8-6a06-5b0e-b911-acaed11c675f", 00:11:24.880 "is_configured": true, 00:11:24.880 "data_offset": 0, 00:11:24.880 "data_size": 65536 00:11:24.880 }, 00:11:24.880 { 00:11:24.880 "name": "BaseBdev4", 00:11:24.880 "uuid": "0f7413ce-1741-52fe-8fa6-07c7944b623f", 00:11:24.880 "is_configured": true, 00:11:24.880 "data_offset": 0, 00:11:24.880 "data_size": 65536 00:11:24.880 } 00:11:24.880 ] 00:11:24.880 }' 00:11:24.880 23:45:12 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:24.880 23:45:12 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:25.139 23:45:13 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:11:25.139 23:45:13 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:11:25.139 23:45:13 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:25.139 23:45:13 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:25.139 [2024-11-26 23:45:13.253395] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:11:25.397 23:45:13 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:25.397 23:45:13 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=65536 00:11:25.397 23:45:13 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:11:25.397 23:45:13 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:25.397 23:45:13 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:25.397 23:45:13 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:25.397 23:45:13 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:25.397 23:45:13 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@619 -- # data_offset=0 00:11:25.397 23:45:13 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:11:25.397 23:45:13 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:11:25.397 23:45:13 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:11:25.397 23:45:13 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:11:25.397 23:45:13 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:11:25.397 23:45:13 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:11:25.397 23:45:13 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@10 -- # local bdev_list 00:11:25.397 23:45:13 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:11:25.397 23:45:13 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@11 -- # local nbd_list 00:11:25.397 23:45:13 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@12 -- # local i 00:11:25.397 23:45:13 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:11:25.397 23:45:13 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:11:25.398 23:45:13 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:11:25.657 [2024-11-26 23:45:13.536606] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000027a0 00:11:25.657 /dev/nbd0 00:11:25.657 23:45:13 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:11:25.657 23:45:13 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:11:25.657 23:45:13 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:11:25.657 23:45:13 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@873 -- # local i 00:11:25.657 23:45:13 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:11:25.657 23:45:13 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:11:25.657 23:45:13 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:11:25.657 23:45:13 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@877 -- # break 00:11:25.657 23:45:13 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:11:25.657 23:45:13 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:11:25.657 23:45:13 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:11:25.657 1+0 records in 00:11:25.657 1+0 records out 00:11:25.657 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000511696 s, 8.0 MB/s 00:11:25.657 23:45:13 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:25.657 23:45:13 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@890 -- # size=4096 00:11:25.657 23:45:13 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:25.657 23:45:13 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:11:25.657 23:45:13 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@893 -- # return 0 00:11:25.657 23:45:13 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:11:25.657 23:45:13 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:11:25.657 23:45:13 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@629 -- # '[' raid1 = raid5f ']' 00:11:25.657 23:45:13 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@633 -- # write_unit_size=1 00:11:25.657 23:45:13 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=512 count=65536 oflag=direct 00:11:30.966 65536+0 records in 00:11:30.966 65536+0 records out 00:11:30.966 33554432 bytes (34 MB, 32 MiB) copied, 5.31027 s, 6.3 MB/s 00:11:30.966 23:45:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:11:30.966 23:45:18 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:11:30.966 23:45:18 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:11:30.966 23:45:18 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@50 -- # local nbd_list 00:11:30.966 23:45:18 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@51 -- # local i 00:11:30.966 23:45:18 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:11:30.966 23:45:18 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:11:31.223 [2024-11-26 23:45:19.112542] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:31.223 23:45:19 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:11:31.223 23:45:19 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:11:31.223 23:45:19 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:11:31.223 23:45:19 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:11:31.223 23:45:19 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:11:31.223 23:45:19 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:11:31.223 23:45:19 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:11:31.223 23:45:19 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:11:31.223 23:45:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:11:31.223 23:45:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:31.223 23:45:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:31.223 [2024-11-26 23:45:19.144555] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:11:31.223 23:45:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:31.223 23:45:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:11:31.223 23:45:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:31.223 23:45:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:31.223 23:45:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:31.223 23:45:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:31.223 23:45:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:11:31.223 23:45:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:31.223 23:45:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:31.223 23:45:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:31.223 23:45:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:31.223 23:45:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:31.223 23:45:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:31.223 23:45:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:31.223 23:45:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:31.223 23:45:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:31.223 23:45:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:31.223 "name": "raid_bdev1", 00:11:31.223 "uuid": "49c74909-b545-42a5-8fa7-cf975caea55f", 00:11:31.223 "strip_size_kb": 0, 00:11:31.223 "state": "online", 00:11:31.223 "raid_level": "raid1", 00:11:31.223 "superblock": false, 00:11:31.223 "num_base_bdevs": 4, 00:11:31.223 "num_base_bdevs_discovered": 3, 00:11:31.223 "num_base_bdevs_operational": 3, 00:11:31.223 "base_bdevs_list": [ 00:11:31.223 { 00:11:31.223 "name": null, 00:11:31.223 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:31.223 "is_configured": false, 00:11:31.223 "data_offset": 0, 00:11:31.223 "data_size": 65536 00:11:31.223 }, 00:11:31.223 { 00:11:31.223 "name": "BaseBdev2", 00:11:31.223 "uuid": "da29134b-b34b-5dc6-9377-4447cd1702f1", 00:11:31.223 "is_configured": true, 00:11:31.223 "data_offset": 0, 00:11:31.223 "data_size": 65536 00:11:31.223 }, 00:11:31.223 { 00:11:31.223 "name": "BaseBdev3", 00:11:31.223 "uuid": "6a57c1b8-6a06-5b0e-b911-acaed11c675f", 00:11:31.223 "is_configured": true, 00:11:31.223 "data_offset": 0, 00:11:31.223 "data_size": 65536 00:11:31.223 }, 00:11:31.223 { 00:11:31.223 "name": "BaseBdev4", 00:11:31.223 "uuid": "0f7413ce-1741-52fe-8fa6-07c7944b623f", 00:11:31.223 "is_configured": true, 00:11:31.223 "data_offset": 0, 00:11:31.223 "data_size": 65536 00:11:31.223 } 00:11:31.223 ] 00:11:31.223 }' 00:11:31.223 23:45:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:31.223 23:45:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:31.482 23:45:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:11:31.482 23:45:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:31.482 23:45:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:31.482 [2024-11-26 23:45:19.579831] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:11:31.482 [2024-11-26 23:45:19.584047] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000d063c0 00:11:31.482 23:45:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:31.482 23:45:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@647 -- # sleep 1 00:11:31.482 [2024-11-26 23:45:19.586058] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:11:32.874 23:45:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:32.874 23:45:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:32.874 23:45:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:32.874 23:45:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:32.874 23:45:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:32.874 23:45:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:32.874 23:45:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:32.874 23:45:20 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:32.874 23:45:20 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:32.874 23:45:20 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:32.874 23:45:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:32.874 "name": "raid_bdev1", 00:11:32.874 "uuid": "49c74909-b545-42a5-8fa7-cf975caea55f", 00:11:32.874 "strip_size_kb": 0, 00:11:32.874 "state": "online", 00:11:32.874 "raid_level": "raid1", 00:11:32.874 "superblock": false, 00:11:32.874 "num_base_bdevs": 4, 00:11:32.874 "num_base_bdevs_discovered": 4, 00:11:32.874 "num_base_bdevs_operational": 4, 00:11:32.874 "process": { 00:11:32.874 "type": "rebuild", 00:11:32.874 "target": "spare", 00:11:32.874 "progress": { 00:11:32.874 "blocks": 20480, 00:11:32.874 "percent": 31 00:11:32.874 } 00:11:32.874 }, 00:11:32.874 "base_bdevs_list": [ 00:11:32.874 { 00:11:32.874 "name": "spare", 00:11:32.874 "uuid": "e0a953bb-2acb-5119-b408-a87da80aa838", 00:11:32.874 "is_configured": true, 00:11:32.874 "data_offset": 0, 00:11:32.874 "data_size": 65536 00:11:32.874 }, 00:11:32.874 { 00:11:32.874 "name": "BaseBdev2", 00:11:32.874 "uuid": "da29134b-b34b-5dc6-9377-4447cd1702f1", 00:11:32.874 "is_configured": true, 00:11:32.874 "data_offset": 0, 00:11:32.874 "data_size": 65536 00:11:32.874 }, 00:11:32.874 { 00:11:32.874 "name": "BaseBdev3", 00:11:32.874 "uuid": "6a57c1b8-6a06-5b0e-b911-acaed11c675f", 00:11:32.874 "is_configured": true, 00:11:32.874 "data_offset": 0, 00:11:32.874 "data_size": 65536 00:11:32.874 }, 00:11:32.874 { 00:11:32.874 "name": "BaseBdev4", 00:11:32.874 "uuid": "0f7413ce-1741-52fe-8fa6-07c7944b623f", 00:11:32.874 "is_configured": true, 00:11:32.874 "data_offset": 0, 00:11:32.874 "data_size": 65536 00:11:32.874 } 00:11:32.874 ] 00:11:32.874 }' 00:11:32.874 23:45:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:32.874 23:45:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:32.874 23:45:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:32.874 23:45:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:32.874 23:45:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:11:32.874 23:45:20 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:32.874 23:45:20 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:32.874 [2024-11-26 23:45:20.754786] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:11:32.874 [2024-11-26 23:45:20.790625] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:11:32.874 [2024-11-26 23:45:20.790700] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:32.874 [2024-11-26 23:45:20.790719] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:11:32.874 [2024-11-26 23:45:20.790726] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:11:32.874 23:45:20 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:32.874 23:45:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:11:32.874 23:45:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:32.874 23:45:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:32.874 23:45:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:32.874 23:45:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:32.874 23:45:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:11:32.874 23:45:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:32.874 23:45:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:32.874 23:45:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:32.874 23:45:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:32.874 23:45:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:32.874 23:45:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:32.874 23:45:20 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:32.874 23:45:20 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:32.874 23:45:20 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:32.874 23:45:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:32.874 "name": "raid_bdev1", 00:11:32.874 "uuid": "49c74909-b545-42a5-8fa7-cf975caea55f", 00:11:32.874 "strip_size_kb": 0, 00:11:32.875 "state": "online", 00:11:32.875 "raid_level": "raid1", 00:11:32.875 "superblock": false, 00:11:32.875 "num_base_bdevs": 4, 00:11:32.875 "num_base_bdevs_discovered": 3, 00:11:32.875 "num_base_bdevs_operational": 3, 00:11:32.875 "base_bdevs_list": [ 00:11:32.875 { 00:11:32.875 "name": null, 00:11:32.875 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:32.875 "is_configured": false, 00:11:32.875 "data_offset": 0, 00:11:32.875 "data_size": 65536 00:11:32.875 }, 00:11:32.875 { 00:11:32.875 "name": "BaseBdev2", 00:11:32.875 "uuid": "da29134b-b34b-5dc6-9377-4447cd1702f1", 00:11:32.875 "is_configured": true, 00:11:32.875 "data_offset": 0, 00:11:32.875 "data_size": 65536 00:11:32.875 }, 00:11:32.875 { 00:11:32.875 "name": "BaseBdev3", 00:11:32.875 "uuid": "6a57c1b8-6a06-5b0e-b911-acaed11c675f", 00:11:32.875 "is_configured": true, 00:11:32.875 "data_offset": 0, 00:11:32.875 "data_size": 65536 00:11:32.875 }, 00:11:32.875 { 00:11:32.875 "name": "BaseBdev4", 00:11:32.875 "uuid": "0f7413ce-1741-52fe-8fa6-07c7944b623f", 00:11:32.875 "is_configured": true, 00:11:32.875 "data_offset": 0, 00:11:32.875 "data_size": 65536 00:11:32.875 } 00:11:32.875 ] 00:11:32.875 }' 00:11:32.875 23:45:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:32.875 23:45:20 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:33.134 23:45:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:11:33.134 23:45:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:33.134 23:45:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:11:33.134 23:45:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=none 00:11:33.134 23:45:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:33.134 23:45:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:33.134 23:45:21 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:33.134 23:45:21 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:33.134 23:45:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:33.134 23:45:21 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:33.134 23:45:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:33.134 "name": "raid_bdev1", 00:11:33.134 "uuid": "49c74909-b545-42a5-8fa7-cf975caea55f", 00:11:33.134 "strip_size_kb": 0, 00:11:33.134 "state": "online", 00:11:33.134 "raid_level": "raid1", 00:11:33.134 "superblock": false, 00:11:33.134 "num_base_bdevs": 4, 00:11:33.134 "num_base_bdevs_discovered": 3, 00:11:33.134 "num_base_bdevs_operational": 3, 00:11:33.134 "base_bdevs_list": [ 00:11:33.134 { 00:11:33.134 "name": null, 00:11:33.134 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:33.134 "is_configured": false, 00:11:33.134 "data_offset": 0, 00:11:33.134 "data_size": 65536 00:11:33.134 }, 00:11:33.134 { 00:11:33.134 "name": "BaseBdev2", 00:11:33.134 "uuid": "da29134b-b34b-5dc6-9377-4447cd1702f1", 00:11:33.134 "is_configured": true, 00:11:33.134 "data_offset": 0, 00:11:33.134 "data_size": 65536 00:11:33.134 }, 00:11:33.134 { 00:11:33.134 "name": "BaseBdev3", 00:11:33.134 "uuid": "6a57c1b8-6a06-5b0e-b911-acaed11c675f", 00:11:33.134 "is_configured": true, 00:11:33.134 "data_offset": 0, 00:11:33.134 "data_size": 65536 00:11:33.134 }, 00:11:33.134 { 00:11:33.134 "name": "BaseBdev4", 00:11:33.134 "uuid": "0f7413ce-1741-52fe-8fa6-07c7944b623f", 00:11:33.134 "is_configured": true, 00:11:33.134 "data_offset": 0, 00:11:33.134 "data_size": 65536 00:11:33.134 } 00:11:33.134 ] 00:11:33.134 }' 00:11:33.134 23:45:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:33.394 23:45:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:11:33.394 23:45:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:33.394 23:45:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:11:33.394 23:45:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:11:33.394 23:45:21 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:33.394 23:45:21 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:33.394 [2024-11-26 23:45:21.338328] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:11:33.394 [2024-11-26 23:45:21.342490] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000d06490 00:11:33.394 23:45:21 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:33.394 23:45:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@663 -- # sleep 1 00:11:33.394 [2024-11-26 23:45:21.344357] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:11:34.350 23:45:22 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:34.350 23:45:22 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:34.350 23:45:22 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:34.350 23:45:22 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:34.350 23:45:22 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:34.350 23:45:22 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:34.350 23:45:22 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:34.350 23:45:22 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:34.350 23:45:22 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:34.350 23:45:22 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:34.351 23:45:22 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:34.351 "name": "raid_bdev1", 00:11:34.351 "uuid": "49c74909-b545-42a5-8fa7-cf975caea55f", 00:11:34.351 "strip_size_kb": 0, 00:11:34.351 "state": "online", 00:11:34.351 "raid_level": "raid1", 00:11:34.351 "superblock": false, 00:11:34.351 "num_base_bdevs": 4, 00:11:34.351 "num_base_bdevs_discovered": 4, 00:11:34.351 "num_base_bdevs_operational": 4, 00:11:34.351 "process": { 00:11:34.351 "type": "rebuild", 00:11:34.351 "target": "spare", 00:11:34.351 "progress": { 00:11:34.351 "blocks": 20480, 00:11:34.351 "percent": 31 00:11:34.351 } 00:11:34.351 }, 00:11:34.351 "base_bdevs_list": [ 00:11:34.351 { 00:11:34.351 "name": "spare", 00:11:34.351 "uuid": "e0a953bb-2acb-5119-b408-a87da80aa838", 00:11:34.351 "is_configured": true, 00:11:34.351 "data_offset": 0, 00:11:34.351 "data_size": 65536 00:11:34.351 }, 00:11:34.351 { 00:11:34.351 "name": "BaseBdev2", 00:11:34.351 "uuid": "da29134b-b34b-5dc6-9377-4447cd1702f1", 00:11:34.351 "is_configured": true, 00:11:34.351 "data_offset": 0, 00:11:34.351 "data_size": 65536 00:11:34.351 }, 00:11:34.351 { 00:11:34.351 "name": "BaseBdev3", 00:11:34.351 "uuid": "6a57c1b8-6a06-5b0e-b911-acaed11c675f", 00:11:34.351 "is_configured": true, 00:11:34.351 "data_offset": 0, 00:11:34.351 "data_size": 65536 00:11:34.351 }, 00:11:34.351 { 00:11:34.351 "name": "BaseBdev4", 00:11:34.351 "uuid": "0f7413ce-1741-52fe-8fa6-07c7944b623f", 00:11:34.351 "is_configured": true, 00:11:34.351 "data_offset": 0, 00:11:34.351 "data_size": 65536 00:11:34.351 } 00:11:34.351 ] 00:11:34.351 }' 00:11:34.351 23:45:22 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:34.351 23:45:22 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:34.351 23:45:22 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:34.351 23:45:22 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:34.351 23:45:22 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@666 -- # '[' false = true ']' 00:11:34.351 23:45:22 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=4 00:11:34.351 23:45:22 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:11:34.351 23:45:22 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@693 -- # '[' 4 -gt 2 ']' 00:11:34.351 23:45:22 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@695 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:11:34.351 23:45:22 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:34.351 23:45:22 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:34.351 [2024-11-26 23:45:22.466068] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:11:34.609 [2024-11-26 23:45:22.548466] bdev_raid.c:1974:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 1 raid_ch: 0x60d000d06490 00:11:34.609 23:45:22 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:34.609 23:45:22 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@698 -- # base_bdevs[1]= 00:11:34.609 23:45:22 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@699 -- # (( num_base_bdevs_operational-- )) 00:11:34.609 23:45:22 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@702 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:34.609 23:45:22 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:34.609 23:45:22 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:34.609 23:45:22 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:34.609 23:45:22 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:34.609 23:45:22 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:34.609 23:45:22 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:34.609 23:45:22 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:34.609 23:45:22 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:34.609 23:45:22 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:34.609 23:45:22 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:34.609 "name": "raid_bdev1", 00:11:34.609 "uuid": "49c74909-b545-42a5-8fa7-cf975caea55f", 00:11:34.609 "strip_size_kb": 0, 00:11:34.609 "state": "online", 00:11:34.609 "raid_level": "raid1", 00:11:34.609 "superblock": false, 00:11:34.609 "num_base_bdevs": 4, 00:11:34.609 "num_base_bdevs_discovered": 3, 00:11:34.609 "num_base_bdevs_operational": 3, 00:11:34.609 "process": { 00:11:34.609 "type": "rebuild", 00:11:34.609 "target": "spare", 00:11:34.609 "progress": { 00:11:34.609 "blocks": 24576, 00:11:34.609 "percent": 37 00:11:34.609 } 00:11:34.609 }, 00:11:34.609 "base_bdevs_list": [ 00:11:34.609 { 00:11:34.609 "name": "spare", 00:11:34.609 "uuid": "e0a953bb-2acb-5119-b408-a87da80aa838", 00:11:34.609 "is_configured": true, 00:11:34.609 "data_offset": 0, 00:11:34.609 "data_size": 65536 00:11:34.609 }, 00:11:34.609 { 00:11:34.609 "name": null, 00:11:34.609 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:34.609 "is_configured": false, 00:11:34.609 "data_offset": 0, 00:11:34.609 "data_size": 65536 00:11:34.609 }, 00:11:34.609 { 00:11:34.609 "name": "BaseBdev3", 00:11:34.609 "uuid": "6a57c1b8-6a06-5b0e-b911-acaed11c675f", 00:11:34.609 "is_configured": true, 00:11:34.609 "data_offset": 0, 00:11:34.609 "data_size": 65536 00:11:34.609 }, 00:11:34.609 { 00:11:34.609 "name": "BaseBdev4", 00:11:34.609 "uuid": "0f7413ce-1741-52fe-8fa6-07c7944b623f", 00:11:34.609 "is_configured": true, 00:11:34.609 "data_offset": 0, 00:11:34.609 "data_size": 65536 00:11:34.609 } 00:11:34.609 ] 00:11:34.609 }' 00:11:34.609 23:45:22 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:34.609 23:45:22 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:34.609 23:45:22 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:34.609 23:45:22 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:34.609 23:45:22 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@706 -- # local timeout=354 00:11:34.609 23:45:22 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:11:34.609 23:45:22 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:34.609 23:45:22 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:34.609 23:45:22 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:34.609 23:45:22 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:34.609 23:45:22 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:34.609 23:45:22 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:34.609 23:45:22 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:34.609 23:45:22 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:34.609 23:45:22 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:34.609 23:45:22 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:34.609 23:45:22 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:34.609 "name": "raid_bdev1", 00:11:34.609 "uuid": "49c74909-b545-42a5-8fa7-cf975caea55f", 00:11:34.609 "strip_size_kb": 0, 00:11:34.609 "state": "online", 00:11:34.609 "raid_level": "raid1", 00:11:34.609 "superblock": false, 00:11:34.609 "num_base_bdevs": 4, 00:11:34.609 "num_base_bdevs_discovered": 3, 00:11:34.609 "num_base_bdevs_operational": 3, 00:11:34.609 "process": { 00:11:34.609 "type": "rebuild", 00:11:34.609 "target": "spare", 00:11:34.609 "progress": { 00:11:34.609 "blocks": 26624, 00:11:34.609 "percent": 40 00:11:34.609 } 00:11:34.609 }, 00:11:34.609 "base_bdevs_list": [ 00:11:34.609 { 00:11:34.609 "name": "spare", 00:11:34.609 "uuid": "e0a953bb-2acb-5119-b408-a87da80aa838", 00:11:34.609 "is_configured": true, 00:11:34.609 "data_offset": 0, 00:11:34.609 "data_size": 65536 00:11:34.609 }, 00:11:34.609 { 00:11:34.609 "name": null, 00:11:34.609 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:34.609 "is_configured": false, 00:11:34.609 "data_offset": 0, 00:11:34.609 "data_size": 65536 00:11:34.609 }, 00:11:34.609 { 00:11:34.609 "name": "BaseBdev3", 00:11:34.609 "uuid": "6a57c1b8-6a06-5b0e-b911-acaed11c675f", 00:11:34.609 "is_configured": true, 00:11:34.609 "data_offset": 0, 00:11:34.609 "data_size": 65536 00:11:34.609 }, 00:11:34.609 { 00:11:34.609 "name": "BaseBdev4", 00:11:34.609 "uuid": "0f7413ce-1741-52fe-8fa6-07c7944b623f", 00:11:34.609 "is_configured": true, 00:11:34.609 "data_offset": 0, 00:11:34.609 "data_size": 65536 00:11:34.609 } 00:11:34.609 ] 00:11:34.609 }' 00:11:34.609 23:45:22 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:34.868 23:45:22 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:34.868 23:45:22 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:34.868 23:45:22 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:34.868 23:45:22 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:11:35.826 23:45:23 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:11:35.826 23:45:23 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:35.826 23:45:23 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:35.826 23:45:23 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:35.826 23:45:23 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:35.826 23:45:23 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:35.826 23:45:23 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:35.826 23:45:23 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:35.826 23:45:23 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:35.827 23:45:23 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:35.827 23:45:23 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:35.827 23:45:23 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:35.827 "name": "raid_bdev1", 00:11:35.827 "uuid": "49c74909-b545-42a5-8fa7-cf975caea55f", 00:11:35.827 "strip_size_kb": 0, 00:11:35.827 "state": "online", 00:11:35.827 "raid_level": "raid1", 00:11:35.827 "superblock": false, 00:11:35.827 "num_base_bdevs": 4, 00:11:35.827 "num_base_bdevs_discovered": 3, 00:11:35.827 "num_base_bdevs_operational": 3, 00:11:35.827 "process": { 00:11:35.827 "type": "rebuild", 00:11:35.827 "target": "spare", 00:11:35.827 "progress": { 00:11:35.827 "blocks": 49152, 00:11:35.827 "percent": 75 00:11:35.827 } 00:11:35.827 }, 00:11:35.827 "base_bdevs_list": [ 00:11:35.827 { 00:11:35.827 "name": "spare", 00:11:35.827 "uuid": "e0a953bb-2acb-5119-b408-a87da80aa838", 00:11:35.827 "is_configured": true, 00:11:35.827 "data_offset": 0, 00:11:35.827 "data_size": 65536 00:11:35.827 }, 00:11:35.827 { 00:11:35.827 "name": null, 00:11:35.827 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:35.827 "is_configured": false, 00:11:35.827 "data_offset": 0, 00:11:35.827 "data_size": 65536 00:11:35.827 }, 00:11:35.827 { 00:11:35.827 "name": "BaseBdev3", 00:11:35.827 "uuid": "6a57c1b8-6a06-5b0e-b911-acaed11c675f", 00:11:35.827 "is_configured": true, 00:11:35.827 "data_offset": 0, 00:11:35.827 "data_size": 65536 00:11:35.827 }, 00:11:35.827 { 00:11:35.827 "name": "BaseBdev4", 00:11:35.827 "uuid": "0f7413ce-1741-52fe-8fa6-07c7944b623f", 00:11:35.827 "is_configured": true, 00:11:35.827 "data_offset": 0, 00:11:35.827 "data_size": 65536 00:11:35.827 } 00:11:35.827 ] 00:11:35.827 }' 00:11:35.827 23:45:23 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:35.827 23:45:23 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:35.827 23:45:23 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:36.088 23:45:23 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:36.088 23:45:23 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:11:36.655 [2024-11-26 23:45:24.555263] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:11:36.655 [2024-11-26 23:45:24.555417] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:11:36.655 [2024-11-26 23:45:24.555490] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:36.915 23:45:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:11:36.915 23:45:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:36.915 23:45:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:36.915 23:45:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:36.915 23:45:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:36.915 23:45:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:36.915 23:45:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:36.915 23:45:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:36.915 23:45:24 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:36.915 23:45:24 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:36.915 23:45:24 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:36.915 23:45:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:36.915 "name": "raid_bdev1", 00:11:36.915 "uuid": "49c74909-b545-42a5-8fa7-cf975caea55f", 00:11:36.915 "strip_size_kb": 0, 00:11:36.915 "state": "online", 00:11:36.915 "raid_level": "raid1", 00:11:36.915 "superblock": false, 00:11:36.915 "num_base_bdevs": 4, 00:11:36.915 "num_base_bdevs_discovered": 3, 00:11:36.915 "num_base_bdevs_operational": 3, 00:11:36.915 "base_bdevs_list": [ 00:11:36.915 { 00:11:36.915 "name": "spare", 00:11:36.915 "uuid": "e0a953bb-2acb-5119-b408-a87da80aa838", 00:11:36.915 "is_configured": true, 00:11:36.915 "data_offset": 0, 00:11:36.915 "data_size": 65536 00:11:36.915 }, 00:11:36.915 { 00:11:36.915 "name": null, 00:11:36.915 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:36.915 "is_configured": false, 00:11:36.915 "data_offset": 0, 00:11:36.915 "data_size": 65536 00:11:36.915 }, 00:11:36.915 { 00:11:36.915 "name": "BaseBdev3", 00:11:36.915 "uuid": "6a57c1b8-6a06-5b0e-b911-acaed11c675f", 00:11:36.915 "is_configured": true, 00:11:36.915 "data_offset": 0, 00:11:36.915 "data_size": 65536 00:11:36.915 }, 00:11:36.915 { 00:11:36.915 "name": "BaseBdev4", 00:11:36.915 "uuid": "0f7413ce-1741-52fe-8fa6-07c7944b623f", 00:11:36.915 "is_configured": true, 00:11:36.915 "data_offset": 0, 00:11:36.915 "data_size": 65536 00:11:36.915 } 00:11:36.915 ] 00:11:36.915 }' 00:11:36.915 23:45:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:37.176 23:45:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:11:37.176 23:45:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:37.176 23:45:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:11:37.176 23:45:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@709 -- # break 00:11:37.176 23:45:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:11:37.176 23:45:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:37.176 23:45:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:11:37.176 23:45:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=none 00:11:37.176 23:45:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:37.176 23:45:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:37.176 23:45:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:37.176 23:45:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:37.176 23:45:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:37.176 23:45:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:37.176 23:45:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:37.176 "name": "raid_bdev1", 00:11:37.176 "uuid": "49c74909-b545-42a5-8fa7-cf975caea55f", 00:11:37.176 "strip_size_kb": 0, 00:11:37.176 "state": "online", 00:11:37.176 "raid_level": "raid1", 00:11:37.176 "superblock": false, 00:11:37.176 "num_base_bdevs": 4, 00:11:37.176 "num_base_bdevs_discovered": 3, 00:11:37.176 "num_base_bdevs_operational": 3, 00:11:37.176 "base_bdevs_list": [ 00:11:37.176 { 00:11:37.176 "name": "spare", 00:11:37.176 "uuid": "e0a953bb-2acb-5119-b408-a87da80aa838", 00:11:37.176 "is_configured": true, 00:11:37.176 "data_offset": 0, 00:11:37.176 "data_size": 65536 00:11:37.176 }, 00:11:37.176 { 00:11:37.176 "name": null, 00:11:37.176 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:37.176 "is_configured": false, 00:11:37.176 "data_offset": 0, 00:11:37.176 "data_size": 65536 00:11:37.176 }, 00:11:37.176 { 00:11:37.176 "name": "BaseBdev3", 00:11:37.176 "uuid": "6a57c1b8-6a06-5b0e-b911-acaed11c675f", 00:11:37.176 "is_configured": true, 00:11:37.176 "data_offset": 0, 00:11:37.176 "data_size": 65536 00:11:37.176 }, 00:11:37.176 { 00:11:37.176 "name": "BaseBdev4", 00:11:37.176 "uuid": "0f7413ce-1741-52fe-8fa6-07c7944b623f", 00:11:37.176 "is_configured": true, 00:11:37.176 "data_offset": 0, 00:11:37.176 "data_size": 65536 00:11:37.176 } 00:11:37.176 ] 00:11:37.176 }' 00:11:37.176 23:45:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:37.176 23:45:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:11:37.176 23:45:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:37.176 23:45:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:11:37.176 23:45:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:11:37.176 23:45:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:37.176 23:45:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:37.176 23:45:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:37.176 23:45:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:37.176 23:45:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:11:37.176 23:45:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:37.176 23:45:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:37.176 23:45:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:37.176 23:45:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:37.176 23:45:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:37.176 23:45:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:37.176 23:45:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:37.176 23:45:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:37.176 23:45:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:37.435 23:45:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:37.435 "name": "raid_bdev1", 00:11:37.435 "uuid": "49c74909-b545-42a5-8fa7-cf975caea55f", 00:11:37.435 "strip_size_kb": 0, 00:11:37.435 "state": "online", 00:11:37.435 "raid_level": "raid1", 00:11:37.435 "superblock": false, 00:11:37.435 "num_base_bdevs": 4, 00:11:37.435 "num_base_bdevs_discovered": 3, 00:11:37.435 "num_base_bdevs_operational": 3, 00:11:37.435 "base_bdevs_list": [ 00:11:37.435 { 00:11:37.435 "name": "spare", 00:11:37.435 "uuid": "e0a953bb-2acb-5119-b408-a87da80aa838", 00:11:37.435 "is_configured": true, 00:11:37.435 "data_offset": 0, 00:11:37.435 "data_size": 65536 00:11:37.435 }, 00:11:37.435 { 00:11:37.435 "name": null, 00:11:37.435 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:37.435 "is_configured": false, 00:11:37.435 "data_offset": 0, 00:11:37.435 "data_size": 65536 00:11:37.435 }, 00:11:37.435 { 00:11:37.435 "name": "BaseBdev3", 00:11:37.435 "uuid": "6a57c1b8-6a06-5b0e-b911-acaed11c675f", 00:11:37.435 "is_configured": true, 00:11:37.435 "data_offset": 0, 00:11:37.435 "data_size": 65536 00:11:37.435 }, 00:11:37.435 { 00:11:37.435 "name": "BaseBdev4", 00:11:37.435 "uuid": "0f7413ce-1741-52fe-8fa6-07c7944b623f", 00:11:37.435 "is_configured": true, 00:11:37.435 "data_offset": 0, 00:11:37.435 "data_size": 65536 00:11:37.435 } 00:11:37.435 ] 00:11:37.435 }' 00:11:37.435 23:45:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:37.435 23:45:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:37.695 23:45:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:11:37.695 23:45:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:37.695 23:45:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:37.695 [2024-11-26 23:45:25.705877] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:11:37.695 [2024-11-26 23:45:25.705961] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:11:37.695 [2024-11-26 23:45:25.706075] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:11:37.695 [2024-11-26 23:45:25.706168] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:11:37.695 [2024-11-26 23:45:25.706221] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:11:37.695 23:45:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:37.695 23:45:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:37.695 23:45:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:37.695 23:45:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:37.695 23:45:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@720 -- # jq length 00:11:37.695 23:45:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:37.695 23:45:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:11:37.695 23:45:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:11:37.695 23:45:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:11:37.695 23:45:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:11:37.695 23:45:25 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:11:37.695 23:45:25 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:11:37.695 23:45:25 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@10 -- # local bdev_list 00:11:37.695 23:45:25 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:11:37.695 23:45:25 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@11 -- # local nbd_list 00:11:37.695 23:45:25 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@12 -- # local i 00:11:37.695 23:45:25 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:11:37.695 23:45:25 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:11:37.695 23:45:25 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:11:37.954 /dev/nbd0 00:11:37.954 23:45:25 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:11:37.954 23:45:25 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:11:37.954 23:45:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:11:37.954 23:45:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@873 -- # local i 00:11:37.954 23:45:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:11:37.954 23:45:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:11:37.954 23:45:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:11:37.954 23:45:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@877 -- # break 00:11:37.954 23:45:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:11:37.954 23:45:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:11:37.954 23:45:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:11:37.954 1+0 records in 00:11:37.954 1+0 records out 00:11:37.954 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000578585 s, 7.1 MB/s 00:11:37.954 23:45:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:37.954 23:45:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@890 -- # size=4096 00:11:37.954 23:45:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:37.954 23:45:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:11:37.955 23:45:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@893 -- # return 0 00:11:37.955 23:45:25 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:11:37.955 23:45:25 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:11:37.955 23:45:25 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:11:38.213 /dev/nbd1 00:11:38.213 23:45:26 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:11:38.213 23:45:26 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:11:38.213 23:45:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:11:38.213 23:45:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@873 -- # local i 00:11:38.213 23:45:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:11:38.213 23:45:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:11:38.213 23:45:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:11:38.213 23:45:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@877 -- # break 00:11:38.213 23:45:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:11:38.213 23:45:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:11:38.213 23:45:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:11:38.213 1+0 records in 00:11:38.213 1+0 records out 00:11:38.213 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000426825 s, 9.6 MB/s 00:11:38.213 23:45:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:38.213 23:45:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@890 -- # size=4096 00:11:38.213 23:45:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:38.213 23:45:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:11:38.213 23:45:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@893 -- # return 0 00:11:38.213 23:45:26 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:11:38.213 23:45:26 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:11:38.213 23:45:26 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@738 -- # cmp -i 0 /dev/nbd0 /dev/nbd1 00:11:38.213 23:45:26 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:11:38.213 23:45:26 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:11:38.213 23:45:26 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:11:38.213 23:45:26 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@50 -- # local nbd_list 00:11:38.213 23:45:26 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@51 -- # local i 00:11:38.213 23:45:26 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:11:38.213 23:45:26 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:11:38.472 23:45:26 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:11:38.472 23:45:26 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:11:38.472 23:45:26 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:11:38.472 23:45:26 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:11:38.472 23:45:26 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:11:38.472 23:45:26 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:11:38.472 23:45:26 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:11:38.472 23:45:26 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:11:38.472 23:45:26 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:11:38.472 23:45:26 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:11:38.734 23:45:26 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:11:38.734 23:45:26 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:11:38.734 23:45:26 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:11:38.734 23:45:26 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:11:38.734 23:45:26 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:11:38.734 23:45:26 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:11:38.734 23:45:26 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:11:38.734 23:45:26 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:11:38.734 23:45:26 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@743 -- # '[' false = true ']' 00:11:38.734 23:45:26 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@784 -- # killprocess 87876 00:11:38.734 23:45:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@954 -- # '[' -z 87876 ']' 00:11:38.734 23:45:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@958 -- # kill -0 87876 00:11:38.734 23:45:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@959 -- # uname 00:11:38.734 23:45:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:11:38.734 23:45:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 87876 00:11:38.734 23:45:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:11:38.734 23:45:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:11:38.734 23:45:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 87876' 00:11:38.734 killing process with pid 87876 00:11:38.734 23:45:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@973 -- # kill 87876 00:11:38.734 Received shutdown signal, test time was about 60.000000 seconds 00:11:38.734 00:11:38.734 Latency(us) 00:11:38.734 [2024-11-26T23:45:26.866Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:11:38.734 [2024-11-26T23:45:26.866Z] =================================================================================================================== 00:11:38.734 [2024-11-26T23:45:26.866Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:11:38.734 [2024-11-26 23:45:26.775460] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:11:38.734 23:45:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@978 -- # wait 87876 00:11:38.734 [2024-11-26 23:45:26.824791] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:11:38.994 23:45:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@786 -- # return 0 00:11:38.994 00:11:38.994 real 0m15.335s 00:11:38.994 user 0m17.429s 00:11:38.994 sys 0m2.892s 00:11:38.994 23:45:27 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:11:38.994 ************************************ 00:11:38.994 END TEST raid_rebuild_test 00:11:38.994 ************************************ 00:11:38.994 23:45:27 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:38.994 23:45:27 bdev_raid -- bdev/bdev_raid.sh@979 -- # run_test raid_rebuild_test_sb raid_rebuild_test raid1 4 true false true 00:11:38.994 23:45:27 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:11:38.994 23:45:27 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:11:38.994 23:45:27 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:11:38.994 ************************************ 00:11:38.994 START TEST raid_rebuild_test_sb 00:11:38.994 ************************************ 00:11:38.994 23:45:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@1129 -- # raid_rebuild_test raid1 4 true false true 00:11:38.994 23:45:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:11:38.994 23:45:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=4 00:11:38.994 23:45:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@571 -- # local superblock=true 00:11:38.994 23:45:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:11:38.994 23:45:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@573 -- # local verify=true 00:11:38.994 23:45:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:11:38.994 23:45:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:11:38.994 23:45:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:11:38.994 23:45:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:11:38.994 23:45:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:11:38.994 23:45:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:11:38.994 23:45:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:11:38.994 23:45:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:11:38.994 23:45:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev3 00:11:38.994 23:45:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:11:38.994 23:45:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:11:38.994 23:45:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev4 00:11:38.994 23:45:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:11:38.994 23:45:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:11:38.994 23:45:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:11:38.994 23:45:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:11:38.994 23:45:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:11:38.994 23:45:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # local strip_size 00:11:38.994 23:45:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@577 -- # local create_arg 00:11:38.994 23:45:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:11:38.994 23:45:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@579 -- # local data_offset 00:11:38.994 23:45:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:11:38.994 23:45:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:11:38.994 23:45:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@592 -- # '[' true = true ']' 00:11:38.994 23:45:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@593 -- # create_arg+=' -s' 00:11:38.994 23:45:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@597 -- # raid_pid=88300 00:11:38.994 23:45:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:11:38.994 23:45:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@598 -- # waitforlisten 88300 00:11:38.994 23:45:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@835 -- # '[' -z 88300 ']' 00:11:39.254 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:11:39.254 23:45:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:11:39.254 23:45:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@840 -- # local max_retries=100 00:11:39.254 23:45:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:11:39.254 23:45:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@844 -- # xtrace_disable 00:11:39.254 23:45:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:39.254 I/O size of 3145728 is greater than zero copy threshold (65536). 00:11:39.254 Zero copy mechanism will not be used. 00:11:39.254 [2024-11-26 23:45:27.199184] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:11:39.254 [2024-11-26 23:45:27.199322] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid88300 ] 00:11:39.254 [2024-11-26 23:45:27.354005] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:11:39.254 [2024-11-26 23:45:27.378385] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:11:39.513 [2024-11-26 23:45:27.420138] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:39.513 [2024-11-26 23:45:27.420256] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:40.083 23:45:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:11:40.083 23:45:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@868 -- # return 0 00:11:40.083 23:45:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:11:40.083 23:45:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:11:40.083 23:45:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:40.083 23:45:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:40.083 BaseBdev1_malloc 00:11:40.083 23:45:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:40.083 23:45:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:11:40.083 23:45:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:40.083 23:45:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:40.083 [2024-11-26 23:45:28.038943] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:11:40.083 [2024-11-26 23:45:28.039085] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:40.083 [2024-11-26 23:45:28.039143] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:11:40.083 [2024-11-26 23:45:28.039185] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:40.083 [2024-11-26 23:45:28.041204] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:40.083 [2024-11-26 23:45:28.041274] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:11:40.083 BaseBdev1 00:11:40.083 23:45:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:40.083 23:45:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:11:40.083 23:45:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:11:40.083 23:45:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:40.083 23:45:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:40.083 BaseBdev2_malloc 00:11:40.083 23:45:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:40.083 23:45:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:11:40.083 23:45:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:40.083 23:45:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:40.083 [2024-11-26 23:45:28.067297] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:11:40.083 [2024-11-26 23:45:28.067438] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:40.083 [2024-11-26 23:45:28.067481] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:11:40.083 [2024-11-26 23:45:28.067512] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:40.083 [2024-11-26 23:45:28.069514] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:40.084 [2024-11-26 23:45:28.069587] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:11:40.084 BaseBdev2 00:11:40.084 23:45:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:40.084 23:45:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:11:40.084 23:45:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:11:40.084 23:45:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:40.084 23:45:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:40.084 BaseBdev3_malloc 00:11:40.084 23:45:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:40.084 23:45:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev3_malloc -p BaseBdev3 00:11:40.084 23:45:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:40.084 23:45:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:40.084 [2024-11-26 23:45:28.095573] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev3_malloc 00:11:40.084 [2024-11-26 23:45:28.095625] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:40.084 [2024-11-26 23:45:28.095645] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:11:40.084 [2024-11-26 23:45:28.095654] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:40.084 [2024-11-26 23:45:28.097663] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:40.084 [2024-11-26 23:45:28.097696] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:11:40.084 BaseBdev3 00:11:40.084 23:45:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:40.084 23:45:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:11:40.084 23:45:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:11:40.084 23:45:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:40.084 23:45:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:40.084 BaseBdev4_malloc 00:11:40.084 23:45:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:40.084 23:45:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev4_malloc -p BaseBdev4 00:11:40.084 23:45:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:40.084 23:45:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:40.084 [2024-11-26 23:45:28.142284] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev4_malloc 00:11:40.084 [2024-11-26 23:45:28.142512] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:40.084 [2024-11-26 23:45:28.142567] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:11:40.084 [2024-11-26 23:45:28.142587] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:40.084 [2024-11-26 23:45:28.146718] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:40.084 [2024-11-26 23:45:28.146780] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:11:40.084 BaseBdev4 00:11:40.084 23:45:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:40.084 23:45:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:11:40.084 23:45:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:40.084 23:45:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:40.084 spare_malloc 00:11:40.084 23:45:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:40.084 23:45:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:11:40.084 23:45:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:40.084 23:45:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:40.084 spare_delay 00:11:40.084 23:45:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:40.084 23:45:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:11:40.084 23:45:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:40.084 23:45:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:40.084 [2024-11-26 23:45:28.183694] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:11:40.084 [2024-11-26 23:45:28.183809] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:40.084 [2024-11-26 23:45:28.183845] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009c80 00:11:40.084 [2024-11-26 23:45:28.183871] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:40.084 [2024-11-26 23:45:28.185882] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:40.084 [2024-11-26 23:45:28.185948] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:11:40.084 spare 00:11:40.084 23:45:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:40.084 23:45:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 00:11:40.084 23:45:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:40.084 23:45:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:40.084 [2024-11-26 23:45:28.195745] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:11:40.084 [2024-11-26 23:45:28.197549] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:11:40.084 [2024-11-26 23:45:28.197607] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:11:40.084 [2024-11-26 23:45:28.197650] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:11:40.084 [2024-11-26 23:45:28.197803] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:11:40.084 [2024-11-26 23:45:28.197814] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:11:40.084 [2024-11-26 23:45:28.198034] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:11:40.084 [2024-11-26 23:45:28.198155] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:11:40.084 [2024-11-26 23:45:28.198167] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:11:40.084 [2024-11-26 23:45:28.198293] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:40.084 23:45:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:40.084 23:45:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:11:40.084 23:45:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:40.084 23:45:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:40.084 23:45:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:40.084 23:45:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:40.084 23:45:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:40.084 23:45:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:40.084 23:45:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:40.084 23:45:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:40.084 23:45:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:40.084 23:45:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:40.084 23:45:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:40.084 23:45:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:40.084 23:45:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:40.346 23:45:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:40.347 23:45:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:40.347 "name": "raid_bdev1", 00:11:40.347 "uuid": "32f38c29-1954-4acc-8a9f-f847beecd7e8", 00:11:40.347 "strip_size_kb": 0, 00:11:40.347 "state": "online", 00:11:40.347 "raid_level": "raid1", 00:11:40.347 "superblock": true, 00:11:40.347 "num_base_bdevs": 4, 00:11:40.347 "num_base_bdevs_discovered": 4, 00:11:40.347 "num_base_bdevs_operational": 4, 00:11:40.347 "base_bdevs_list": [ 00:11:40.347 { 00:11:40.347 "name": "BaseBdev1", 00:11:40.347 "uuid": "c8241c0b-57fd-5a7e-a2a8-82b0cb909633", 00:11:40.347 "is_configured": true, 00:11:40.347 "data_offset": 2048, 00:11:40.347 "data_size": 63488 00:11:40.347 }, 00:11:40.347 { 00:11:40.347 "name": "BaseBdev2", 00:11:40.347 "uuid": "530c797d-986b-5b78-a50d-8649f527d38a", 00:11:40.347 "is_configured": true, 00:11:40.347 "data_offset": 2048, 00:11:40.347 "data_size": 63488 00:11:40.347 }, 00:11:40.347 { 00:11:40.347 "name": "BaseBdev3", 00:11:40.347 "uuid": "f56c9a76-6413-5d98-aff0-d9130d853e73", 00:11:40.347 "is_configured": true, 00:11:40.347 "data_offset": 2048, 00:11:40.347 "data_size": 63488 00:11:40.347 }, 00:11:40.347 { 00:11:40.347 "name": "BaseBdev4", 00:11:40.347 "uuid": "f3f95ef6-5797-5126-a8c1-b9ab3fd97b66", 00:11:40.347 "is_configured": true, 00:11:40.347 "data_offset": 2048, 00:11:40.347 "data_size": 63488 00:11:40.347 } 00:11:40.347 ] 00:11:40.347 }' 00:11:40.347 23:45:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:40.347 23:45:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:40.606 23:45:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:11:40.606 23:45:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:11:40.606 23:45:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:40.606 23:45:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:40.606 [2024-11-26 23:45:28.675234] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:11:40.606 23:45:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:40.606 23:45:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=63488 00:11:40.606 23:45:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:40.606 23:45:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:40.606 23:45:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:40.606 23:45:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:11:40.606 23:45:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:40.866 23:45:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # data_offset=2048 00:11:40.866 23:45:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:11:40.866 23:45:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:11:40.866 23:45:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:11:40.866 23:45:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:11:40.866 23:45:28 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:11:40.866 23:45:28 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:11:40.866 23:45:28 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # local bdev_list 00:11:40.866 23:45:28 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:11:40.866 23:45:28 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # local nbd_list 00:11:40.866 23:45:28 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@12 -- # local i 00:11:40.866 23:45:28 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:11:40.866 23:45:28 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:11:40.866 23:45:28 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:11:40.866 [2024-11-26 23:45:28.942526] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000027a0 00:11:40.866 /dev/nbd0 00:11:40.866 23:45:28 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:11:40.866 23:45:28 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:11:40.866 23:45:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:11:40.866 23:45:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@873 -- # local i 00:11:40.866 23:45:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:11:40.866 23:45:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:11:40.866 23:45:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:11:40.866 23:45:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@877 -- # break 00:11:40.866 23:45:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:11:40.866 23:45:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:11:40.866 23:45:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:11:40.866 1+0 records in 00:11:40.866 1+0 records out 00:11:40.866 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000449009 s, 9.1 MB/s 00:11:40.866 23:45:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:41.126 23:45:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@890 -- # size=4096 00:11:41.126 23:45:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:41.126 23:45:29 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:11:41.126 23:45:29 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@893 -- # return 0 00:11:41.126 23:45:29 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:11:41.126 23:45:29 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:11:41.126 23:45:29 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@629 -- # '[' raid1 = raid5f ']' 00:11:41.126 23:45:29 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@633 -- # write_unit_size=1 00:11:41.126 23:45:29 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=512 count=63488 oflag=direct 00:11:46.403 63488+0 records in 00:11:46.403 63488+0 records out 00:11:46.403 32505856 bytes (33 MB, 31 MiB) copied, 4.87258 s, 6.7 MB/s 00:11:46.403 23:45:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:11:46.403 23:45:33 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:11:46.403 23:45:33 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:11:46.403 23:45:33 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # local nbd_list 00:11:46.403 23:45:33 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@51 -- # local i 00:11:46.403 23:45:33 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:11:46.403 23:45:33 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:11:46.403 [2024-11-26 23:45:34.070935] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:46.403 23:45:34 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:11:46.403 23:45:34 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:11:46.403 23:45:34 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:11:46.403 23:45:34 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:11:46.403 23:45:34 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:11:46.403 23:45:34 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:11:46.403 23:45:34 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:11:46.403 23:45:34 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:11:46.403 23:45:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:11:46.403 23:45:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:46.403 23:45:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:46.403 [2024-11-26 23:45:34.108141] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:11:46.403 23:45:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:46.403 23:45:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:11:46.403 23:45:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:46.403 23:45:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:46.403 23:45:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:46.403 23:45:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:46.403 23:45:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:11:46.403 23:45:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:46.403 23:45:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:46.403 23:45:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:46.403 23:45:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:46.403 23:45:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:46.403 23:45:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:46.403 23:45:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:46.403 23:45:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:46.403 23:45:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:46.403 23:45:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:46.403 "name": "raid_bdev1", 00:11:46.403 "uuid": "32f38c29-1954-4acc-8a9f-f847beecd7e8", 00:11:46.403 "strip_size_kb": 0, 00:11:46.403 "state": "online", 00:11:46.403 "raid_level": "raid1", 00:11:46.403 "superblock": true, 00:11:46.403 "num_base_bdevs": 4, 00:11:46.403 "num_base_bdevs_discovered": 3, 00:11:46.403 "num_base_bdevs_operational": 3, 00:11:46.403 "base_bdevs_list": [ 00:11:46.403 { 00:11:46.403 "name": null, 00:11:46.403 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:46.403 "is_configured": false, 00:11:46.403 "data_offset": 0, 00:11:46.403 "data_size": 63488 00:11:46.403 }, 00:11:46.403 { 00:11:46.403 "name": "BaseBdev2", 00:11:46.403 "uuid": "530c797d-986b-5b78-a50d-8649f527d38a", 00:11:46.403 "is_configured": true, 00:11:46.403 "data_offset": 2048, 00:11:46.403 "data_size": 63488 00:11:46.403 }, 00:11:46.403 { 00:11:46.403 "name": "BaseBdev3", 00:11:46.403 "uuid": "f56c9a76-6413-5d98-aff0-d9130d853e73", 00:11:46.403 "is_configured": true, 00:11:46.403 "data_offset": 2048, 00:11:46.403 "data_size": 63488 00:11:46.403 }, 00:11:46.403 { 00:11:46.403 "name": "BaseBdev4", 00:11:46.403 "uuid": "f3f95ef6-5797-5126-a8c1-b9ab3fd97b66", 00:11:46.403 "is_configured": true, 00:11:46.403 "data_offset": 2048, 00:11:46.403 "data_size": 63488 00:11:46.404 } 00:11:46.404 ] 00:11:46.404 }' 00:11:46.404 23:45:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:46.404 23:45:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:46.663 23:45:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:11:46.663 23:45:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:46.663 23:45:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:46.663 [2024-11-26 23:45:34.539482] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:11:46.663 [2024-11-26 23:45:34.543808] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000c3e420 00:11:46.664 23:45:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:46.664 23:45:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@647 -- # sleep 1 00:11:46.664 [2024-11-26 23:45:34.545807] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:11:47.600 23:45:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:47.600 23:45:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:47.600 23:45:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:47.600 23:45:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:47.600 23:45:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:47.600 23:45:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:47.600 23:45:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:47.600 23:45:35 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:47.600 23:45:35 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:47.600 23:45:35 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:47.600 23:45:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:47.600 "name": "raid_bdev1", 00:11:47.600 "uuid": "32f38c29-1954-4acc-8a9f-f847beecd7e8", 00:11:47.600 "strip_size_kb": 0, 00:11:47.600 "state": "online", 00:11:47.600 "raid_level": "raid1", 00:11:47.600 "superblock": true, 00:11:47.600 "num_base_bdevs": 4, 00:11:47.600 "num_base_bdevs_discovered": 4, 00:11:47.600 "num_base_bdevs_operational": 4, 00:11:47.600 "process": { 00:11:47.600 "type": "rebuild", 00:11:47.600 "target": "spare", 00:11:47.600 "progress": { 00:11:47.600 "blocks": 20480, 00:11:47.600 "percent": 32 00:11:47.600 } 00:11:47.600 }, 00:11:47.600 "base_bdevs_list": [ 00:11:47.600 { 00:11:47.600 "name": "spare", 00:11:47.600 "uuid": "b4091172-2cc8-5287-a78c-5cf3dae99027", 00:11:47.600 "is_configured": true, 00:11:47.600 "data_offset": 2048, 00:11:47.600 "data_size": 63488 00:11:47.600 }, 00:11:47.600 { 00:11:47.600 "name": "BaseBdev2", 00:11:47.600 "uuid": "530c797d-986b-5b78-a50d-8649f527d38a", 00:11:47.600 "is_configured": true, 00:11:47.600 "data_offset": 2048, 00:11:47.600 "data_size": 63488 00:11:47.600 }, 00:11:47.600 { 00:11:47.600 "name": "BaseBdev3", 00:11:47.600 "uuid": "f56c9a76-6413-5d98-aff0-d9130d853e73", 00:11:47.600 "is_configured": true, 00:11:47.600 "data_offset": 2048, 00:11:47.600 "data_size": 63488 00:11:47.600 }, 00:11:47.600 { 00:11:47.600 "name": "BaseBdev4", 00:11:47.600 "uuid": "f3f95ef6-5797-5126-a8c1-b9ab3fd97b66", 00:11:47.600 "is_configured": true, 00:11:47.600 "data_offset": 2048, 00:11:47.600 "data_size": 63488 00:11:47.600 } 00:11:47.600 ] 00:11:47.600 }' 00:11:47.600 23:45:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:47.600 23:45:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:47.600 23:45:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:47.600 23:45:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:47.600 23:45:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:11:47.600 23:45:35 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:47.600 23:45:35 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:47.600 [2024-11-26 23:45:35.686486] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:11:47.860 [2024-11-26 23:45:35.750333] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:11:47.860 [2024-11-26 23:45:35.750451] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:47.860 [2024-11-26 23:45:35.750490] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:11:47.860 [2024-11-26 23:45:35.750513] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:11:47.860 23:45:35 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:47.860 23:45:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:11:47.860 23:45:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:47.860 23:45:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:47.860 23:45:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:47.860 23:45:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:47.860 23:45:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:11:47.860 23:45:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:47.860 23:45:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:47.860 23:45:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:47.860 23:45:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:47.860 23:45:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:47.860 23:45:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:47.860 23:45:35 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:47.860 23:45:35 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:47.860 23:45:35 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:47.860 23:45:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:47.860 "name": "raid_bdev1", 00:11:47.860 "uuid": "32f38c29-1954-4acc-8a9f-f847beecd7e8", 00:11:47.860 "strip_size_kb": 0, 00:11:47.860 "state": "online", 00:11:47.860 "raid_level": "raid1", 00:11:47.860 "superblock": true, 00:11:47.860 "num_base_bdevs": 4, 00:11:47.860 "num_base_bdevs_discovered": 3, 00:11:47.860 "num_base_bdevs_operational": 3, 00:11:47.860 "base_bdevs_list": [ 00:11:47.860 { 00:11:47.860 "name": null, 00:11:47.860 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:47.860 "is_configured": false, 00:11:47.860 "data_offset": 0, 00:11:47.860 "data_size": 63488 00:11:47.860 }, 00:11:47.860 { 00:11:47.860 "name": "BaseBdev2", 00:11:47.860 "uuid": "530c797d-986b-5b78-a50d-8649f527d38a", 00:11:47.860 "is_configured": true, 00:11:47.860 "data_offset": 2048, 00:11:47.860 "data_size": 63488 00:11:47.860 }, 00:11:47.860 { 00:11:47.860 "name": "BaseBdev3", 00:11:47.860 "uuid": "f56c9a76-6413-5d98-aff0-d9130d853e73", 00:11:47.860 "is_configured": true, 00:11:47.860 "data_offset": 2048, 00:11:47.860 "data_size": 63488 00:11:47.860 }, 00:11:47.860 { 00:11:47.860 "name": "BaseBdev4", 00:11:47.860 "uuid": "f3f95ef6-5797-5126-a8c1-b9ab3fd97b66", 00:11:47.860 "is_configured": true, 00:11:47.860 "data_offset": 2048, 00:11:47.860 "data_size": 63488 00:11:47.860 } 00:11:47.860 ] 00:11:47.860 }' 00:11:47.860 23:45:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:47.860 23:45:35 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:48.119 23:45:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:11:48.119 23:45:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:48.119 23:45:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:11:48.119 23:45:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:11:48.119 23:45:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:48.119 23:45:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:48.119 23:45:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:48.119 23:45:36 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:48.119 23:45:36 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:48.119 23:45:36 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:48.119 23:45:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:48.119 "name": "raid_bdev1", 00:11:48.119 "uuid": "32f38c29-1954-4acc-8a9f-f847beecd7e8", 00:11:48.119 "strip_size_kb": 0, 00:11:48.119 "state": "online", 00:11:48.119 "raid_level": "raid1", 00:11:48.119 "superblock": true, 00:11:48.119 "num_base_bdevs": 4, 00:11:48.119 "num_base_bdevs_discovered": 3, 00:11:48.119 "num_base_bdevs_operational": 3, 00:11:48.119 "base_bdevs_list": [ 00:11:48.119 { 00:11:48.119 "name": null, 00:11:48.119 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:48.119 "is_configured": false, 00:11:48.119 "data_offset": 0, 00:11:48.119 "data_size": 63488 00:11:48.119 }, 00:11:48.119 { 00:11:48.119 "name": "BaseBdev2", 00:11:48.119 "uuid": "530c797d-986b-5b78-a50d-8649f527d38a", 00:11:48.119 "is_configured": true, 00:11:48.119 "data_offset": 2048, 00:11:48.119 "data_size": 63488 00:11:48.119 }, 00:11:48.119 { 00:11:48.119 "name": "BaseBdev3", 00:11:48.119 "uuid": "f56c9a76-6413-5d98-aff0-d9130d853e73", 00:11:48.119 "is_configured": true, 00:11:48.119 "data_offset": 2048, 00:11:48.119 "data_size": 63488 00:11:48.119 }, 00:11:48.119 { 00:11:48.119 "name": "BaseBdev4", 00:11:48.119 "uuid": "f3f95ef6-5797-5126-a8c1-b9ab3fd97b66", 00:11:48.119 "is_configured": true, 00:11:48.119 "data_offset": 2048, 00:11:48.119 "data_size": 63488 00:11:48.120 } 00:11:48.120 ] 00:11:48.120 }' 00:11:48.379 23:45:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:48.379 23:45:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:11:48.379 23:45:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:48.379 23:45:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:11:48.379 23:45:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:11:48.379 23:45:36 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:48.379 23:45:36 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:48.379 [2024-11-26 23:45:36.314159] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:11:48.379 [2024-11-26 23:45:36.318219] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000c3e4f0 00:11:48.379 23:45:36 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:48.379 23:45:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@663 -- # sleep 1 00:11:48.379 [2024-11-26 23:45:36.320101] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:11:49.316 23:45:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:49.316 23:45:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:49.316 23:45:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:49.316 23:45:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:49.316 23:45:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:49.316 23:45:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:49.316 23:45:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:49.316 23:45:37 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:49.316 23:45:37 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:49.316 23:45:37 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:49.316 23:45:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:49.316 "name": "raid_bdev1", 00:11:49.316 "uuid": "32f38c29-1954-4acc-8a9f-f847beecd7e8", 00:11:49.316 "strip_size_kb": 0, 00:11:49.316 "state": "online", 00:11:49.316 "raid_level": "raid1", 00:11:49.316 "superblock": true, 00:11:49.316 "num_base_bdevs": 4, 00:11:49.316 "num_base_bdevs_discovered": 4, 00:11:49.316 "num_base_bdevs_operational": 4, 00:11:49.316 "process": { 00:11:49.316 "type": "rebuild", 00:11:49.316 "target": "spare", 00:11:49.316 "progress": { 00:11:49.316 "blocks": 20480, 00:11:49.316 "percent": 32 00:11:49.316 } 00:11:49.316 }, 00:11:49.316 "base_bdevs_list": [ 00:11:49.316 { 00:11:49.316 "name": "spare", 00:11:49.316 "uuid": "b4091172-2cc8-5287-a78c-5cf3dae99027", 00:11:49.316 "is_configured": true, 00:11:49.316 "data_offset": 2048, 00:11:49.316 "data_size": 63488 00:11:49.316 }, 00:11:49.316 { 00:11:49.316 "name": "BaseBdev2", 00:11:49.316 "uuid": "530c797d-986b-5b78-a50d-8649f527d38a", 00:11:49.316 "is_configured": true, 00:11:49.316 "data_offset": 2048, 00:11:49.316 "data_size": 63488 00:11:49.316 }, 00:11:49.316 { 00:11:49.316 "name": "BaseBdev3", 00:11:49.316 "uuid": "f56c9a76-6413-5d98-aff0-d9130d853e73", 00:11:49.316 "is_configured": true, 00:11:49.316 "data_offset": 2048, 00:11:49.316 "data_size": 63488 00:11:49.316 }, 00:11:49.316 { 00:11:49.316 "name": "BaseBdev4", 00:11:49.316 "uuid": "f3f95ef6-5797-5126-a8c1-b9ab3fd97b66", 00:11:49.316 "is_configured": true, 00:11:49.316 "data_offset": 2048, 00:11:49.316 "data_size": 63488 00:11:49.316 } 00:11:49.316 ] 00:11:49.316 }' 00:11:49.316 23:45:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:49.316 23:45:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:49.316 23:45:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:49.576 23:45:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:49.576 23:45:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@666 -- # '[' true = true ']' 00:11:49.576 23:45:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@666 -- # '[' = false ']' 00:11:49.576 /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh: line 666: [: =: unary operator expected 00:11:49.576 23:45:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=4 00:11:49.576 23:45:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:11:49.576 23:45:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@693 -- # '[' 4 -gt 2 ']' 00:11:49.576 23:45:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@695 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:11:49.576 23:45:37 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:49.576 23:45:37 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:49.576 [2024-11-26 23:45:37.476907] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:11:49.576 [2024-11-26 23:45:37.623927] bdev_raid.c:1974:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 1 raid_ch: 0x60d000c3e4f0 00:11:49.576 23:45:37 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:49.576 23:45:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@698 -- # base_bdevs[1]= 00:11:49.576 23:45:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@699 -- # (( num_base_bdevs_operational-- )) 00:11:49.576 23:45:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@702 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:49.576 23:45:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:49.576 23:45:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:49.576 23:45:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:49.576 23:45:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:49.576 23:45:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:49.576 23:45:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:49.576 23:45:37 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:49.576 23:45:37 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:49.576 23:45:37 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:49.576 23:45:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:49.576 "name": "raid_bdev1", 00:11:49.576 "uuid": "32f38c29-1954-4acc-8a9f-f847beecd7e8", 00:11:49.576 "strip_size_kb": 0, 00:11:49.576 "state": "online", 00:11:49.576 "raid_level": "raid1", 00:11:49.576 "superblock": true, 00:11:49.576 "num_base_bdevs": 4, 00:11:49.576 "num_base_bdevs_discovered": 3, 00:11:49.576 "num_base_bdevs_operational": 3, 00:11:49.576 "process": { 00:11:49.576 "type": "rebuild", 00:11:49.576 "target": "spare", 00:11:49.576 "progress": { 00:11:49.576 "blocks": 24576, 00:11:49.576 "percent": 38 00:11:49.576 } 00:11:49.576 }, 00:11:49.576 "base_bdevs_list": [ 00:11:49.576 { 00:11:49.576 "name": "spare", 00:11:49.576 "uuid": "b4091172-2cc8-5287-a78c-5cf3dae99027", 00:11:49.576 "is_configured": true, 00:11:49.576 "data_offset": 2048, 00:11:49.576 "data_size": 63488 00:11:49.576 }, 00:11:49.576 { 00:11:49.576 "name": null, 00:11:49.576 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:49.576 "is_configured": false, 00:11:49.576 "data_offset": 0, 00:11:49.576 "data_size": 63488 00:11:49.576 }, 00:11:49.576 { 00:11:49.576 "name": "BaseBdev3", 00:11:49.576 "uuid": "f56c9a76-6413-5d98-aff0-d9130d853e73", 00:11:49.576 "is_configured": true, 00:11:49.576 "data_offset": 2048, 00:11:49.576 "data_size": 63488 00:11:49.576 }, 00:11:49.576 { 00:11:49.576 "name": "BaseBdev4", 00:11:49.576 "uuid": "f3f95ef6-5797-5126-a8c1-b9ab3fd97b66", 00:11:49.576 "is_configured": true, 00:11:49.576 "data_offset": 2048, 00:11:49.576 "data_size": 63488 00:11:49.576 } 00:11:49.576 ] 00:11:49.576 }' 00:11:49.576 23:45:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:49.848 23:45:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:49.848 23:45:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:49.848 23:45:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:49.848 23:45:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@706 -- # local timeout=369 00:11:49.848 23:45:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:11:49.848 23:45:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:49.848 23:45:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:49.848 23:45:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:49.848 23:45:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:49.848 23:45:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:49.848 23:45:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:49.848 23:45:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:49.848 23:45:37 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:49.848 23:45:37 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:49.848 23:45:37 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:49.848 23:45:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:49.848 "name": "raid_bdev1", 00:11:49.848 "uuid": "32f38c29-1954-4acc-8a9f-f847beecd7e8", 00:11:49.848 "strip_size_kb": 0, 00:11:49.848 "state": "online", 00:11:49.848 "raid_level": "raid1", 00:11:49.848 "superblock": true, 00:11:49.848 "num_base_bdevs": 4, 00:11:49.848 "num_base_bdevs_discovered": 3, 00:11:49.848 "num_base_bdevs_operational": 3, 00:11:49.848 "process": { 00:11:49.848 "type": "rebuild", 00:11:49.848 "target": "spare", 00:11:49.848 "progress": { 00:11:49.848 "blocks": 26624, 00:11:49.848 "percent": 41 00:11:49.848 } 00:11:49.848 }, 00:11:49.848 "base_bdevs_list": [ 00:11:49.848 { 00:11:49.848 "name": "spare", 00:11:49.848 "uuid": "b4091172-2cc8-5287-a78c-5cf3dae99027", 00:11:49.848 "is_configured": true, 00:11:49.848 "data_offset": 2048, 00:11:49.848 "data_size": 63488 00:11:49.848 }, 00:11:49.848 { 00:11:49.848 "name": null, 00:11:49.848 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:49.848 "is_configured": false, 00:11:49.848 "data_offset": 0, 00:11:49.848 "data_size": 63488 00:11:49.848 }, 00:11:49.848 { 00:11:49.848 "name": "BaseBdev3", 00:11:49.848 "uuid": "f56c9a76-6413-5d98-aff0-d9130d853e73", 00:11:49.848 "is_configured": true, 00:11:49.848 "data_offset": 2048, 00:11:49.848 "data_size": 63488 00:11:49.848 }, 00:11:49.848 { 00:11:49.848 "name": "BaseBdev4", 00:11:49.848 "uuid": "f3f95ef6-5797-5126-a8c1-b9ab3fd97b66", 00:11:49.848 "is_configured": true, 00:11:49.848 "data_offset": 2048, 00:11:49.848 "data_size": 63488 00:11:49.848 } 00:11:49.848 ] 00:11:49.848 }' 00:11:49.848 23:45:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:49.848 23:45:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:49.848 23:45:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:49.848 23:45:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:49.848 23:45:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:11:50.785 23:45:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:11:50.786 23:45:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:50.786 23:45:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:50.786 23:45:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:50.786 23:45:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:50.786 23:45:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:50.786 23:45:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:50.786 23:45:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:50.786 23:45:38 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:50.786 23:45:38 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:50.786 23:45:38 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:51.045 23:45:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:51.045 "name": "raid_bdev1", 00:11:51.045 "uuid": "32f38c29-1954-4acc-8a9f-f847beecd7e8", 00:11:51.045 "strip_size_kb": 0, 00:11:51.045 "state": "online", 00:11:51.045 "raid_level": "raid1", 00:11:51.045 "superblock": true, 00:11:51.045 "num_base_bdevs": 4, 00:11:51.045 "num_base_bdevs_discovered": 3, 00:11:51.045 "num_base_bdevs_operational": 3, 00:11:51.045 "process": { 00:11:51.045 "type": "rebuild", 00:11:51.045 "target": "spare", 00:11:51.045 "progress": { 00:11:51.045 "blocks": 49152, 00:11:51.045 "percent": 77 00:11:51.045 } 00:11:51.045 }, 00:11:51.045 "base_bdevs_list": [ 00:11:51.045 { 00:11:51.045 "name": "spare", 00:11:51.045 "uuid": "b4091172-2cc8-5287-a78c-5cf3dae99027", 00:11:51.045 "is_configured": true, 00:11:51.045 "data_offset": 2048, 00:11:51.045 "data_size": 63488 00:11:51.045 }, 00:11:51.045 { 00:11:51.045 "name": null, 00:11:51.045 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:51.045 "is_configured": false, 00:11:51.045 "data_offset": 0, 00:11:51.045 "data_size": 63488 00:11:51.045 }, 00:11:51.045 { 00:11:51.045 "name": "BaseBdev3", 00:11:51.045 "uuid": "f56c9a76-6413-5d98-aff0-d9130d853e73", 00:11:51.045 "is_configured": true, 00:11:51.045 "data_offset": 2048, 00:11:51.045 "data_size": 63488 00:11:51.045 }, 00:11:51.045 { 00:11:51.045 "name": "BaseBdev4", 00:11:51.045 "uuid": "f3f95ef6-5797-5126-a8c1-b9ab3fd97b66", 00:11:51.045 "is_configured": true, 00:11:51.045 "data_offset": 2048, 00:11:51.045 "data_size": 63488 00:11:51.045 } 00:11:51.045 ] 00:11:51.045 }' 00:11:51.045 23:45:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:51.045 23:45:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:51.045 23:45:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:51.045 23:45:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:51.045 23:45:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:11:51.610 [2024-11-26 23:45:39.530750] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:11:51.610 [2024-11-26 23:45:39.530915] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:11:51.610 [2024-11-26 23:45:39.531038] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:52.177 23:45:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:11:52.177 23:45:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:52.177 23:45:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:52.177 23:45:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:52.177 23:45:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:52.177 23:45:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:52.177 23:45:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:52.177 23:45:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:52.177 23:45:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:52.177 23:45:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:52.177 23:45:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:52.177 23:45:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:52.177 "name": "raid_bdev1", 00:11:52.177 "uuid": "32f38c29-1954-4acc-8a9f-f847beecd7e8", 00:11:52.177 "strip_size_kb": 0, 00:11:52.177 "state": "online", 00:11:52.177 "raid_level": "raid1", 00:11:52.177 "superblock": true, 00:11:52.177 "num_base_bdevs": 4, 00:11:52.177 "num_base_bdevs_discovered": 3, 00:11:52.177 "num_base_bdevs_operational": 3, 00:11:52.177 "base_bdevs_list": [ 00:11:52.177 { 00:11:52.177 "name": "spare", 00:11:52.177 "uuid": "b4091172-2cc8-5287-a78c-5cf3dae99027", 00:11:52.177 "is_configured": true, 00:11:52.177 "data_offset": 2048, 00:11:52.177 "data_size": 63488 00:11:52.177 }, 00:11:52.177 { 00:11:52.177 "name": null, 00:11:52.177 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:52.177 "is_configured": false, 00:11:52.177 "data_offset": 0, 00:11:52.177 "data_size": 63488 00:11:52.177 }, 00:11:52.177 { 00:11:52.177 "name": "BaseBdev3", 00:11:52.177 "uuid": "f56c9a76-6413-5d98-aff0-d9130d853e73", 00:11:52.177 "is_configured": true, 00:11:52.177 "data_offset": 2048, 00:11:52.178 "data_size": 63488 00:11:52.178 }, 00:11:52.178 { 00:11:52.178 "name": "BaseBdev4", 00:11:52.178 "uuid": "f3f95ef6-5797-5126-a8c1-b9ab3fd97b66", 00:11:52.178 "is_configured": true, 00:11:52.178 "data_offset": 2048, 00:11:52.178 "data_size": 63488 00:11:52.178 } 00:11:52.178 ] 00:11:52.178 }' 00:11:52.178 23:45:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:52.178 23:45:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:11:52.178 23:45:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:52.178 23:45:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:11:52.178 23:45:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@709 -- # break 00:11:52.178 23:45:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:11:52.178 23:45:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:52.178 23:45:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:11:52.178 23:45:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:11:52.178 23:45:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:52.178 23:45:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:52.178 23:45:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:52.178 23:45:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:52.178 23:45:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:52.178 23:45:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:52.178 23:45:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:52.178 "name": "raid_bdev1", 00:11:52.178 "uuid": "32f38c29-1954-4acc-8a9f-f847beecd7e8", 00:11:52.178 "strip_size_kb": 0, 00:11:52.178 "state": "online", 00:11:52.178 "raid_level": "raid1", 00:11:52.178 "superblock": true, 00:11:52.178 "num_base_bdevs": 4, 00:11:52.178 "num_base_bdevs_discovered": 3, 00:11:52.178 "num_base_bdevs_operational": 3, 00:11:52.178 "base_bdevs_list": [ 00:11:52.178 { 00:11:52.178 "name": "spare", 00:11:52.178 "uuid": "b4091172-2cc8-5287-a78c-5cf3dae99027", 00:11:52.178 "is_configured": true, 00:11:52.178 "data_offset": 2048, 00:11:52.178 "data_size": 63488 00:11:52.178 }, 00:11:52.178 { 00:11:52.178 "name": null, 00:11:52.178 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:52.178 "is_configured": false, 00:11:52.178 "data_offset": 0, 00:11:52.178 "data_size": 63488 00:11:52.178 }, 00:11:52.178 { 00:11:52.178 "name": "BaseBdev3", 00:11:52.178 "uuid": "f56c9a76-6413-5d98-aff0-d9130d853e73", 00:11:52.178 "is_configured": true, 00:11:52.178 "data_offset": 2048, 00:11:52.178 "data_size": 63488 00:11:52.178 }, 00:11:52.178 { 00:11:52.178 "name": "BaseBdev4", 00:11:52.178 "uuid": "f3f95ef6-5797-5126-a8c1-b9ab3fd97b66", 00:11:52.178 "is_configured": true, 00:11:52.178 "data_offset": 2048, 00:11:52.178 "data_size": 63488 00:11:52.178 } 00:11:52.178 ] 00:11:52.178 }' 00:11:52.178 23:45:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:52.178 23:45:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:11:52.178 23:45:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:52.178 23:45:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:11:52.178 23:45:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:11:52.178 23:45:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:52.178 23:45:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:52.178 23:45:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:52.178 23:45:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:52.178 23:45:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:11:52.178 23:45:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:52.178 23:45:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:52.178 23:45:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:52.178 23:45:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:52.178 23:45:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:52.178 23:45:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:52.178 23:45:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:52.178 23:45:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:52.451 23:45:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:52.451 23:45:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:52.451 "name": "raid_bdev1", 00:11:52.451 "uuid": "32f38c29-1954-4acc-8a9f-f847beecd7e8", 00:11:52.451 "strip_size_kb": 0, 00:11:52.451 "state": "online", 00:11:52.451 "raid_level": "raid1", 00:11:52.451 "superblock": true, 00:11:52.451 "num_base_bdevs": 4, 00:11:52.451 "num_base_bdevs_discovered": 3, 00:11:52.451 "num_base_bdevs_operational": 3, 00:11:52.451 "base_bdevs_list": [ 00:11:52.451 { 00:11:52.451 "name": "spare", 00:11:52.451 "uuid": "b4091172-2cc8-5287-a78c-5cf3dae99027", 00:11:52.451 "is_configured": true, 00:11:52.451 "data_offset": 2048, 00:11:52.451 "data_size": 63488 00:11:52.451 }, 00:11:52.451 { 00:11:52.451 "name": null, 00:11:52.451 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:52.451 "is_configured": false, 00:11:52.451 "data_offset": 0, 00:11:52.451 "data_size": 63488 00:11:52.451 }, 00:11:52.451 { 00:11:52.451 "name": "BaseBdev3", 00:11:52.451 "uuid": "f56c9a76-6413-5d98-aff0-d9130d853e73", 00:11:52.451 "is_configured": true, 00:11:52.451 "data_offset": 2048, 00:11:52.451 "data_size": 63488 00:11:52.451 }, 00:11:52.451 { 00:11:52.451 "name": "BaseBdev4", 00:11:52.451 "uuid": "f3f95ef6-5797-5126-a8c1-b9ab3fd97b66", 00:11:52.451 "is_configured": true, 00:11:52.451 "data_offset": 2048, 00:11:52.451 "data_size": 63488 00:11:52.451 } 00:11:52.451 ] 00:11:52.451 }' 00:11:52.451 23:45:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:52.451 23:45:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:52.720 23:45:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:11:52.720 23:45:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:52.720 23:45:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:52.720 [2024-11-26 23:45:40.745094] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:11:52.720 [2024-11-26 23:45:40.745127] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:11:52.720 [2024-11-26 23:45:40.745215] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:11:52.720 [2024-11-26 23:45:40.745289] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:11:52.720 [2024-11-26 23:45:40.745302] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:11:52.720 23:45:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:52.720 23:45:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # jq length 00:11:52.720 23:45:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:52.720 23:45:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:52.720 23:45:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:52.720 23:45:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:52.720 23:45:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:11:52.720 23:45:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:11:52.720 23:45:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:11:52.720 23:45:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:11:52.720 23:45:40 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:11:52.720 23:45:40 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:11:52.720 23:45:40 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # local bdev_list 00:11:52.720 23:45:40 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:11:52.720 23:45:40 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # local nbd_list 00:11:52.720 23:45:40 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@12 -- # local i 00:11:52.720 23:45:40 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:11:52.720 23:45:40 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:11:52.720 23:45:40 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:11:52.980 /dev/nbd0 00:11:52.980 23:45:41 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:11:52.980 23:45:41 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:11:52.980 23:45:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:11:52.980 23:45:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@873 -- # local i 00:11:52.980 23:45:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:11:52.980 23:45:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:11:52.980 23:45:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:11:52.980 23:45:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@877 -- # break 00:11:52.980 23:45:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:11:52.980 23:45:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:11:52.980 23:45:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:11:52.980 1+0 records in 00:11:52.980 1+0 records out 00:11:52.980 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000369446 s, 11.1 MB/s 00:11:52.980 23:45:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:52.980 23:45:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@890 -- # size=4096 00:11:52.980 23:45:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:52.980 23:45:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:11:52.980 23:45:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@893 -- # return 0 00:11:52.980 23:45:41 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:11:52.980 23:45:41 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:11:52.980 23:45:41 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:11:53.240 /dev/nbd1 00:11:53.240 23:45:41 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:11:53.240 23:45:41 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:11:53.240 23:45:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:11:53.240 23:45:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@873 -- # local i 00:11:53.240 23:45:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:11:53.240 23:45:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:11:53.240 23:45:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:11:53.240 23:45:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@877 -- # break 00:11:53.240 23:45:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:11:53.240 23:45:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:11:53.240 23:45:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:11:53.240 1+0 records in 00:11:53.240 1+0 records out 00:11:53.240 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000391055 s, 10.5 MB/s 00:11:53.240 23:45:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:53.240 23:45:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@890 -- # size=4096 00:11:53.240 23:45:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:53.240 23:45:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:11:53.240 23:45:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@893 -- # return 0 00:11:53.240 23:45:41 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:11:53.240 23:45:41 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:11:53.240 23:45:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@738 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:11:53.240 23:45:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:11:53.240 23:45:41 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:11:53.240 23:45:41 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:11:53.240 23:45:41 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # local nbd_list 00:11:53.240 23:45:41 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@51 -- # local i 00:11:53.240 23:45:41 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:11:53.240 23:45:41 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:11:53.499 23:45:41 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:11:53.499 23:45:41 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:11:53.499 23:45:41 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:11:53.499 23:45:41 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:11:53.499 23:45:41 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:11:53.499 23:45:41 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:11:53.499 23:45:41 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:11:53.499 23:45:41 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:11:53.499 23:45:41 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:11:53.499 23:45:41 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:11:53.762 23:45:41 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:11:53.762 23:45:41 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:11:53.762 23:45:41 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:11:53.762 23:45:41 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:11:53.762 23:45:41 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:11:53.762 23:45:41 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:11:53.762 23:45:41 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:11:53.762 23:45:41 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:11:53.762 23:45:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@743 -- # '[' true = true ']' 00:11:53.762 23:45:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@745 -- # rpc_cmd bdev_passthru_delete spare 00:11:53.762 23:45:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:53.762 23:45:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:53.762 23:45:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:53.762 23:45:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@746 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:11:53.762 23:45:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:53.762 23:45:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:53.762 [2024-11-26 23:45:41.813379] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:11:53.762 [2024-11-26 23:45:41.813445] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:53.762 [2024-11-26 23:45:41.813466] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ae80 00:11:53.762 [2024-11-26 23:45:41.813480] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:53.762 [2024-11-26 23:45:41.815591] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:53.762 [2024-11-26 23:45:41.815675] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:11:53.762 [2024-11-26 23:45:41.815760] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:11:53.762 [2024-11-26 23:45:41.815805] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:11:53.762 [2024-11-26 23:45:41.815927] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:11:53.762 [2024-11-26 23:45:41.816034] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:11:53.762 spare 00:11:53.762 23:45:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:53.762 23:45:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@747 -- # rpc_cmd bdev_wait_for_examine 00:11:53.762 23:45:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:53.762 23:45:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:54.035 [2024-11-26 23:45:41.915920] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001580 00:11:54.035 [2024-11-26 23:45:41.915997] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:11:54.035 [2024-11-26 23:45:41.916284] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000caeb00 00:11:54.035 [2024-11-26 23:45:41.916450] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001580 00:11:54.035 [2024-11-26 23:45:41.916462] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001580 00:11:54.035 [2024-11-26 23:45:41.916595] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:54.035 23:45:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:54.035 23:45:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@749 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:11:54.035 23:45:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:54.035 23:45:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:54.035 23:45:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:54.035 23:45:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:54.035 23:45:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:11:54.035 23:45:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:54.035 23:45:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:54.035 23:45:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:54.035 23:45:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:54.035 23:45:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:54.035 23:45:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:54.035 23:45:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:54.035 23:45:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:54.035 23:45:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:54.035 23:45:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:54.035 "name": "raid_bdev1", 00:11:54.035 "uuid": "32f38c29-1954-4acc-8a9f-f847beecd7e8", 00:11:54.035 "strip_size_kb": 0, 00:11:54.035 "state": "online", 00:11:54.035 "raid_level": "raid1", 00:11:54.035 "superblock": true, 00:11:54.035 "num_base_bdevs": 4, 00:11:54.035 "num_base_bdevs_discovered": 3, 00:11:54.035 "num_base_bdevs_operational": 3, 00:11:54.035 "base_bdevs_list": [ 00:11:54.035 { 00:11:54.035 "name": "spare", 00:11:54.035 "uuid": "b4091172-2cc8-5287-a78c-5cf3dae99027", 00:11:54.035 "is_configured": true, 00:11:54.035 "data_offset": 2048, 00:11:54.035 "data_size": 63488 00:11:54.035 }, 00:11:54.035 { 00:11:54.035 "name": null, 00:11:54.035 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:54.035 "is_configured": false, 00:11:54.035 "data_offset": 2048, 00:11:54.035 "data_size": 63488 00:11:54.035 }, 00:11:54.035 { 00:11:54.035 "name": "BaseBdev3", 00:11:54.035 "uuid": "f56c9a76-6413-5d98-aff0-d9130d853e73", 00:11:54.035 "is_configured": true, 00:11:54.035 "data_offset": 2048, 00:11:54.035 "data_size": 63488 00:11:54.035 }, 00:11:54.035 { 00:11:54.035 "name": "BaseBdev4", 00:11:54.035 "uuid": "f3f95ef6-5797-5126-a8c1-b9ab3fd97b66", 00:11:54.035 "is_configured": true, 00:11:54.035 "data_offset": 2048, 00:11:54.035 "data_size": 63488 00:11:54.035 } 00:11:54.035 ] 00:11:54.035 }' 00:11:54.035 23:45:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:54.035 23:45:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:54.309 23:45:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@750 -- # verify_raid_bdev_process raid_bdev1 none none 00:11:54.309 23:45:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:54.309 23:45:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:11:54.309 23:45:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:11:54.309 23:45:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:54.309 23:45:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:54.309 23:45:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:54.309 23:45:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:54.309 23:45:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:54.309 23:45:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:54.309 23:45:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:54.309 "name": "raid_bdev1", 00:11:54.309 "uuid": "32f38c29-1954-4acc-8a9f-f847beecd7e8", 00:11:54.309 "strip_size_kb": 0, 00:11:54.309 "state": "online", 00:11:54.309 "raid_level": "raid1", 00:11:54.309 "superblock": true, 00:11:54.309 "num_base_bdevs": 4, 00:11:54.309 "num_base_bdevs_discovered": 3, 00:11:54.309 "num_base_bdevs_operational": 3, 00:11:54.309 "base_bdevs_list": [ 00:11:54.309 { 00:11:54.309 "name": "spare", 00:11:54.309 "uuid": "b4091172-2cc8-5287-a78c-5cf3dae99027", 00:11:54.309 "is_configured": true, 00:11:54.309 "data_offset": 2048, 00:11:54.309 "data_size": 63488 00:11:54.309 }, 00:11:54.309 { 00:11:54.309 "name": null, 00:11:54.309 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:54.309 "is_configured": false, 00:11:54.309 "data_offset": 2048, 00:11:54.309 "data_size": 63488 00:11:54.309 }, 00:11:54.309 { 00:11:54.309 "name": "BaseBdev3", 00:11:54.309 "uuid": "f56c9a76-6413-5d98-aff0-d9130d853e73", 00:11:54.309 "is_configured": true, 00:11:54.309 "data_offset": 2048, 00:11:54.309 "data_size": 63488 00:11:54.309 }, 00:11:54.309 { 00:11:54.309 "name": "BaseBdev4", 00:11:54.309 "uuid": "f3f95ef6-5797-5126-a8c1-b9ab3fd97b66", 00:11:54.309 "is_configured": true, 00:11:54.309 "data_offset": 2048, 00:11:54.309 "data_size": 63488 00:11:54.309 } 00:11:54.309 ] 00:11:54.309 }' 00:11:54.309 23:45:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:54.309 23:45:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:11:54.309 23:45:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:54.568 23:45:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:11:54.568 23:45:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:54.568 23:45:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:54.568 23:45:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:54.568 23:45:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # jq -r '.[].base_bdevs_list[0].name' 00:11:54.568 23:45:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:54.568 23:45:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # [[ spare == \s\p\a\r\e ]] 00:11:54.568 23:45:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@754 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:11:54.568 23:45:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:54.568 23:45:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:54.568 [2024-11-26 23:45:42.512222] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:11:54.568 23:45:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:54.568 23:45:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@755 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:11:54.568 23:45:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:54.568 23:45:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:54.568 23:45:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:54.568 23:45:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:54.568 23:45:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:11:54.568 23:45:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:54.568 23:45:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:54.568 23:45:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:54.568 23:45:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:54.568 23:45:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:54.568 23:45:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:54.568 23:45:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:54.568 23:45:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:54.568 23:45:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:54.568 23:45:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:54.568 "name": "raid_bdev1", 00:11:54.568 "uuid": "32f38c29-1954-4acc-8a9f-f847beecd7e8", 00:11:54.568 "strip_size_kb": 0, 00:11:54.568 "state": "online", 00:11:54.568 "raid_level": "raid1", 00:11:54.568 "superblock": true, 00:11:54.568 "num_base_bdevs": 4, 00:11:54.568 "num_base_bdevs_discovered": 2, 00:11:54.568 "num_base_bdevs_operational": 2, 00:11:54.568 "base_bdevs_list": [ 00:11:54.568 { 00:11:54.568 "name": null, 00:11:54.568 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:54.568 "is_configured": false, 00:11:54.568 "data_offset": 0, 00:11:54.568 "data_size": 63488 00:11:54.568 }, 00:11:54.568 { 00:11:54.568 "name": null, 00:11:54.568 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:54.568 "is_configured": false, 00:11:54.568 "data_offset": 2048, 00:11:54.568 "data_size": 63488 00:11:54.568 }, 00:11:54.568 { 00:11:54.568 "name": "BaseBdev3", 00:11:54.568 "uuid": "f56c9a76-6413-5d98-aff0-d9130d853e73", 00:11:54.568 "is_configured": true, 00:11:54.568 "data_offset": 2048, 00:11:54.568 "data_size": 63488 00:11:54.568 }, 00:11:54.568 { 00:11:54.568 "name": "BaseBdev4", 00:11:54.568 "uuid": "f3f95ef6-5797-5126-a8c1-b9ab3fd97b66", 00:11:54.568 "is_configured": true, 00:11:54.568 "data_offset": 2048, 00:11:54.568 "data_size": 63488 00:11:54.568 } 00:11:54.568 ] 00:11:54.568 }' 00:11:54.568 23:45:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:54.568 23:45:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:54.827 23:45:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@756 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:11:54.827 23:45:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:54.827 23:45:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:54.827 [2024-11-26 23:45:42.939502] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:11:54.827 [2024-11-26 23:45:42.939778] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (5) smaller than existing raid bdev raid_bdev1 (6) 00:11:54.827 [2024-11-26 23:45:42.939848] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:11:54.827 [2024-11-26 23:45:42.939927] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:11:54.827 [2024-11-26 23:45:42.943873] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000caebd0 00:11:54.827 23:45:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:54.827 23:45:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@757 -- # sleep 1 00:11:54.827 [2024-11-26 23:45:42.945852] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:11:56.205 23:45:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@758 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:56.205 23:45:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:56.205 23:45:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:56.205 23:45:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:56.205 23:45:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:56.205 23:45:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:56.205 23:45:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:56.205 23:45:43 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:56.205 23:45:43 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:56.205 23:45:43 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:56.205 23:45:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:56.205 "name": "raid_bdev1", 00:11:56.205 "uuid": "32f38c29-1954-4acc-8a9f-f847beecd7e8", 00:11:56.205 "strip_size_kb": 0, 00:11:56.205 "state": "online", 00:11:56.205 "raid_level": "raid1", 00:11:56.205 "superblock": true, 00:11:56.205 "num_base_bdevs": 4, 00:11:56.205 "num_base_bdevs_discovered": 3, 00:11:56.205 "num_base_bdevs_operational": 3, 00:11:56.205 "process": { 00:11:56.205 "type": "rebuild", 00:11:56.205 "target": "spare", 00:11:56.205 "progress": { 00:11:56.205 "blocks": 20480, 00:11:56.205 "percent": 32 00:11:56.205 } 00:11:56.205 }, 00:11:56.205 "base_bdevs_list": [ 00:11:56.205 { 00:11:56.205 "name": "spare", 00:11:56.205 "uuid": "b4091172-2cc8-5287-a78c-5cf3dae99027", 00:11:56.205 "is_configured": true, 00:11:56.205 "data_offset": 2048, 00:11:56.205 "data_size": 63488 00:11:56.205 }, 00:11:56.205 { 00:11:56.205 "name": null, 00:11:56.205 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:56.205 "is_configured": false, 00:11:56.205 "data_offset": 2048, 00:11:56.205 "data_size": 63488 00:11:56.205 }, 00:11:56.205 { 00:11:56.205 "name": "BaseBdev3", 00:11:56.205 "uuid": "f56c9a76-6413-5d98-aff0-d9130d853e73", 00:11:56.205 "is_configured": true, 00:11:56.205 "data_offset": 2048, 00:11:56.205 "data_size": 63488 00:11:56.205 }, 00:11:56.205 { 00:11:56.205 "name": "BaseBdev4", 00:11:56.205 "uuid": "f3f95ef6-5797-5126-a8c1-b9ab3fd97b66", 00:11:56.205 "is_configured": true, 00:11:56.205 "data_offset": 2048, 00:11:56.205 "data_size": 63488 00:11:56.205 } 00:11:56.205 ] 00:11:56.205 }' 00:11:56.205 23:45:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:56.205 23:45:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:56.205 23:45:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:56.205 23:45:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:56.205 23:45:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@761 -- # rpc_cmd bdev_passthru_delete spare 00:11:56.205 23:45:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:56.205 23:45:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:56.205 [2024-11-26 23:45:44.091331] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:11:56.205 [2024-11-26 23:45:44.149853] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:11:56.205 [2024-11-26 23:45:44.149971] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:56.205 [2024-11-26 23:45:44.149988] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:11:56.205 [2024-11-26 23:45:44.149997] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:11:56.205 23:45:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:56.205 23:45:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@762 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:11:56.205 23:45:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:56.205 23:45:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:56.205 23:45:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:56.205 23:45:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:56.205 23:45:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:11:56.205 23:45:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:56.205 23:45:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:56.205 23:45:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:56.205 23:45:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:56.205 23:45:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:56.205 23:45:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:56.205 23:45:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:56.205 23:45:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:56.205 23:45:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:56.205 23:45:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:56.205 "name": "raid_bdev1", 00:11:56.205 "uuid": "32f38c29-1954-4acc-8a9f-f847beecd7e8", 00:11:56.205 "strip_size_kb": 0, 00:11:56.205 "state": "online", 00:11:56.205 "raid_level": "raid1", 00:11:56.205 "superblock": true, 00:11:56.205 "num_base_bdevs": 4, 00:11:56.205 "num_base_bdevs_discovered": 2, 00:11:56.205 "num_base_bdevs_operational": 2, 00:11:56.205 "base_bdevs_list": [ 00:11:56.205 { 00:11:56.205 "name": null, 00:11:56.205 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:56.205 "is_configured": false, 00:11:56.205 "data_offset": 0, 00:11:56.205 "data_size": 63488 00:11:56.205 }, 00:11:56.205 { 00:11:56.205 "name": null, 00:11:56.205 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:56.205 "is_configured": false, 00:11:56.205 "data_offset": 2048, 00:11:56.205 "data_size": 63488 00:11:56.205 }, 00:11:56.205 { 00:11:56.205 "name": "BaseBdev3", 00:11:56.205 "uuid": "f56c9a76-6413-5d98-aff0-d9130d853e73", 00:11:56.205 "is_configured": true, 00:11:56.205 "data_offset": 2048, 00:11:56.205 "data_size": 63488 00:11:56.205 }, 00:11:56.205 { 00:11:56.205 "name": "BaseBdev4", 00:11:56.205 "uuid": "f3f95ef6-5797-5126-a8c1-b9ab3fd97b66", 00:11:56.205 "is_configured": true, 00:11:56.205 "data_offset": 2048, 00:11:56.205 "data_size": 63488 00:11:56.205 } 00:11:56.205 ] 00:11:56.205 }' 00:11:56.205 23:45:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:56.205 23:45:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:56.464 23:45:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@763 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:11:56.464 23:45:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:56.464 23:45:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:56.464 [2024-11-26 23:45:44.589426] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:11:56.464 [2024-11-26 23:45:44.589558] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:56.464 [2024-11-26 23:45:44.589598] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b780 00:11:56.464 [2024-11-26 23:45:44.589629] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:56.464 [2024-11-26 23:45:44.590066] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:56.464 [2024-11-26 23:45:44.590127] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:11:56.464 [2024-11-26 23:45:44.590245] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:11:56.464 [2024-11-26 23:45:44.590292] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (5) smaller than existing raid bdev raid_bdev1 (6) 00:11:56.464 [2024-11-26 23:45:44.590334] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:11:56.464 [2024-11-26 23:45:44.590430] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:11:56.724 [2024-11-26 23:45:44.594217] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000caeca0 00:11:56.724 spare 00:11:56.724 23:45:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:56.724 23:45:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@764 -- # sleep 1 00:11:56.724 [2024-11-26 23:45:44.596091] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:11:57.663 23:45:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@765 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:57.663 23:45:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:57.663 23:45:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:57.663 23:45:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:57.663 23:45:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:57.663 23:45:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:57.663 23:45:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:57.663 23:45:45 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:57.663 23:45:45 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:57.663 23:45:45 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:57.663 23:45:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:57.663 "name": "raid_bdev1", 00:11:57.663 "uuid": "32f38c29-1954-4acc-8a9f-f847beecd7e8", 00:11:57.663 "strip_size_kb": 0, 00:11:57.663 "state": "online", 00:11:57.663 "raid_level": "raid1", 00:11:57.663 "superblock": true, 00:11:57.663 "num_base_bdevs": 4, 00:11:57.663 "num_base_bdevs_discovered": 3, 00:11:57.663 "num_base_bdevs_operational": 3, 00:11:57.663 "process": { 00:11:57.663 "type": "rebuild", 00:11:57.663 "target": "spare", 00:11:57.663 "progress": { 00:11:57.663 "blocks": 20480, 00:11:57.663 "percent": 32 00:11:57.663 } 00:11:57.663 }, 00:11:57.663 "base_bdevs_list": [ 00:11:57.663 { 00:11:57.663 "name": "spare", 00:11:57.663 "uuid": "b4091172-2cc8-5287-a78c-5cf3dae99027", 00:11:57.663 "is_configured": true, 00:11:57.663 "data_offset": 2048, 00:11:57.663 "data_size": 63488 00:11:57.663 }, 00:11:57.663 { 00:11:57.663 "name": null, 00:11:57.663 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:57.663 "is_configured": false, 00:11:57.663 "data_offset": 2048, 00:11:57.663 "data_size": 63488 00:11:57.663 }, 00:11:57.663 { 00:11:57.663 "name": "BaseBdev3", 00:11:57.663 "uuid": "f56c9a76-6413-5d98-aff0-d9130d853e73", 00:11:57.663 "is_configured": true, 00:11:57.663 "data_offset": 2048, 00:11:57.663 "data_size": 63488 00:11:57.663 }, 00:11:57.663 { 00:11:57.663 "name": "BaseBdev4", 00:11:57.663 "uuid": "f3f95ef6-5797-5126-a8c1-b9ab3fd97b66", 00:11:57.663 "is_configured": true, 00:11:57.664 "data_offset": 2048, 00:11:57.664 "data_size": 63488 00:11:57.664 } 00:11:57.664 ] 00:11:57.664 }' 00:11:57.664 23:45:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:57.664 23:45:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:57.664 23:45:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:57.664 23:45:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:57.664 23:45:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@768 -- # rpc_cmd bdev_passthru_delete spare 00:11:57.664 23:45:45 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:57.664 23:45:45 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:57.664 [2024-11-26 23:45:45.760515] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:11:57.924 [2024-11-26 23:45:45.800344] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:11:57.924 [2024-11-26 23:45:45.800460] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:57.924 [2024-11-26 23:45:45.800482] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:11:57.924 [2024-11-26 23:45:45.800490] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:11:57.924 23:45:45 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:57.924 23:45:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@769 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:11:57.924 23:45:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:57.924 23:45:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:57.924 23:45:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:57.924 23:45:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:57.924 23:45:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:11:57.924 23:45:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:57.924 23:45:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:57.924 23:45:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:57.924 23:45:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:57.924 23:45:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:57.924 23:45:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:57.924 23:45:45 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:57.924 23:45:45 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:57.924 23:45:45 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:57.924 23:45:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:57.924 "name": "raid_bdev1", 00:11:57.924 "uuid": "32f38c29-1954-4acc-8a9f-f847beecd7e8", 00:11:57.924 "strip_size_kb": 0, 00:11:57.924 "state": "online", 00:11:57.924 "raid_level": "raid1", 00:11:57.924 "superblock": true, 00:11:57.924 "num_base_bdevs": 4, 00:11:57.924 "num_base_bdevs_discovered": 2, 00:11:57.924 "num_base_bdevs_operational": 2, 00:11:57.924 "base_bdevs_list": [ 00:11:57.924 { 00:11:57.924 "name": null, 00:11:57.924 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:57.924 "is_configured": false, 00:11:57.924 "data_offset": 0, 00:11:57.924 "data_size": 63488 00:11:57.924 }, 00:11:57.924 { 00:11:57.924 "name": null, 00:11:57.924 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:57.924 "is_configured": false, 00:11:57.924 "data_offset": 2048, 00:11:57.924 "data_size": 63488 00:11:57.924 }, 00:11:57.924 { 00:11:57.924 "name": "BaseBdev3", 00:11:57.924 "uuid": "f56c9a76-6413-5d98-aff0-d9130d853e73", 00:11:57.924 "is_configured": true, 00:11:57.924 "data_offset": 2048, 00:11:57.924 "data_size": 63488 00:11:57.924 }, 00:11:57.924 { 00:11:57.924 "name": "BaseBdev4", 00:11:57.924 "uuid": "f3f95ef6-5797-5126-a8c1-b9ab3fd97b66", 00:11:57.924 "is_configured": true, 00:11:57.924 "data_offset": 2048, 00:11:57.924 "data_size": 63488 00:11:57.924 } 00:11:57.924 ] 00:11:57.924 }' 00:11:57.924 23:45:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:57.924 23:45:45 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:58.183 23:45:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@770 -- # verify_raid_bdev_process raid_bdev1 none none 00:11:58.183 23:45:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:58.183 23:45:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:11:58.183 23:45:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:11:58.183 23:45:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:58.183 23:45:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:58.183 23:45:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:58.183 23:45:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:58.183 23:45:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:58.183 23:45:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:58.183 23:45:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:58.183 "name": "raid_bdev1", 00:11:58.183 "uuid": "32f38c29-1954-4acc-8a9f-f847beecd7e8", 00:11:58.183 "strip_size_kb": 0, 00:11:58.183 "state": "online", 00:11:58.183 "raid_level": "raid1", 00:11:58.183 "superblock": true, 00:11:58.183 "num_base_bdevs": 4, 00:11:58.183 "num_base_bdevs_discovered": 2, 00:11:58.183 "num_base_bdevs_operational": 2, 00:11:58.183 "base_bdevs_list": [ 00:11:58.183 { 00:11:58.183 "name": null, 00:11:58.183 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:58.183 "is_configured": false, 00:11:58.183 "data_offset": 0, 00:11:58.183 "data_size": 63488 00:11:58.183 }, 00:11:58.183 { 00:11:58.183 "name": null, 00:11:58.183 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:58.183 "is_configured": false, 00:11:58.183 "data_offset": 2048, 00:11:58.183 "data_size": 63488 00:11:58.183 }, 00:11:58.183 { 00:11:58.183 "name": "BaseBdev3", 00:11:58.183 "uuid": "f56c9a76-6413-5d98-aff0-d9130d853e73", 00:11:58.183 "is_configured": true, 00:11:58.183 "data_offset": 2048, 00:11:58.183 "data_size": 63488 00:11:58.183 }, 00:11:58.183 { 00:11:58.183 "name": "BaseBdev4", 00:11:58.183 "uuid": "f3f95ef6-5797-5126-a8c1-b9ab3fd97b66", 00:11:58.183 "is_configured": true, 00:11:58.183 "data_offset": 2048, 00:11:58.183 "data_size": 63488 00:11:58.183 } 00:11:58.183 ] 00:11:58.183 }' 00:11:58.183 23:45:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:58.183 23:45:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:11:58.183 23:45:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:58.467 23:45:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:11:58.467 23:45:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@773 -- # rpc_cmd bdev_passthru_delete BaseBdev1 00:11:58.467 23:45:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:58.467 23:45:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:58.467 23:45:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:58.467 23:45:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@774 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:11:58.467 23:45:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:58.467 23:45:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:58.467 [2024-11-26 23:45:46.363790] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:11:58.467 [2024-11-26 23:45:46.363894] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:58.467 [2024-11-26 23:45:46.363918] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000bd80 00:11:58.467 [2024-11-26 23:45:46.363929] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:58.467 [2024-11-26 23:45:46.364329] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:58.467 [2024-11-26 23:45:46.364346] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:11:58.467 [2024-11-26 23:45:46.364430] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:11:58.467 [2024-11-26 23:45:46.364443] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (6) 00:11:58.467 [2024-11-26 23:45:46.364453] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:11:58.467 [2024-11-26 23:45:46.364462] bdev_raid.c:3894:raid_bdev_examine_done: *ERROR*: Failed to examine bdev BaseBdev1: Invalid argument 00:11:58.467 BaseBdev1 00:11:58.467 23:45:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:58.467 23:45:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@775 -- # sleep 1 00:11:59.414 23:45:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@776 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:11:59.414 23:45:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:59.414 23:45:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:59.414 23:45:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:59.414 23:45:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:59.414 23:45:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:11:59.414 23:45:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:59.414 23:45:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:59.414 23:45:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:59.414 23:45:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:59.414 23:45:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:59.414 23:45:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:59.414 23:45:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:59.414 23:45:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:59.414 23:45:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:59.414 23:45:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:59.414 "name": "raid_bdev1", 00:11:59.414 "uuid": "32f38c29-1954-4acc-8a9f-f847beecd7e8", 00:11:59.414 "strip_size_kb": 0, 00:11:59.414 "state": "online", 00:11:59.414 "raid_level": "raid1", 00:11:59.414 "superblock": true, 00:11:59.414 "num_base_bdevs": 4, 00:11:59.414 "num_base_bdevs_discovered": 2, 00:11:59.414 "num_base_bdevs_operational": 2, 00:11:59.414 "base_bdevs_list": [ 00:11:59.414 { 00:11:59.414 "name": null, 00:11:59.414 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:59.414 "is_configured": false, 00:11:59.414 "data_offset": 0, 00:11:59.414 "data_size": 63488 00:11:59.414 }, 00:11:59.414 { 00:11:59.414 "name": null, 00:11:59.414 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:59.414 "is_configured": false, 00:11:59.414 "data_offset": 2048, 00:11:59.414 "data_size": 63488 00:11:59.414 }, 00:11:59.414 { 00:11:59.414 "name": "BaseBdev3", 00:11:59.414 "uuid": "f56c9a76-6413-5d98-aff0-d9130d853e73", 00:11:59.414 "is_configured": true, 00:11:59.414 "data_offset": 2048, 00:11:59.414 "data_size": 63488 00:11:59.414 }, 00:11:59.414 { 00:11:59.414 "name": "BaseBdev4", 00:11:59.414 "uuid": "f3f95ef6-5797-5126-a8c1-b9ab3fd97b66", 00:11:59.414 "is_configured": true, 00:11:59.414 "data_offset": 2048, 00:11:59.414 "data_size": 63488 00:11:59.414 } 00:11:59.414 ] 00:11:59.414 }' 00:11:59.414 23:45:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:59.414 23:45:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:59.984 23:45:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@777 -- # verify_raid_bdev_process raid_bdev1 none none 00:11:59.984 23:45:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:59.984 23:45:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:11:59.984 23:45:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:11:59.984 23:45:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:59.984 23:45:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:59.984 23:45:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:59.984 23:45:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:59.984 23:45:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:59.984 23:45:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:59.984 23:45:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:59.984 "name": "raid_bdev1", 00:11:59.984 "uuid": "32f38c29-1954-4acc-8a9f-f847beecd7e8", 00:11:59.984 "strip_size_kb": 0, 00:11:59.984 "state": "online", 00:11:59.984 "raid_level": "raid1", 00:11:59.984 "superblock": true, 00:11:59.984 "num_base_bdevs": 4, 00:11:59.984 "num_base_bdevs_discovered": 2, 00:11:59.984 "num_base_bdevs_operational": 2, 00:11:59.984 "base_bdevs_list": [ 00:11:59.984 { 00:11:59.984 "name": null, 00:11:59.984 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:59.984 "is_configured": false, 00:11:59.984 "data_offset": 0, 00:11:59.984 "data_size": 63488 00:11:59.984 }, 00:11:59.984 { 00:11:59.984 "name": null, 00:11:59.984 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:59.984 "is_configured": false, 00:11:59.984 "data_offset": 2048, 00:11:59.984 "data_size": 63488 00:11:59.984 }, 00:11:59.984 { 00:11:59.984 "name": "BaseBdev3", 00:11:59.984 "uuid": "f56c9a76-6413-5d98-aff0-d9130d853e73", 00:11:59.984 "is_configured": true, 00:11:59.984 "data_offset": 2048, 00:11:59.984 "data_size": 63488 00:11:59.984 }, 00:11:59.984 { 00:11:59.985 "name": "BaseBdev4", 00:11:59.985 "uuid": "f3f95ef6-5797-5126-a8c1-b9ab3fd97b66", 00:11:59.985 "is_configured": true, 00:11:59.985 "data_offset": 2048, 00:11:59.985 "data_size": 63488 00:11:59.985 } 00:11:59.985 ] 00:11:59.985 }' 00:11:59.985 23:45:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:59.985 23:45:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:11:59.985 23:45:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:59.985 23:45:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:11:59.985 23:45:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@778 -- # NOT rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:11:59.985 23:45:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@652 -- # local es=0 00:11:59.985 23:45:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:11:59.985 23:45:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:11:59.985 23:45:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:11:59.985 23:45:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:11:59.985 23:45:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:11:59.985 23:45:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:11:59.985 23:45:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:59.985 23:45:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:59.985 [2024-11-26 23:45:48.009069] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:11:59.985 [2024-11-26 23:45:48.009285] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (6) 00:11:59.985 [2024-11-26 23:45:48.009362] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:11:59.985 request: 00:11:59.985 { 00:11:59.985 "base_bdev": "BaseBdev1", 00:11:59.985 "raid_bdev": "raid_bdev1", 00:11:59.985 "method": "bdev_raid_add_base_bdev", 00:11:59.985 "req_id": 1 00:11:59.985 } 00:11:59.985 Got JSON-RPC error response 00:11:59.985 response: 00:11:59.985 { 00:11:59.985 "code": -22, 00:11:59.985 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:11:59.985 } 00:11:59.985 23:45:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:11:59.985 23:45:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@655 -- # es=1 00:11:59.985 23:45:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:11:59.985 23:45:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:11:59.985 23:45:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:11:59.985 23:45:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@779 -- # sleep 1 00:12:00.920 23:45:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@780 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:12:00.920 23:45:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:00.920 23:45:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:00.920 23:45:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:00.920 23:45:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:00.920 23:45:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:12:00.920 23:45:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:00.920 23:45:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:00.920 23:45:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:00.920 23:45:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:00.920 23:45:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:00.920 23:45:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:00.920 23:45:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:00.920 23:45:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:00.920 23:45:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:01.184 23:45:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:01.184 "name": "raid_bdev1", 00:12:01.184 "uuid": "32f38c29-1954-4acc-8a9f-f847beecd7e8", 00:12:01.184 "strip_size_kb": 0, 00:12:01.184 "state": "online", 00:12:01.184 "raid_level": "raid1", 00:12:01.184 "superblock": true, 00:12:01.184 "num_base_bdevs": 4, 00:12:01.184 "num_base_bdevs_discovered": 2, 00:12:01.184 "num_base_bdevs_operational": 2, 00:12:01.184 "base_bdevs_list": [ 00:12:01.184 { 00:12:01.184 "name": null, 00:12:01.184 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:01.184 "is_configured": false, 00:12:01.184 "data_offset": 0, 00:12:01.184 "data_size": 63488 00:12:01.184 }, 00:12:01.184 { 00:12:01.184 "name": null, 00:12:01.184 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:01.184 "is_configured": false, 00:12:01.184 "data_offset": 2048, 00:12:01.184 "data_size": 63488 00:12:01.184 }, 00:12:01.184 { 00:12:01.184 "name": "BaseBdev3", 00:12:01.184 "uuid": "f56c9a76-6413-5d98-aff0-d9130d853e73", 00:12:01.184 "is_configured": true, 00:12:01.184 "data_offset": 2048, 00:12:01.184 "data_size": 63488 00:12:01.184 }, 00:12:01.184 { 00:12:01.184 "name": "BaseBdev4", 00:12:01.184 "uuid": "f3f95ef6-5797-5126-a8c1-b9ab3fd97b66", 00:12:01.184 "is_configured": true, 00:12:01.184 "data_offset": 2048, 00:12:01.184 "data_size": 63488 00:12:01.184 } 00:12:01.184 ] 00:12:01.184 }' 00:12:01.184 23:45:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:01.184 23:45:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:01.449 23:45:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@781 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:01.449 23:45:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:01.449 23:45:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:01.449 23:45:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:01.449 23:45:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:01.449 23:45:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:01.449 23:45:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:01.449 23:45:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:01.449 23:45:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:01.449 23:45:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:01.449 23:45:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:01.449 "name": "raid_bdev1", 00:12:01.449 "uuid": "32f38c29-1954-4acc-8a9f-f847beecd7e8", 00:12:01.449 "strip_size_kb": 0, 00:12:01.449 "state": "online", 00:12:01.449 "raid_level": "raid1", 00:12:01.449 "superblock": true, 00:12:01.449 "num_base_bdevs": 4, 00:12:01.449 "num_base_bdevs_discovered": 2, 00:12:01.449 "num_base_bdevs_operational": 2, 00:12:01.449 "base_bdevs_list": [ 00:12:01.449 { 00:12:01.449 "name": null, 00:12:01.449 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:01.449 "is_configured": false, 00:12:01.449 "data_offset": 0, 00:12:01.449 "data_size": 63488 00:12:01.449 }, 00:12:01.449 { 00:12:01.449 "name": null, 00:12:01.449 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:01.449 "is_configured": false, 00:12:01.449 "data_offset": 2048, 00:12:01.449 "data_size": 63488 00:12:01.449 }, 00:12:01.449 { 00:12:01.449 "name": "BaseBdev3", 00:12:01.449 "uuid": "f56c9a76-6413-5d98-aff0-d9130d853e73", 00:12:01.449 "is_configured": true, 00:12:01.449 "data_offset": 2048, 00:12:01.449 "data_size": 63488 00:12:01.449 }, 00:12:01.449 { 00:12:01.449 "name": "BaseBdev4", 00:12:01.449 "uuid": "f3f95ef6-5797-5126-a8c1-b9ab3fd97b66", 00:12:01.449 "is_configured": true, 00:12:01.449 "data_offset": 2048, 00:12:01.449 "data_size": 63488 00:12:01.449 } 00:12:01.449 ] 00:12:01.449 }' 00:12:01.449 23:45:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:01.449 23:45:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:01.449 23:45:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:01.708 23:45:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:01.708 23:45:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@784 -- # killprocess 88300 00:12:01.708 23:45:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@954 -- # '[' -z 88300 ']' 00:12:01.708 23:45:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@958 -- # kill -0 88300 00:12:01.708 23:45:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@959 -- # uname 00:12:01.708 23:45:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:12:01.708 23:45:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 88300 00:12:01.708 killing process with pid 88300 00:12:01.708 Received shutdown signal, test time was about 60.000000 seconds 00:12:01.708 00:12:01.708 Latency(us) 00:12:01.708 [2024-11-26T23:45:49.840Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:12:01.708 [2024-11-26T23:45:49.840Z] =================================================================================================================== 00:12:01.708 [2024-11-26T23:45:49.840Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:12:01.708 23:45:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:12:01.708 23:45:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:12:01.708 23:45:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@972 -- # echo 'killing process with pid 88300' 00:12:01.708 23:45:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@973 -- # kill 88300 00:12:01.708 [2024-11-26 23:45:49.639094] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:12:01.708 [2024-11-26 23:45:49.639211] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:12:01.708 23:45:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@978 -- # wait 88300 00:12:01.708 [2024-11-26 23:45:49.639278] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:12:01.708 [2024-11-26 23:45:49.639291] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state offline 00:12:01.708 [2024-11-26 23:45:49.689796] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:12:01.968 23:45:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@786 -- # return 0 00:12:01.968 00:12:01.968 real 0m22.788s 00:12:01.968 user 0m27.999s 00:12:01.968 sys 0m3.453s 00:12:01.968 ************************************ 00:12:01.968 END TEST raid_rebuild_test_sb 00:12:01.968 ************************************ 00:12:01.968 23:45:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@1130 -- # xtrace_disable 00:12:01.968 23:45:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:01.968 23:45:49 bdev_raid -- bdev/bdev_raid.sh@980 -- # run_test raid_rebuild_test_io raid_rebuild_test raid1 4 false true true 00:12:01.968 23:45:49 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:12:01.968 23:45:49 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:12:01.968 23:45:49 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:12:01.968 ************************************ 00:12:01.968 START TEST raid_rebuild_test_io 00:12:01.968 ************************************ 00:12:01.968 23:45:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@1129 -- # raid_rebuild_test raid1 4 false true true 00:12:01.968 23:45:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:12:01.968 23:45:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=4 00:12:01.968 23:45:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@571 -- # local superblock=false 00:12:01.968 23:45:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@572 -- # local background_io=true 00:12:01.968 23:45:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@573 -- # local verify=true 00:12:01.968 23:45:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:12:01.968 23:45:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:01.968 23:45:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:12:01.968 23:45:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:01.968 23:45:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:01.968 23:45:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:12:01.968 23:45:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:01.968 23:45:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:01.968 23:45:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev3 00:12:01.968 23:45:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:01.968 23:45:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:01.968 23:45:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev4 00:12:01.968 23:45:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:01.968 23:45:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:01.968 23:45:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:12:01.968 23:45:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:12:01.968 23:45:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:12:01.968 23:45:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@576 -- # local strip_size 00:12:01.968 23:45:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@577 -- # local create_arg 00:12:01.968 23:45:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:12:01.968 23:45:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@579 -- # local data_offset 00:12:01.968 23:45:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:12:01.968 23:45:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:12:01.968 23:45:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@592 -- # '[' false = true ']' 00:12:01.968 23:45:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@597 -- # raid_pid=89036 00:12:01.968 23:45:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@598 -- # waitforlisten 89036 00:12:01.968 23:45:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:12:01.968 23:45:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@835 -- # '[' -z 89036 ']' 00:12:01.968 23:45:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:12:01.968 23:45:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@840 -- # local max_retries=100 00:12:01.968 23:45:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:12:01.968 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:12:01.968 23:45:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@844 -- # xtrace_disable 00:12:01.968 23:45:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:01.968 [2024-11-26 23:45:50.061400] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:12:01.968 [2024-11-26 23:45:50.061636] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid89036 ] 00:12:01.968 I/O size of 3145728 is greater than zero copy threshold (65536). 00:12:01.968 Zero copy mechanism will not be used. 00:12:02.227 [2024-11-26 23:45:50.216445] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:12:02.227 [2024-11-26 23:45:50.244685] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:12:02.227 [2024-11-26 23:45:50.286386] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:12:02.227 [2024-11-26 23:45:50.286422] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:12:02.796 23:45:50 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:12:02.796 23:45:50 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@868 -- # return 0 00:12:02.796 23:45:50 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:12:02.796 23:45:50 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:12:02.796 23:45:50 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:02.796 23:45:50 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:02.796 BaseBdev1_malloc 00:12:02.796 23:45:50 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:02.796 23:45:50 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:12:02.796 23:45:50 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:02.796 23:45:50 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:02.796 [2024-11-26 23:45:50.917683] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:12:02.796 [2024-11-26 23:45:50.917830] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:02.796 [2024-11-26 23:45:50.917886] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:12:02.796 [2024-11-26 23:45:50.917929] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:02.796 [2024-11-26 23:45:50.920001] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:02.796 [2024-11-26 23:45:50.920074] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:12:02.796 BaseBdev1 00:12:02.796 23:45:50 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:02.796 23:45:50 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:12:02.796 23:45:50 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:12:02.796 23:45:50 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:02.796 23:45:50 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:03.056 BaseBdev2_malloc 00:12:03.056 23:45:50 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:03.056 23:45:50 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:12:03.056 23:45:50 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:03.056 23:45:50 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:03.056 [2024-11-26 23:45:50.945998] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:12:03.056 [2024-11-26 23:45:50.946128] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:03.056 [2024-11-26 23:45:50.946169] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:12:03.056 [2024-11-26 23:45:50.946222] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:03.056 [2024-11-26 23:45:50.948225] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:03.056 [2024-11-26 23:45:50.948300] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:12:03.056 BaseBdev2 00:12:03.056 23:45:50 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:03.056 23:45:50 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:12:03.056 23:45:50 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:12:03.056 23:45:50 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:03.056 23:45:50 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:03.056 BaseBdev3_malloc 00:12:03.056 23:45:50 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:03.056 23:45:50 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev3_malloc -p BaseBdev3 00:12:03.056 23:45:50 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:03.056 23:45:50 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:03.056 [2024-11-26 23:45:50.974427] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev3_malloc 00:12:03.056 [2024-11-26 23:45:50.974483] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:03.056 [2024-11-26 23:45:50.974506] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:12:03.056 [2024-11-26 23:45:50.974516] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:03.056 [2024-11-26 23:45:50.976520] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:03.056 [2024-11-26 23:45:50.976601] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:12:03.056 BaseBdev3 00:12:03.056 23:45:50 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:03.056 23:45:50 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:12:03.056 23:45:50 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:12:03.056 23:45:50 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:03.056 23:45:50 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:03.056 BaseBdev4_malloc 00:12:03.056 23:45:51 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:03.056 23:45:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev4_malloc -p BaseBdev4 00:12:03.056 23:45:51 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:03.056 23:45:51 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:03.056 [2024-11-26 23:45:51.012100] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev4_malloc 00:12:03.056 [2024-11-26 23:45:51.012158] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:03.056 [2024-11-26 23:45:51.012182] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:12:03.056 [2024-11-26 23:45:51.012190] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:03.056 [2024-11-26 23:45:51.014187] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:03.057 [2024-11-26 23:45:51.014229] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:12:03.057 BaseBdev4 00:12:03.057 23:45:51 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:03.057 23:45:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:12:03.057 23:45:51 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:03.057 23:45:51 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:03.057 spare_malloc 00:12:03.057 23:45:51 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:03.057 23:45:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:12:03.057 23:45:51 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:03.057 23:45:51 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:03.057 spare_delay 00:12:03.057 23:45:51 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:03.057 23:45:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:12:03.057 23:45:51 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:03.057 23:45:51 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:03.057 [2024-11-26 23:45:51.052431] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:12:03.057 [2024-11-26 23:45:51.052482] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:03.057 [2024-11-26 23:45:51.052517] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009c80 00:12:03.057 [2024-11-26 23:45:51.052525] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:03.057 [2024-11-26 23:45:51.054529] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:03.057 [2024-11-26 23:45:51.054563] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:12:03.057 spare 00:12:03.057 23:45:51 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:03.057 23:45:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 00:12:03.057 23:45:51 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:03.057 23:45:51 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:03.057 [2024-11-26 23:45:51.064482] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:12:03.057 [2024-11-26 23:45:51.066318] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:12:03.057 [2024-11-26 23:45:51.066393] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:12:03.057 [2024-11-26 23:45:51.066440] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:12:03.057 [2024-11-26 23:45:51.066523] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:12:03.057 [2024-11-26 23:45:51.066536] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:12:03.057 [2024-11-26 23:45:51.066786] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:12:03.057 [2024-11-26 23:45:51.066923] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:12:03.057 [2024-11-26 23:45:51.066935] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:12:03.057 [2024-11-26 23:45:51.067048] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:03.057 23:45:51 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:03.057 23:45:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:12:03.057 23:45:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:03.057 23:45:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:03.057 23:45:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:03.057 23:45:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:03.057 23:45:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:12:03.057 23:45:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:03.057 23:45:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:03.057 23:45:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:03.057 23:45:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:03.057 23:45:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:03.057 23:45:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:03.057 23:45:51 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:03.057 23:45:51 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:03.057 23:45:51 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:03.057 23:45:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:03.057 "name": "raid_bdev1", 00:12:03.057 "uuid": "4650b49a-7c8d-4f7f-ad8c-c30b7a6941bf", 00:12:03.057 "strip_size_kb": 0, 00:12:03.057 "state": "online", 00:12:03.057 "raid_level": "raid1", 00:12:03.057 "superblock": false, 00:12:03.057 "num_base_bdevs": 4, 00:12:03.057 "num_base_bdevs_discovered": 4, 00:12:03.057 "num_base_bdevs_operational": 4, 00:12:03.057 "base_bdevs_list": [ 00:12:03.057 { 00:12:03.057 "name": "BaseBdev1", 00:12:03.057 "uuid": "9f675dd5-542b-5852-bccd-63a18ca8e27e", 00:12:03.057 "is_configured": true, 00:12:03.057 "data_offset": 0, 00:12:03.057 "data_size": 65536 00:12:03.057 }, 00:12:03.057 { 00:12:03.057 "name": "BaseBdev2", 00:12:03.057 "uuid": "8a192a92-adb4-5450-8a36-041c761e79b8", 00:12:03.057 "is_configured": true, 00:12:03.057 "data_offset": 0, 00:12:03.057 "data_size": 65536 00:12:03.057 }, 00:12:03.057 { 00:12:03.057 "name": "BaseBdev3", 00:12:03.057 "uuid": "608fda61-30ed-58a9-b933-0875d2911b28", 00:12:03.057 "is_configured": true, 00:12:03.057 "data_offset": 0, 00:12:03.057 "data_size": 65536 00:12:03.057 }, 00:12:03.057 { 00:12:03.057 "name": "BaseBdev4", 00:12:03.057 "uuid": "0f2c8ada-28b8-5e7e-897d-50f6ef189e2c", 00:12:03.057 "is_configured": true, 00:12:03.057 "data_offset": 0, 00:12:03.057 "data_size": 65536 00:12:03.057 } 00:12:03.057 ] 00:12:03.057 }' 00:12:03.057 23:45:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:03.057 23:45:51 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:03.625 23:45:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:12:03.625 23:45:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:12:03.625 23:45:51 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:03.625 23:45:51 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:03.625 [2024-11-26 23:45:51.575900] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:12:03.625 23:45:51 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:03.625 23:45:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=65536 00:12:03.625 23:45:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:12:03.625 23:45:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:03.625 23:45:51 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:03.625 23:45:51 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:03.625 23:45:51 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:03.625 23:45:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@619 -- # data_offset=0 00:12:03.625 23:45:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@621 -- # '[' true = true ']' 00:12:03.625 23:45:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:12:03.625 23:45:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@623 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:12:03.625 23:45:51 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:03.625 23:45:51 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:03.625 [2024-11-26 23:45:51.675399] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:12:03.625 23:45:51 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:03.625 23:45:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:12:03.625 23:45:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:03.625 23:45:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:03.625 23:45:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:03.625 23:45:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:03.625 23:45:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:03.625 23:45:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:03.625 23:45:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:03.625 23:45:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:03.625 23:45:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:03.625 23:45:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:03.625 23:45:51 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:03.625 23:45:51 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:03.625 23:45:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:03.625 23:45:51 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:03.625 23:45:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:03.625 "name": "raid_bdev1", 00:12:03.625 "uuid": "4650b49a-7c8d-4f7f-ad8c-c30b7a6941bf", 00:12:03.625 "strip_size_kb": 0, 00:12:03.625 "state": "online", 00:12:03.625 "raid_level": "raid1", 00:12:03.625 "superblock": false, 00:12:03.625 "num_base_bdevs": 4, 00:12:03.625 "num_base_bdevs_discovered": 3, 00:12:03.625 "num_base_bdevs_operational": 3, 00:12:03.625 "base_bdevs_list": [ 00:12:03.625 { 00:12:03.625 "name": null, 00:12:03.625 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:03.625 "is_configured": false, 00:12:03.625 "data_offset": 0, 00:12:03.625 "data_size": 65536 00:12:03.625 }, 00:12:03.625 { 00:12:03.625 "name": "BaseBdev2", 00:12:03.625 "uuid": "8a192a92-adb4-5450-8a36-041c761e79b8", 00:12:03.625 "is_configured": true, 00:12:03.625 "data_offset": 0, 00:12:03.625 "data_size": 65536 00:12:03.625 }, 00:12:03.625 { 00:12:03.625 "name": "BaseBdev3", 00:12:03.625 "uuid": "608fda61-30ed-58a9-b933-0875d2911b28", 00:12:03.625 "is_configured": true, 00:12:03.625 "data_offset": 0, 00:12:03.625 "data_size": 65536 00:12:03.625 }, 00:12:03.625 { 00:12:03.625 "name": "BaseBdev4", 00:12:03.625 "uuid": "0f2c8ada-28b8-5e7e-897d-50f6ef189e2c", 00:12:03.625 "is_configured": true, 00:12:03.625 "data_offset": 0, 00:12:03.625 "data_size": 65536 00:12:03.625 } 00:12:03.625 ] 00:12:03.625 }' 00:12:03.625 23:45:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:03.625 23:45:51 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:03.884 [2024-11-26 23:45:51.769186] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002870 00:12:03.884 I/O size of 3145728 is greater than zero copy threshold (65536). 00:12:03.884 Zero copy mechanism will not be used. 00:12:03.884 Running I/O for 60 seconds... 00:12:04.143 23:45:52 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:12:04.143 23:45:52 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:04.143 23:45:52 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:04.143 [2024-11-26 23:45:52.136093] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:04.143 23:45:52 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:04.143 23:45:52 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@647 -- # sleep 1 00:12:04.143 [2024-11-26 23:45:52.190045] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002940 00:12:04.143 [2024-11-26 23:45:52.192145] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:12:04.402 [2024-11-26 23:45:52.307040] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:12:04.402 [2024-11-26 23:45:52.307669] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:12:04.402 [2024-11-26 23:45:52.517940] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:12:04.402 [2024-11-26 23:45:52.518373] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:12:04.662 173.00 IOPS, 519.00 MiB/s [2024-11-26T23:45:52.794Z] [2024-11-26 23:45:52.771306] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:12:04.662 [2024-11-26 23:45:52.772502] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:12:04.921 [2024-11-26 23:45:52.986394] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:12:04.921 [2024-11-26 23:45:52.987176] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:12:05.181 23:45:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:05.181 23:45:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:05.181 23:45:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:05.181 23:45:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:05.181 23:45:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:05.181 23:45:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:05.181 23:45:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:05.181 23:45:53 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:05.181 23:45:53 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:05.181 23:45:53 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:05.181 23:45:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:05.181 "name": "raid_bdev1", 00:12:05.181 "uuid": "4650b49a-7c8d-4f7f-ad8c-c30b7a6941bf", 00:12:05.181 "strip_size_kb": 0, 00:12:05.181 "state": "online", 00:12:05.181 "raid_level": "raid1", 00:12:05.181 "superblock": false, 00:12:05.181 "num_base_bdevs": 4, 00:12:05.181 "num_base_bdevs_discovered": 4, 00:12:05.181 "num_base_bdevs_operational": 4, 00:12:05.181 "process": { 00:12:05.181 "type": "rebuild", 00:12:05.181 "target": "spare", 00:12:05.181 "progress": { 00:12:05.181 "blocks": 10240, 00:12:05.181 "percent": 15 00:12:05.181 } 00:12:05.181 }, 00:12:05.181 "base_bdevs_list": [ 00:12:05.181 { 00:12:05.181 "name": "spare", 00:12:05.181 "uuid": "6bf3b351-1513-5ed7-9bcf-3f8ddb5b7b07", 00:12:05.181 "is_configured": true, 00:12:05.181 "data_offset": 0, 00:12:05.181 "data_size": 65536 00:12:05.181 }, 00:12:05.181 { 00:12:05.181 "name": "BaseBdev2", 00:12:05.181 "uuid": "8a192a92-adb4-5450-8a36-041c761e79b8", 00:12:05.181 "is_configured": true, 00:12:05.181 "data_offset": 0, 00:12:05.181 "data_size": 65536 00:12:05.181 }, 00:12:05.181 { 00:12:05.181 "name": "BaseBdev3", 00:12:05.181 "uuid": "608fda61-30ed-58a9-b933-0875d2911b28", 00:12:05.181 "is_configured": true, 00:12:05.181 "data_offset": 0, 00:12:05.181 "data_size": 65536 00:12:05.181 }, 00:12:05.181 { 00:12:05.181 "name": "BaseBdev4", 00:12:05.181 "uuid": "0f2c8ada-28b8-5e7e-897d-50f6ef189e2c", 00:12:05.181 "is_configured": true, 00:12:05.181 "data_offset": 0, 00:12:05.181 "data_size": 65536 00:12:05.181 } 00:12:05.181 ] 00:12:05.181 }' 00:12:05.181 23:45:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:05.181 23:45:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:05.181 23:45:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:05.440 [2024-11-26 23:45:53.322624] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 14336 offset_begin: 12288 offset_end: 18432 00:12:05.440 23:45:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:05.440 23:45:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:12:05.440 23:45:53 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:05.440 23:45:53 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:05.440 [2024-11-26 23:45:53.338749] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:05.440 [2024-11-26 23:45:53.450314] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:12:05.440 [2024-11-26 23:45:53.564011] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:12:05.698 [2024-11-26 23:45:53.581874] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:05.698 [2024-11-26 23:45:53.581954] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:05.698 [2024-11-26 23:45:53.581970] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:12:05.698 [2024-11-26 23:45:53.600118] bdev_raid.c:1974:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 0 raid_ch: 0x60d000002870 00:12:05.698 23:45:53 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:05.698 23:45:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:12:05.698 23:45:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:05.698 23:45:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:05.698 23:45:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:05.698 23:45:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:05.698 23:45:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:05.698 23:45:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:05.699 23:45:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:05.699 23:45:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:05.699 23:45:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:05.699 23:45:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:05.699 23:45:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:05.699 23:45:53 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:05.699 23:45:53 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:05.699 23:45:53 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:05.699 23:45:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:05.699 "name": "raid_bdev1", 00:12:05.699 "uuid": "4650b49a-7c8d-4f7f-ad8c-c30b7a6941bf", 00:12:05.699 "strip_size_kb": 0, 00:12:05.699 "state": "online", 00:12:05.699 "raid_level": "raid1", 00:12:05.699 "superblock": false, 00:12:05.699 "num_base_bdevs": 4, 00:12:05.699 "num_base_bdevs_discovered": 3, 00:12:05.699 "num_base_bdevs_operational": 3, 00:12:05.699 "base_bdevs_list": [ 00:12:05.699 { 00:12:05.699 "name": null, 00:12:05.699 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:05.699 "is_configured": false, 00:12:05.699 "data_offset": 0, 00:12:05.699 "data_size": 65536 00:12:05.699 }, 00:12:05.699 { 00:12:05.699 "name": "BaseBdev2", 00:12:05.699 "uuid": "8a192a92-adb4-5450-8a36-041c761e79b8", 00:12:05.699 "is_configured": true, 00:12:05.699 "data_offset": 0, 00:12:05.699 "data_size": 65536 00:12:05.699 }, 00:12:05.699 { 00:12:05.699 "name": "BaseBdev3", 00:12:05.699 "uuid": "608fda61-30ed-58a9-b933-0875d2911b28", 00:12:05.699 "is_configured": true, 00:12:05.699 "data_offset": 0, 00:12:05.699 "data_size": 65536 00:12:05.699 }, 00:12:05.699 { 00:12:05.699 "name": "BaseBdev4", 00:12:05.699 "uuid": "0f2c8ada-28b8-5e7e-897d-50f6ef189e2c", 00:12:05.699 "is_configured": true, 00:12:05.699 "data_offset": 0, 00:12:05.699 "data_size": 65536 00:12:05.699 } 00:12:05.699 ] 00:12:05.699 }' 00:12:05.699 23:45:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:05.699 23:45:53 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:05.957 125.00 IOPS, 375.00 MiB/s [2024-11-26T23:45:54.089Z] 23:45:54 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:05.957 23:45:54 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:05.957 23:45:54 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:05.957 23:45:54 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:05.957 23:45:54 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:05.957 23:45:54 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:05.957 23:45:54 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:05.957 23:45:54 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:05.957 23:45:54 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:05.957 23:45:54 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:05.957 23:45:54 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:05.957 "name": "raid_bdev1", 00:12:05.957 "uuid": "4650b49a-7c8d-4f7f-ad8c-c30b7a6941bf", 00:12:05.957 "strip_size_kb": 0, 00:12:05.957 "state": "online", 00:12:05.957 "raid_level": "raid1", 00:12:05.957 "superblock": false, 00:12:05.957 "num_base_bdevs": 4, 00:12:05.957 "num_base_bdevs_discovered": 3, 00:12:05.957 "num_base_bdevs_operational": 3, 00:12:05.957 "base_bdevs_list": [ 00:12:05.957 { 00:12:05.957 "name": null, 00:12:05.957 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:05.957 "is_configured": false, 00:12:05.957 "data_offset": 0, 00:12:05.957 "data_size": 65536 00:12:05.957 }, 00:12:05.957 { 00:12:05.957 "name": "BaseBdev2", 00:12:05.957 "uuid": "8a192a92-adb4-5450-8a36-041c761e79b8", 00:12:05.957 "is_configured": true, 00:12:05.957 "data_offset": 0, 00:12:05.957 "data_size": 65536 00:12:05.957 }, 00:12:05.957 { 00:12:05.957 "name": "BaseBdev3", 00:12:05.957 "uuid": "608fda61-30ed-58a9-b933-0875d2911b28", 00:12:05.957 "is_configured": true, 00:12:05.957 "data_offset": 0, 00:12:05.957 "data_size": 65536 00:12:05.957 }, 00:12:05.957 { 00:12:05.957 "name": "BaseBdev4", 00:12:05.957 "uuid": "0f2c8ada-28b8-5e7e-897d-50f6ef189e2c", 00:12:05.957 "is_configured": true, 00:12:05.957 "data_offset": 0, 00:12:05.957 "data_size": 65536 00:12:05.957 } 00:12:05.957 ] 00:12:05.957 }' 00:12:05.957 23:45:54 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:06.215 23:45:54 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:06.215 23:45:54 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:06.215 23:45:54 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:06.215 23:45:54 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:12:06.215 23:45:54 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:06.215 23:45:54 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:06.215 [2024-11-26 23:45:54.162506] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:06.215 23:45:54 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:06.215 23:45:54 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@663 -- # sleep 1 00:12:06.215 [2024-11-26 23:45:54.217573] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002a10 00:12:06.215 [2024-11-26 23:45:54.219586] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:12:06.215 [2024-11-26 23:45:54.327108] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:12:06.215 [2024-11-26 23:45:54.327652] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:12:06.474 [2024-11-26 23:45:54.560496] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:12:06.731 157.33 IOPS, 472.00 MiB/s [2024-11-26T23:45:54.863Z] [2024-11-26 23:45:54.810119] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:12:06.995 [2024-11-26 23:45:54.937247] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:12:06.995 [2024-11-26 23:45:54.937935] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:12:07.259 23:45:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:07.259 23:45:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:07.259 23:45:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:07.259 23:45:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:07.259 23:45:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:07.259 23:45:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:07.259 23:45:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:07.259 23:45:55 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:07.259 23:45:55 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:07.259 23:45:55 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:07.259 23:45:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:07.259 "name": "raid_bdev1", 00:12:07.259 "uuid": "4650b49a-7c8d-4f7f-ad8c-c30b7a6941bf", 00:12:07.259 "strip_size_kb": 0, 00:12:07.259 "state": "online", 00:12:07.259 "raid_level": "raid1", 00:12:07.259 "superblock": false, 00:12:07.259 "num_base_bdevs": 4, 00:12:07.259 "num_base_bdevs_discovered": 4, 00:12:07.259 "num_base_bdevs_operational": 4, 00:12:07.259 "process": { 00:12:07.259 "type": "rebuild", 00:12:07.259 "target": "spare", 00:12:07.259 "progress": { 00:12:07.259 "blocks": 12288, 00:12:07.259 "percent": 18 00:12:07.259 } 00:12:07.259 }, 00:12:07.259 "base_bdevs_list": [ 00:12:07.259 { 00:12:07.259 "name": "spare", 00:12:07.259 "uuid": "6bf3b351-1513-5ed7-9bcf-3f8ddb5b7b07", 00:12:07.259 "is_configured": true, 00:12:07.259 "data_offset": 0, 00:12:07.259 "data_size": 65536 00:12:07.259 }, 00:12:07.259 { 00:12:07.259 "name": "BaseBdev2", 00:12:07.259 "uuid": "8a192a92-adb4-5450-8a36-041c761e79b8", 00:12:07.259 "is_configured": true, 00:12:07.259 "data_offset": 0, 00:12:07.259 "data_size": 65536 00:12:07.259 }, 00:12:07.259 { 00:12:07.259 "name": "BaseBdev3", 00:12:07.259 "uuid": "608fda61-30ed-58a9-b933-0875d2911b28", 00:12:07.259 "is_configured": true, 00:12:07.259 "data_offset": 0, 00:12:07.259 "data_size": 65536 00:12:07.259 }, 00:12:07.259 { 00:12:07.259 "name": "BaseBdev4", 00:12:07.259 "uuid": "0f2c8ada-28b8-5e7e-897d-50f6ef189e2c", 00:12:07.259 "is_configured": true, 00:12:07.259 "data_offset": 0, 00:12:07.259 "data_size": 65536 00:12:07.259 } 00:12:07.259 ] 00:12:07.259 }' 00:12:07.259 23:45:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:07.259 [2024-11-26 23:45:55.278487] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 14336 offset_begin: 12288 offset_end: 18432 00:12:07.259 23:45:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:07.259 23:45:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:07.259 23:45:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:07.259 23:45:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@666 -- # '[' false = true ']' 00:12:07.259 23:45:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=4 00:12:07.259 23:45:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:12:07.259 23:45:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@693 -- # '[' 4 -gt 2 ']' 00:12:07.259 23:45:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@695 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:12:07.259 23:45:55 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:07.259 23:45:55 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:07.259 [2024-11-26 23:45:55.336922] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:12:07.517 [2024-11-26 23:45:55.392985] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:12:07.517 [2024-11-26 23:45:55.507240] bdev_raid.c:1974:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 1 raid_ch: 0x60d000002870 00:12:07.517 [2024-11-26 23:45:55.507290] bdev_raid.c:1974:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 1 raid_ch: 0x60d000002a10 00:12:07.517 [2024-11-26 23:45:55.507357] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:12:07.517 23:45:55 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:07.517 23:45:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@698 -- # base_bdevs[1]= 00:12:07.517 23:45:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@699 -- # (( num_base_bdevs_operational-- )) 00:12:07.517 23:45:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@702 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:07.517 23:45:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:07.517 23:45:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:07.517 23:45:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:07.517 23:45:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:07.517 23:45:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:07.517 23:45:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:07.517 23:45:55 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:07.517 23:45:55 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:07.517 23:45:55 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:07.517 23:45:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:07.517 "name": "raid_bdev1", 00:12:07.517 "uuid": "4650b49a-7c8d-4f7f-ad8c-c30b7a6941bf", 00:12:07.517 "strip_size_kb": 0, 00:12:07.517 "state": "online", 00:12:07.517 "raid_level": "raid1", 00:12:07.517 "superblock": false, 00:12:07.517 "num_base_bdevs": 4, 00:12:07.517 "num_base_bdevs_discovered": 3, 00:12:07.517 "num_base_bdevs_operational": 3, 00:12:07.517 "process": { 00:12:07.517 "type": "rebuild", 00:12:07.517 "target": "spare", 00:12:07.517 "progress": { 00:12:07.517 "blocks": 16384, 00:12:07.517 "percent": 25 00:12:07.517 } 00:12:07.517 }, 00:12:07.517 "base_bdevs_list": [ 00:12:07.517 { 00:12:07.517 "name": "spare", 00:12:07.517 "uuid": "6bf3b351-1513-5ed7-9bcf-3f8ddb5b7b07", 00:12:07.517 "is_configured": true, 00:12:07.517 "data_offset": 0, 00:12:07.517 "data_size": 65536 00:12:07.517 }, 00:12:07.517 { 00:12:07.517 "name": null, 00:12:07.517 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:07.517 "is_configured": false, 00:12:07.517 "data_offset": 0, 00:12:07.517 "data_size": 65536 00:12:07.517 }, 00:12:07.517 { 00:12:07.517 "name": "BaseBdev3", 00:12:07.517 "uuid": "608fda61-30ed-58a9-b933-0875d2911b28", 00:12:07.517 "is_configured": true, 00:12:07.517 "data_offset": 0, 00:12:07.517 "data_size": 65536 00:12:07.517 }, 00:12:07.517 { 00:12:07.517 "name": "BaseBdev4", 00:12:07.517 "uuid": "0f2c8ada-28b8-5e7e-897d-50f6ef189e2c", 00:12:07.517 "is_configured": true, 00:12:07.517 "data_offset": 0, 00:12:07.517 "data_size": 65536 00:12:07.517 } 00:12:07.517 ] 00:12:07.517 }' 00:12:07.517 23:45:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:07.517 23:45:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:07.517 23:45:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:07.776 23:45:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:07.776 23:45:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@706 -- # local timeout=387 00:12:07.776 23:45:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:12:07.776 23:45:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:07.776 23:45:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:07.776 23:45:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:07.776 23:45:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:07.776 23:45:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:07.776 23:45:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:07.776 23:45:55 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:07.776 23:45:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:07.776 23:45:55 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:07.776 23:45:55 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:07.776 23:45:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:07.776 "name": "raid_bdev1", 00:12:07.776 "uuid": "4650b49a-7c8d-4f7f-ad8c-c30b7a6941bf", 00:12:07.776 "strip_size_kb": 0, 00:12:07.776 "state": "online", 00:12:07.776 "raid_level": "raid1", 00:12:07.776 "superblock": false, 00:12:07.776 "num_base_bdevs": 4, 00:12:07.776 "num_base_bdevs_discovered": 3, 00:12:07.776 "num_base_bdevs_operational": 3, 00:12:07.776 "process": { 00:12:07.776 "type": "rebuild", 00:12:07.776 "target": "spare", 00:12:07.776 "progress": { 00:12:07.776 "blocks": 18432, 00:12:07.776 "percent": 28 00:12:07.776 } 00:12:07.776 }, 00:12:07.776 "base_bdevs_list": [ 00:12:07.776 { 00:12:07.776 "name": "spare", 00:12:07.776 "uuid": "6bf3b351-1513-5ed7-9bcf-3f8ddb5b7b07", 00:12:07.776 "is_configured": true, 00:12:07.776 "data_offset": 0, 00:12:07.776 "data_size": 65536 00:12:07.776 }, 00:12:07.776 { 00:12:07.776 "name": null, 00:12:07.776 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:07.776 "is_configured": false, 00:12:07.776 "data_offset": 0, 00:12:07.776 "data_size": 65536 00:12:07.776 }, 00:12:07.776 { 00:12:07.776 "name": "BaseBdev3", 00:12:07.776 "uuid": "608fda61-30ed-58a9-b933-0875d2911b28", 00:12:07.776 "is_configured": true, 00:12:07.776 "data_offset": 0, 00:12:07.776 "data_size": 65536 00:12:07.776 }, 00:12:07.776 { 00:12:07.776 "name": "BaseBdev4", 00:12:07.776 "uuid": "0f2c8ada-28b8-5e7e-897d-50f6ef189e2c", 00:12:07.776 "is_configured": true, 00:12:07.776 "data_offset": 0, 00:12:07.776 "data_size": 65536 00:12:07.776 } 00:12:07.776 ] 00:12:07.776 }' 00:12:07.776 23:45:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:07.776 23:45:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:07.776 23:45:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:07.776 136.25 IOPS, 408.75 MiB/s [2024-11-26T23:45:55.908Z] 23:45:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:07.776 23:45:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:12:07.776 [2024-11-26 23:45:55.856211] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 22528 offset_begin: 18432 offset_end: 24576 00:12:08.344 [2024-11-26 23:45:56.187401] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 28672 offset_begin: 24576 offset_end: 30720 00:12:08.603 [2024-11-26 23:45:56.623557] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 34816 offset_begin: 30720 offset_end: 36864 00:12:08.860 117.40 IOPS, 352.20 MiB/s [2024-11-26T23:45:56.992Z] 23:45:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:12:08.860 23:45:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:08.860 23:45:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:08.860 23:45:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:08.860 23:45:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:08.860 23:45:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:08.860 23:45:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:08.860 23:45:56 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:08.860 23:45:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:08.860 23:45:56 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:08.860 23:45:56 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:08.860 23:45:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:08.860 "name": "raid_bdev1", 00:12:08.860 "uuid": "4650b49a-7c8d-4f7f-ad8c-c30b7a6941bf", 00:12:08.860 "strip_size_kb": 0, 00:12:08.860 "state": "online", 00:12:08.860 "raid_level": "raid1", 00:12:08.860 "superblock": false, 00:12:08.860 "num_base_bdevs": 4, 00:12:08.860 "num_base_bdevs_discovered": 3, 00:12:08.860 "num_base_bdevs_operational": 3, 00:12:08.860 "process": { 00:12:08.860 "type": "rebuild", 00:12:08.860 "target": "spare", 00:12:08.860 "progress": { 00:12:08.860 "blocks": 34816, 00:12:08.860 "percent": 53 00:12:08.860 } 00:12:08.860 }, 00:12:08.860 "base_bdevs_list": [ 00:12:08.860 { 00:12:08.860 "name": "spare", 00:12:08.860 "uuid": "6bf3b351-1513-5ed7-9bcf-3f8ddb5b7b07", 00:12:08.861 "is_configured": true, 00:12:08.861 "data_offset": 0, 00:12:08.861 "data_size": 65536 00:12:08.861 }, 00:12:08.861 { 00:12:08.861 "name": null, 00:12:08.861 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:08.861 "is_configured": false, 00:12:08.861 "data_offset": 0, 00:12:08.861 "data_size": 65536 00:12:08.861 }, 00:12:08.861 { 00:12:08.861 "name": "BaseBdev3", 00:12:08.861 "uuid": "608fda61-30ed-58a9-b933-0875d2911b28", 00:12:08.861 "is_configured": true, 00:12:08.861 "data_offset": 0, 00:12:08.861 "data_size": 65536 00:12:08.861 }, 00:12:08.861 { 00:12:08.861 "name": "BaseBdev4", 00:12:08.861 "uuid": "0f2c8ada-28b8-5e7e-897d-50f6ef189e2c", 00:12:08.861 "is_configured": true, 00:12:08.861 "data_offset": 0, 00:12:08.861 "data_size": 65536 00:12:08.861 } 00:12:08.861 ] 00:12:08.861 }' 00:12:08.861 23:45:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:08.861 23:45:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:08.861 23:45:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:08.861 [2024-11-26 23:45:56.958429] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 38912 offset_begin: 36864 offset_end: 43008 00:12:08.861 23:45:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:08.861 23:45:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:12:09.119 [2024-11-26 23:45:57.188004] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 40960 offset_begin: 36864 offset_end: 43008 00:12:09.688 [2024-11-26 23:45:57.523388] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 47104 offset_begin: 43008 offset_end: 49152 00:12:09.946 104.33 IOPS, 313.00 MiB/s [2024-11-26T23:45:58.078Z] 23:45:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:12:09.947 23:45:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:09.947 23:45:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:09.947 23:45:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:09.947 23:45:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:09.947 23:45:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:09.947 23:45:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:09.947 23:45:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:09.947 23:45:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:09.947 23:45:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:09.947 23:45:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:09.947 23:45:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:09.947 "name": "raid_bdev1", 00:12:09.947 "uuid": "4650b49a-7c8d-4f7f-ad8c-c30b7a6941bf", 00:12:09.947 "strip_size_kb": 0, 00:12:09.947 "state": "online", 00:12:09.947 "raid_level": "raid1", 00:12:09.947 "superblock": false, 00:12:09.947 "num_base_bdevs": 4, 00:12:09.947 "num_base_bdevs_discovered": 3, 00:12:09.947 "num_base_bdevs_operational": 3, 00:12:09.947 "process": { 00:12:09.947 "type": "rebuild", 00:12:09.947 "target": "spare", 00:12:09.947 "progress": { 00:12:09.947 "blocks": 53248, 00:12:09.947 "percent": 81 00:12:09.947 } 00:12:09.947 }, 00:12:09.947 "base_bdevs_list": [ 00:12:09.947 { 00:12:09.947 "name": "spare", 00:12:09.947 "uuid": "6bf3b351-1513-5ed7-9bcf-3f8ddb5b7b07", 00:12:09.947 "is_configured": true, 00:12:09.947 "data_offset": 0, 00:12:09.947 "data_size": 65536 00:12:09.947 }, 00:12:09.947 { 00:12:09.947 "name": null, 00:12:09.947 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:09.947 "is_configured": false, 00:12:09.947 "data_offset": 0, 00:12:09.947 "data_size": 65536 00:12:09.947 }, 00:12:09.947 { 00:12:09.947 "name": "BaseBdev3", 00:12:09.947 "uuid": "608fda61-30ed-58a9-b933-0875d2911b28", 00:12:09.947 "is_configured": true, 00:12:09.947 "data_offset": 0, 00:12:09.947 "data_size": 65536 00:12:09.947 }, 00:12:09.947 { 00:12:09.947 "name": "BaseBdev4", 00:12:09.947 "uuid": "0f2c8ada-28b8-5e7e-897d-50f6ef189e2c", 00:12:09.947 "is_configured": true, 00:12:09.947 "data_offset": 0, 00:12:09.947 "data_size": 65536 00:12:09.947 } 00:12:09.947 ] 00:12:09.947 }' 00:12:09.947 23:45:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:09.947 23:45:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:09.947 23:45:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:10.205 23:45:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:10.205 23:45:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:12:10.464 [2024-11-26 23:45:58.593381] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:12:10.722 [2024-11-26 23:45:58.614581] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:12:10.722 [2024-11-26 23:45:58.616746] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:11.291 94.86 IOPS, 284.57 MiB/s [2024-11-26T23:45:59.423Z] 23:45:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:12:11.291 23:45:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:11.291 23:45:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:11.291 23:45:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:11.291 23:45:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:11.291 23:45:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:11.291 23:45:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:11.291 23:45:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:11.291 23:45:59 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:11.291 23:45:59 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:11.291 23:45:59 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:11.291 23:45:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:11.291 "name": "raid_bdev1", 00:12:11.291 "uuid": "4650b49a-7c8d-4f7f-ad8c-c30b7a6941bf", 00:12:11.291 "strip_size_kb": 0, 00:12:11.291 "state": "online", 00:12:11.291 "raid_level": "raid1", 00:12:11.291 "superblock": false, 00:12:11.291 "num_base_bdevs": 4, 00:12:11.291 "num_base_bdevs_discovered": 3, 00:12:11.291 "num_base_bdevs_operational": 3, 00:12:11.291 "base_bdevs_list": [ 00:12:11.291 { 00:12:11.291 "name": "spare", 00:12:11.291 "uuid": "6bf3b351-1513-5ed7-9bcf-3f8ddb5b7b07", 00:12:11.291 "is_configured": true, 00:12:11.291 "data_offset": 0, 00:12:11.291 "data_size": 65536 00:12:11.291 }, 00:12:11.291 { 00:12:11.291 "name": null, 00:12:11.291 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:11.291 "is_configured": false, 00:12:11.291 "data_offset": 0, 00:12:11.291 "data_size": 65536 00:12:11.291 }, 00:12:11.291 { 00:12:11.291 "name": "BaseBdev3", 00:12:11.291 "uuid": "608fda61-30ed-58a9-b933-0875d2911b28", 00:12:11.291 "is_configured": true, 00:12:11.291 "data_offset": 0, 00:12:11.291 "data_size": 65536 00:12:11.291 }, 00:12:11.291 { 00:12:11.291 "name": "BaseBdev4", 00:12:11.291 "uuid": "0f2c8ada-28b8-5e7e-897d-50f6ef189e2c", 00:12:11.291 "is_configured": true, 00:12:11.291 "data_offset": 0, 00:12:11.291 "data_size": 65536 00:12:11.291 } 00:12:11.291 ] 00:12:11.291 }' 00:12:11.291 23:45:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:11.291 23:45:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:12:11.291 23:45:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:11.291 23:45:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:12:11.291 23:45:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@709 -- # break 00:12:11.291 23:45:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:11.291 23:45:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:11.291 23:45:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:11.291 23:45:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:11.291 23:45:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:11.291 23:45:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:11.291 23:45:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:11.291 23:45:59 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:11.291 23:45:59 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:11.291 23:45:59 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:11.291 23:45:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:11.291 "name": "raid_bdev1", 00:12:11.291 "uuid": "4650b49a-7c8d-4f7f-ad8c-c30b7a6941bf", 00:12:11.291 "strip_size_kb": 0, 00:12:11.291 "state": "online", 00:12:11.291 "raid_level": "raid1", 00:12:11.291 "superblock": false, 00:12:11.291 "num_base_bdevs": 4, 00:12:11.291 "num_base_bdevs_discovered": 3, 00:12:11.291 "num_base_bdevs_operational": 3, 00:12:11.291 "base_bdevs_list": [ 00:12:11.291 { 00:12:11.291 "name": "spare", 00:12:11.291 "uuid": "6bf3b351-1513-5ed7-9bcf-3f8ddb5b7b07", 00:12:11.291 "is_configured": true, 00:12:11.291 "data_offset": 0, 00:12:11.291 "data_size": 65536 00:12:11.291 }, 00:12:11.291 { 00:12:11.291 "name": null, 00:12:11.291 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:11.291 "is_configured": false, 00:12:11.291 "data_offset": 0, 00:12:11.291 "data_size": 65536 00:12:11.291 }, 00:12:11.291 { 00:12:11.291 "name": "BaseBdev3", 00:12:11.291 "uuid": "608fda61-30ed-58a9-b933-0875d2911b28", 00:12:11.291 "is_configured": true, 00:12:11.291 "data_offset": 0, 00:12:11.291 "data_size": 65536 00:12:11.291 }, 00:12:11.291 { 00:12:11.291 "name": "BaseBdev4", 00:12:11.291 "uuid": "0f2c8ada-28b8-5e7e-897d-50f6ef189e2c", 00:12:11.291 "is_configured": true, 00:12:11.291 "data_offset": 0, 00:12:11.291 "data_size": 65536 00:12:11.291 } 00:12:11.291 ] 00:12:11.291 }' 00:12:11.291 23:45:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:11.291 23:45:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:11.291 23:45:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:11.291 23:45:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:11.291 23:45:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:12:11.291 23:45:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:11.291 23:45:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:11.291 23:45:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:11.291 23:45:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:11.291 23:45:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:11.291 23:45:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:11.291 23:45:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:11.291 23:45:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:11.291 23:45:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:11.291 23:45:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:11.291 23:45:59 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:11.291 23:45:59 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:11.291 23:45:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:11.550 23:45:59 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:11.550 23:45:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:11.550 "name": "raid_bdev1", 00:12:11.550 "uuid": "4650b49a-7c8d-4f7f-ad8c-c30b7a6941bf", 00:12:11.550 "strip_size_kb": 0, 00:12:11.550 "state": "online", 00:12:11.550 "raid_level": "raid1", 00:12:11.550 "superblock": false, 00:12:11.550 "num_base_bdevs": 4, 00:12:11.550 "num_base_bdevs_discovered": 3, 00:12:11.550 "num_base_bdevs_operational": 3, 00:12:11.550 "base_bdevs_list": [ 00:12:11.550 { 00:12:11.550 "name": "spare", 00:12:11.550 "uuid": "6bf3b351-1513-5ed7-9bcf-3f8ddb5b7b07", 00:12:11.550 "is_configured": true, 00:12:11.550 "data_offset": 0, 00:12:11.550 "data_size": 65536 00:12:11.550 }, 00:12:11.550 { 00:12:11.550 "name": null, 00:12:11.550 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:11.550 "is_configured": false, 00:12:11.550 "data_offset": 0, 00:12:11.550 "data_size": 65536 00:12:11.550 }, 00:12:11.550 { 00:12:11.550 "name": "BaseBdev3", 00:12:11.550 "uuid": "608fda61-30ed-58a9-b933-0875d2911b28", 00:12:11.550 "is_configured": true, 00:12:11.550 "data_offset": 0, 00:12:11.550 "data_size": 65536 00:12:11.550 }, 00:12:11.550 { 00:12:11.550 "name": "BaseBdev4", 00:12:11.550 "uuid": "0f2c8ada-28b8-5e7e-897d-50f6ef189e2c", 00:12:11.550 "is_configured": true, 00:12:11.550 "data_offset": 0, 00:12:11.550 "data_size": 65536 00:12:11.550 } 00:12:11.550 ] 00:12:11.550 }' 00:12:11.550 23:45:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:11.550 23:45:59 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:11.810 86.88 IOPS, 260.62 MiB/s [2024-11-26T23:45:59.942Z] 23:45:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:12:11.810 23:45:59 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:11.810 23:45:59 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:11.810 [2024-11-26 23:45:59.840778] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:12:11.810 [2024-11-26 23:45:59.840890] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:12:11.810 00:12:11.810 Latency(us) 00:12:11.810 [2024-11-26T23:45:59.942Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:12:11.810 Job: raid_bdev1 (Core Mask 0x1, workload: randrw, percentage: 50, depth: 2, IO size: 3145728) 00:12:11.810 raid_bdev1 : 8.15 85.74 257.21 0.00 0.00 15302.94 282.61 109436.53 00:12:11.810 [2024-11-26T23:45:59.942Z] =================================================================================================================== 00:12:11.810 [2024-11-26T23:45:59.942Z] Total : 85.74 257.21 0.00 0.00 15302.94 282.61 109436.53 00:12:11.810 [2024-11-26 23:45:59.912103] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:12:11.810 [2024-11-26 23:45:59.912204] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:11.810 [2024-11-26 23:45:59.912325] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:12:11.810 [2024-11-26 23:45:59.912388] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:12:11.810 { 00:12:11.810 "results": [ 00:12:11.810 { 00:12:11.810 "job": "raid_bdev1", 00:12:11.810 "core_mask": "0x1", 00:12:11.810 "workload": "randrw", 00:12:11.810 "percentage": 50, 00:12:11.810 "status": "finished", 00:12:11.810 "queue_depth": 2, 00:12:11.810 "io_size": 3145728, 00:12:11.810 "runtime": 8.152972, 00:12:11.810 "iops": 85.73560659842816, 00:12:11.810 "mibps": 257.2068197952845, 00:12:11.810 "io_failed": 0, 00:12:11.810 "io_timeout": 0, 00:12:11.810 "avg_latency_us": 15302.941159860311, 00:12:11.810 "min_latency_us": 282.6061135371179, 00:12:11.810 "max_latency_us": 109436.5344978166 00:12:11.810 } 00:12:11.810 ], 00:12:11.810 "core_count": 1 00:12:11.810 } 00:12:11.810 23:45:59 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:11.810 23:45:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:11.810 23:45:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@720 -- # jq length 00:12:11.810 23:45:59 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:11.810 23:45:59 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:11.810 23:45:59 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:12.070 23:45:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:12:12.070 23:45:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:12:12.070 23:45:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@723 -- # '[' true = true ']' 00:12:12.070 23:45:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@725 -- # nbd_start_disks /var/tmp/spdk.sock spare /dev/nbd0 00:12:12.070 23:45:59 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:12:12.070 23:45:59 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # bdev_list=('spare') 00:12:12.070 23:45:59 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:12:12.070 23:45:59 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:12:12.070 23:45:59 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:12:12.070 23:45:59 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@12 -- # local i 00:12:12.070 23:45:59 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:12:12.070 23:45:59 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:12:12.070 23:45:59 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd0 00:12:12.070 /dev/nbd0 00:12:12.329 23:46:00 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:12:12.329 23:46:00 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:12:12.329 23:46:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:12:12.329 23:46:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@873 -- # local i 00:12:12.329 23:46:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:12:12.329 23:46:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:12:12.329 23:46:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:12:12.330 23:46:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@877 -- # break 00:12:12.330 23:46:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:12:12.330 23:46:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:12:12.330 23:46:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:12:12.330 1+0 records in 00:12:12.330 1+0 records out 00:12:12.330 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000589888 s, 6.9 MB/s 00:12:12.330 23:46:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:12.330 23:46:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@890 -- # size=4096 00:12:12.330 23:46:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:12.330 23:46:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:12:12.330 23:46:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@893 -- # return 0 00:12:12.330 23:46:00 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:12:12.330 23:46:00 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:12:12.330 23:46:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@726 -- # for bdev in "${base_bdevs[@]:1}" 00:12:12.330 23:46:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@727 -- # '[' -z '' ']' 00:12:12.330 23:46:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@728 -- # continue 00:12:12.330 23:46:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@726 -- # for bdev in "${base_bdevs[@]:1}" 00:12:12.330 23:46:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@727 -- # '[' -z BaseBdev3 ']' 00:12:12.330 23:46:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@730 -- # nbd_start_disks /var/tmp/spdk.sock BaseBdev3 /dev/nbd1 00:12:12.330 23:46:00 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:12:12.330 23:46:00 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev3') 00:12:12.330 23:46:00 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:12:12.330 23:46:00 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd1') 00:12:12.330 23:46:00 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:12:12.330 23:46:00 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@12 -- # local i 00:12:12.330 23:46:00 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:12:12.330 23:46:00 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:12:12.330 23:46:00 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev3 /dev/nbd1 00:12:12.330 /dev/nbd1 00:12:12.588 23:46:00 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:12:12.588 23:46:00 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:12:12.588 23:46:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:12:12.588 23:46:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@873 -- # local i 00:12:12.588 23:46:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:12:12.588 23:46:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:12:12.588 23:46:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:12:12.588 23:46:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@877 -- # break 00:12:12.588 23:46:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:12:12.588 23:46:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:12:12.588 23:46:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:12:12.588 1+0 records in 00:12:12.588 1+0 records out 00:12:12.588 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000278537 s, 14.7 MB/s 00:12:12.588 23:46:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:12.588 23:46:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@890 -- # size=4096 00:12:12.588 23:46:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:12.588 23:46:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:12:12.588 23:46:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@893 -- # return 0 00:12:12.588 23:46:00 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:12:12.588 23:46:00 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:12:12.588 23:46:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@731 -- # cmp -i 0 /dev/nbd0 /dev/nbd1 00:12:12.588 23:46:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@732 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd1 00:12:12.588 23:46:00 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:12:12.588 23:46:00 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd1') 00:12:12.588 23:46:00 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:12:12.588 23:46:00 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@51 -- # local i 00:12:12.588 23:46:00 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:12:12.588 23:46:00 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:12:12.848 23:46:00 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:12:12.848 23:46:00 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:12:12.848 23:46:00 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:12:12.848 23:46:00 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:12:12.848 23:46:00 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:12:12.848 23:46:00 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:12:12.848 23:46:00 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@41 -- # break 00:12:12.848 23:46:00 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@45 -- # return 0 00:12:12.848 23:46:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@726 -- # for bdev in "${base_bdevs[@]:1}" 00:12:12.848 23:46:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@727 -- # '[' -z BaseBdev4 ']' 00:12:12.848 23:46:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@730 -- # nbd_start_disks /var/tmp/spdk.sock BaseBdev4 /dev/nbd1 00:12:12.848 23:46:00 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:12:12.848 23:46:00 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev4') 00:12:12.848 23:46:00 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:12:12.848 23:46:00 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd1') 00:12:12.848 23:46:00 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:12:12.848 23:46:00 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@12 -- # local i 00:12:12.848 23:46:00 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:12:12.848 23:46:00 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:12:12.848 23:46:00 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev4 /dev/nbd1 00:12:12.848 /dev/nbd1 00:12:13.108 23:46:00 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:12:13.108 23:46:00 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:12:13.108 23:46:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:12:13.108 23:46:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@873 -- # local i 00:12:13.108 23:46:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:12:13.108 23:46:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:12:13.108 23:46:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:12:13.108 23:46:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@877 -- # break 00:12:13.108 23:46:01 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:12:13.108 23:46:01 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:12:13.108 23:46:01 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:12:13.108 1+0 records in 00:12:13.108 1+0 records out 00:12:13.108 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000353436 s, 11.6 MB/s 00:12:13.108 23:46:01 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:13.108 23:46:01 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@890 -- # size=4096 00:12:13.108 23:46:01 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:13.108 23:46:01 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:12:13.108 23:46:01 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@893 -- # return 0 00:12:13.108 23:46:01 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:12:13.108 23:46:01 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:12:13.108 23:46:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@731 -- # cmp -i 0 /dev/nbd0 /dev/nbd1 00:12:13.108 23:46:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@732 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd1 00:12:13.108 23:46:01 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:12:13.108 23:46:01 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd1') 00:12:13.108 23:46:01 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:12:13.108 23:46:01 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@51 -- # local i 00:12:13.108 23:46:01 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:12:13.108 23:46:01 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:12:13.368 23:46:01 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:12:13.368 23:46:01 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:12:13.368 23:46:01 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:12:13.368 23:46:01 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:12:13.368 23:46:01 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:12:13.368 23:46:01 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:12:13.368 23:46:01 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@41 -- # break 00:12:13.368 23:46:01 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@45 -- # return 0 00:12:13.368 23:46:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@734 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:12:13.368 23:46:01 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:12:13.368 23:46:01 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:12:13.368 23:46:01 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:12:13.368 23:46:01 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@51 -- # local i 00:12:13.368 23:46:01 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:12:13.368 23:46:01 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:12:13.628 23:46:01 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:12:13.628 23:46:01 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:12:13.628 23:46:01 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:12:13.628 23:46:01 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:12:13.628 23:46:01 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:12:13.628 23:46:01 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:12:13.628 23:46:01 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@41 -- # break 00:12:13.628 23:46:01 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@45 -- # return 0 00:12:13.628 23:46:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@743 -- # '[' false = true ']' 00:12:13.628 23:46:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@784 -- # killprocess 89036 00:12:13.628 23:46:01 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@954 -- # '[' -z 89036 ']' 00:12:13.628 23:46:01 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@958 -- # kill -0 89036 00:12:13.628 23:46:01 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@959 -- # uname 00:12:13.628 23:46:01 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:12:13.628 23:46:01 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 89036 00:12:13.628 23:46:01 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:12:13.628 23:46:01 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:12:13.628 23:46:01 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@972 -- # echo 'killing process with pid 89036' 00:12:13.628 killing process with pid 89036 00:12:13.628 23:46:01 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@973 -- # kill 89036 00:12:13.628 Received shutdown signal, test time was about 9.797664 seconds 00:12:13.628 00:12:13.628 Latency(us) 00:12:13.628 [2024-11-26T23:46:01.760Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:12:13.628 [2024-11-26T23:46:01.760Z] =================================================================================================================== 00:12:13.628 [2024-11-26T23:46:01.760Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:12:13.628 23:46:01 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@978 -- # wait 89036 00:12:13.628 [2024-11-26 23:46:01.550185] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:12:13.628 [2024-11-26 23:46:01.596290] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:12:13.888 ************************************ 00:12:13.888 END TEST raid_rebuild_test_io 00:12:13.888 ************************************ 00:12:13.888 23:46:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@786 -- # return 0 00:12:13.888 00:12:13.888 real 0m11.831s 00:12:13.888 user 0m15.409s 00:12:13.888 sys 0m1.708s 00:12:13.888 23:46:01 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@1130 -- # xtrace_disable 00:12:13.888 23:46:01 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:13.888 23:46:01 bdev_raid -- bdev/bdev_raid.sh@981 -- # run_test raid_rebuild_test_sb_io raid_rebuild_test raid1 4 true true true 00:12:13.888 23:46:01 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:12:13.888 23:46:01 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:12:13.888 23:46:01 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:12:13.888 ************************************ 00:12:13.888 START TEST raid_rebuild_test_sb_io 00:12:13.888 ************************************ 00:12:13.888 23:46:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@1129 -- # raid_rebuild_test raid1 4 true true true 00:12:13.888 23:46:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:12:13.888 23:46:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=4 00:12:13.888 23:46:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@571 -- # local superblock=true 00:12:13.888 23:46:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@572 -- # local background_io=true 00:12:13.888 23:46:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@573 -- # local verify=true 00:12:13.888 23:46:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:12:13.888 23:46:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:13.888 23:46:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:12:13.888 23:46:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:13.888 23:46:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:13.888 23:46:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:12:13.888 23:46:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:13.888 23:46:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:13.888 23:46:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev3 00:12:13.888 23:46:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:13.888 23:46:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:13.888 23:46:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev4 00:12:13.888 23:46:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:13.888 23:46:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:13.888 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:12:13.888 23:46:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:12:13.888 23:46:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:12:13.888 23:46:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:12:13.888 23:46:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@576 -- # local strip_size 00:12:13.888 23:46:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@577 -- # local create_arg 00:12:13.888 23:46:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:12:13.888 23:46:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@579 -- # local data_offset 00:12:13.888 23:46:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:12:13.888 23:46:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:12:13.889 23:46:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@592 -- # '[' true = true ']' 00:12:13.889 23:46:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@593 -- # create_arg+=' -s' 00:12:13.889 23:46:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@597 -- # raid_pid=89429 00:12:13.889 23:46:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@598 -- # waitforlisten 89429 00:12:13.889 23:46:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@835 -- # '[' -z 89429 ']' 00:12:13.889 23:46:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:12:13.889 23:46:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@840 -- # local max_retries=100 00:12:13.889 23:46:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:12:13.889 23:46:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@844 -- # xtrace_disable 00:12:13.889 23:46:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:13.889 23:46:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:12:13.889 I/O size of 3145728 is greater than zero copy threshold (65536). 00:12:13.889 Zero copy mechanism will not be used. 00:12:13.889 [2024-11-26 23:46:01.951176] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:12:13.889 [2024-11-26 23:46:01.951315] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid89429 ] 00:12:14.152 [2024-11-26 23:46:02.105934] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:12:14.152 [2024-11-26 23:46:02.131124] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:12:14.152 [2024-11-26 23:46:02.172871] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:12:14.152 [2024-11-26 23:46:02.172909] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:12:14.727 23:46:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:12:14.727 23:46:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@868 -- # return 0 00:12:14.727 23:46:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:12:14.727 23:46:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:12:14.727 23:46:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:14.727 23:46:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:14.727 BaseBdev1_malloc 00:12:14.727 23:46:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:14.727 23:46:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:12:14.727 23:46:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:14.727 23:46:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:14.727 [2024-11-26 23:46:02.795857] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:12:14.727 [2024-11-26 23:46:02.795999] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:14.727 [2024-11-26 23:46:02.796044] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:12:14.727 [2024-11-26 23:46:02.796087] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:14.727 [2024-11-26 23:46:02.798205] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:14.727 [2024-11-26 23:46:02.798284] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:12:14.727 BaseBdev1 00:12:14.727 23:46:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:14.727 23:46:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:12:14.727 23:46:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:12:14.727 23:46:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:14.727 23:46:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:14.727 BaseBdev2_malloc 00:12:14.727 23:46:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:14.727 23:46:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:12:14.727 23:46:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:14.727 23:46:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:14.727 [2024-11-26 23:46:02.824467] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:12:14.727 [2024-11-26 23:46:02.824605] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:14.727 [2024-11-26 23:46:02.824645] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:12:14.727 [2024-11-26 23:46:02.824673] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:14.727 [2024-11-26 23:46:02.826733] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:14.727 [2024-11-26 23:46:02.826809] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:12:14.727 BaseBdev2 00:12:14.727 23:46:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:14.727 23:46:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:12:14.727 23:46:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:12:14.727 23:46:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:14.727 23:46:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:14.727 BaseBdev3_malloc 00:12:14.727 23:46:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:14.727 23:46:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev3_malloc -p BaseBdev3 00:12:14.727 23:46:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:14.727 23:46:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:14.727 [2024-11-26 23:46:02.853103] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev3_malloc 00:12:14.727 [2024-11-26 23:46:02.853176] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:14.727 [2024-11-26 23:46:02.853199] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:12:14.727 [2024-11-26 23:46:02.853210] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:14.727 [2024-11-26 23:46:02.855462] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:14.727 [2024-11-26 23:46:02.855499] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:12:14.987 BaseBdev3 00:12:14.987 23:46:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:14.987 23:46:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:12:14.987 23:46:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:12:14.987 23:46:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:14.987 23:46:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:14.987 BaseBdev4_malloc 00:12:14.987 23:46:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:14.987 23:46:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev4_malloc -p BaseBdev4 00:12:14.987 23:46:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:14.987 23:46:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:14.987 [2024-11-26 23:46:02.900854] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev4_malloc 00:12:14.987 [2024-11-26 23:46:02.900930] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:14.987 [2024-11-26 23:46:02.900951] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:12:14.987 [2024-11-26 23:46:02.900959] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:14.987 [2024-11-26 23:46:02.903018] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:14.987 [2024-11-26 23:46:02.903054] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:12:14.987 BaseBdev4 00:12:14.987 23:46:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:14.987 23:46:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:12:14.987 23:46:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:14.987 23:46:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:14.987 spare_malloc 00:12:14.987 23:46:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:14.987 23:46:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:12:14.987 23:46:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:14.987 23:46:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:14.987 spare_delay 00:12:14.987 23:46:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:14.987 23:46:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:12:14.987 23:46:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:14.987 23:46:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:14.987 [2024-11-26 23:46:02.933568] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:12:14.987 [2024-11-26 23:46:02.933623] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:14.987 [2024-11-26 23:46:02.933642] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009c80 00:12:14.987 [2024-11-26 23:46:02.933651] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:14.987 [2024-11-26 23:46:02.935785] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:14.987 [2024-11-26 23:46:02.935820] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:12:14.987 spare 00:12:14.987 23:46:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:14.987 23:46:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 00:12:14.987 23:46:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:14.987 23:46:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:14.987 [2024-11-26 23:46:02.941638] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:12:14.987 [2024-11-26 23:46:02.943593] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:12:14.987 [2024-11-26 23:46:02.943657] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:12:14.987 [2024-11-26 23:46:02.943703] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:12:14.987 [2024-11-26 23:46:02.943873] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:12:14.987 [2024-11-26 23:46:02.943892] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:12:14.987 [2024-11-26 23:46:02.944119] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:12:14.987 [2024-11-26 23:46:02.944295] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:12:14.987 [2024-11-26 23:46:02.944315] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:12:14.987 [2024-11-26 23:46:02.944441] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:14.987 23:46:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:14.987 23:46:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:12:14.987 23:46:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:14.987 23:46:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:14.987 23:46:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:14.987 23:46:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:14.987 23:46:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:12:14.987 23:46:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:14.987 23:46:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:14.987 23:46:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:14.987 23:46:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:14.987 23:46:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:14.987 23:46:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:14.987 23:46:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:14.987 23:46:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:14.987 23:46:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:14.987 23:46:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:14.987 "name": "raid_bdev1", 00:12:14.987 "uuid": "1ac4fc07-7ba3-43ad-97f9-b1256b4b254d", 00:12:14.987 "strip_size_kb": 0, 00:12:14.987 "state": "online", 00:12:14.987 "raid_level": "raid1", 00:12:14.988 "superblock": true, 00:12:14.988 "num_base_bdevs": 4, 00:12:14.988 "num_base_bdevs_discovered": 4, 00:12:14.988 "num_base_bdevs_operational": 4, 00:12:14.988 "base_bdevs_list": [ 00:12:14.988 { 00:12:14.988 "name": "BaseBdev1", 00:12:14.988 "uuid": "7e9314c8-bc83-5979-9b5b-5ad7c813ce29", 00:12:14.988 "is_configured": true, 00:12:14.988 "data_offset": 2048, 00:12:14.988 "data_size": 63488 00:12:14.988 }, 00:12:14.988 { 00:12:14.988 "name": "BaseBdev2", 00:12:14.988 "uuid": "5197f646-d67b-5e09-af5d-c993bb98c4a5", 00:12:14.988 "is_configured": true, 00:12:14.988 "data_offset": 2048, 00:12:14.988 "data_size": 63488 00:12:14.988 }, 00:12:14.988 { 00:12:14.988 "name": "BaseBdev3", 00:12:14.988 "uuid": "ee6dcd27-62fb-55ac-bd58-5d03fde51b7b", 00:12:14.988 "is_configured": true, 00:12:14.988 "data_offset": 2048, 00:12:14.988 "data_size": 63488 00:12:14.988 }, 00:12:14.988 { 00:12:14.988 "name": "BaseBdev4", 00:12:14.988 "uuid": "5a25f2b0-5c65-5f8f-baeb-76b7b36d0157", 00:12:14.988 "is_configured": true, 00:12:14.988 "data_offset": 2048, 00:12:14.988 "data_size": 63488 00:12:14.988 } 00:12:14.988 ] 00:12:14.988 }' 00:12:14.988 23:46:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:14.988 23:46:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:15.247 23:46:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:12:15.247 23:46:03 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:15.247 23:46:03 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:15.247 23:46:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:12:15.247 [2024-11-26 23:46:03.365200] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:12:15.247 23:46:03 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:15.507 23:46:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=63488 00:12:15.507 23:46:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:15.507 23:46:03 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:15.507 23:46:03 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:15.507 23:46:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:12:15.507 23:46:03 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:15.507 23:46:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@619 -- # data_offset=2048 00:12:15.507 23:46:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@621 -- # '[' true = true ']' 00:12:15.507 23:46:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:12:15.507 23:46:03 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:15.507 23:46:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@623 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:12:15.507 23:46:03 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:15.507 [2024-11-26 23:46:03.464683] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:12:15.507 23:46:03 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:15.507 23:46:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:12:15.507 23:46:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:15.507 23:46:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:15.507 23:46:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:15.507 23:46:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:15.507 23:46:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:15.507 23:46:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:15.507 23:46:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:15.507 23:46:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:15.507 23:46:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:15.507 23:46:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:15.507 23:46:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:15.507 23:46:03 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:15.507 23:46:03 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:15.507 23:46:03 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:15.507 23:46:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:15.507 "name": "raid_bdev1", 00:12:15.507 "uuid": "1ac4fc07-7ba3-43ad-97f9-b1256b4b254d", 00:12:15.507 "strip_size_kb": 0, 00:12:15.507 "state": "online", 00:12:15.507 "raid_level": "raid1", 00:12:15.507 "superblock": true, 00:12:15.507 "num_base_bdevs": 4, 00:12:15.507 "num_base_bdevs_discovered": 3, 00:12:15.507 "num_base_bdevs_operational": 3, 00:12:15.507 "base_bdevs_list": [ 00:12:15.507 { 00:12:15.507 "name": null, 00:12:15.507 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:15.507 "is_configured": false, 00:12:15.507 "data_offset": 0, 00:12:15.507 "data_size": 63488 00:12:15.507 }, 00:12:15.507 { 00:12:15.507 "name": "BaseBdev2", 00:12:15.507 "uuid": "5197f646-d67b-5e09-af5d-c993bb98c4a5", 00:12:15.507 "is_configured": true, 00:12:15.507 "data_offset": 2048, 00:12:15.507 "data_size": 63488 00:12:15.507 }, 00:12:15.507 { 00:12:15.507 "name": "BaseBdev3", 00:12:15.507 "uuid": "ee6dcd27-62fb-55ac-bd58-5d03fde51b7b", 00:12:15.507 "is_configured": true, 00:12:15.507 "data_offset": 2048, 00:12:15.507 "data_size": 63488 00:12:15.507 }, 00:12:15.507 { 00:12:15.507 "name": "BaseBdev4", 00:12:15.507 "uuid": "5a25f2b0-5c65-5f8f-baeb-76b7b36d0157", 00:12:15.507 "is_configured": true, 00:12:15.507 "data_offset": 2048, 00:12:15.507 "data_size": 63488 00:12:15.507 } 00:12:15.507 ] 00:12:15.507 }' 00:12:15.507 23:46:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:15.507 23:46:03 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:15.507 [2024-11-26 23:46:03.558602] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002870 00:12:15.507 I/O size of 3145728 is greater than zero copy threshold (65536). 00:12:15.507 Zero copy mechanism will not be used. 00:12:15.507 Running I/O for 60 seconds... 00:12:15.768 23:46:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:12:15.768 23:46:03 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:15.768 23:46:03 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:15.768 [2024-11-26 23:46:03.881580] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:16.027 23:46:03 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:16.027 23:46:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@647 -- # sleep 1 00:12:16.027 [2024-11-26 23:46:03.922364] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002940 00:12:16.027 [2024-11-26 23:46:03.924267] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:12:16.027 [2024-11-26 23:46:04.040121] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:12:16.027 [2024-11-26 23:46:04.041442] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:12:16.286 [2024-11-26 23:46:04.252606] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:12:16.286 [2024-11-26 23:46:04.252933] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:12:16.559 [2024-11-26 23:46:04.476080] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:12:16.559 [2024-11-26 23:46:04.477285] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:12:16.819 127.00 IOPS, 381.00 MiB/s [2024-11-26T23:46:04.951Z] [2024-11-26 23:46:04.693812] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:12:16.819 [2024-11-26 23:46:04.694033] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:12:16.819 23:46:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:16.819 23:46:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:16.819 23:46:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:16.819 23:46:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:16.819 23:46:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:16.819 23:46:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:16.819 23:46:04 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:16.819 23:46:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:16.819 23:46:04 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:16.819 23:46:04 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:16.819 [2024-11-26 23:46:04.946666] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 14336 offset_begin: 12288 offset_end: 18432 00:12:17.079 23:46:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:17.079 "name": "raid_bdev1", 00:12:17.079 "uuid": "1ac4fc07-7ba3-43ad-97f9-b1256b4b254d", 00:12:17.079 "strip_size_kb": 0, 00:12:17.079 "state": "online", 00:12:17.079 "raid_level": "raid1", 00:12:17.079 "superblock": true, 00:12:17.079 "num_base_bdevs": 4, 00:12:17.079 "num_base_bdevs_discovered": 4, 00:12:17.079 "num_base_bdevs_operational": 4, 00:12:17.079 "process": { 00:12:17.079 "type": "rebuild", 00:12:17.079 "target": "spare", 00:12:17.079 "progress": { 00:12:17.079 "blocks": 12288, 00:12:17.079 "percent": 19 00:12:17.079 } 00:12:17.079 }, 00:12:17.079 "base_bdevs_list": [ 00:12:17.079 { 00:12:17.079 "name": "spare", 00:12:17.079 "uuid": "9feca6b8-4d7b-539a-9b6c-0d9b36941f36", 00:12:17.079 "is_configured": true, 00:12:17.079 "data_offset": 2048, 00:12:17.079 "data_size": 63488 00:12:17.079 }, 00:12:17.079 { 00:12:17.079 "name": "BaseBdev2", 00:12:17.079 "uuid": "5197f646-d67b-5e09-af5d-c993bb98c4a5", 00:12:17.079 "is_configured": true, 00:12:17.079 "data_offset": 2048, 00:12:17.079 "data_size": 63488 00:12:17.079 }, 00:12:17.079 { 00:12:17.079 "name": "BaseBdev3", 00:12:17.079 "uuid": "ee6dcd27-62fb-55ac-bd58-5d03fde51b7b", 00:12:17.079 "is_configured": true, 00:12:17.079 "data_offset": 2048, 00:12:17.079 "data_size": 63488 00:12:17.079 }, 00:12:17.079 { 00:12:17.079 "name": "BaseBdev4", 00:12:17.079 "uuid": "5a25f2b0-5c65-5f8f-baeb-76b7b36d0157", 00:12:17.079 "is_configured": true, 00:12:17.079 "data_offset": 2048, 00:12:17.079 "data_size": 63488 00:12:17.079 } 00:12:17.079 ] 00:12:17.079 }' 00:12:17.079 23:46:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:17.079 23:46:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:17.079 23:46:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:17.079 [2024-11-26 23:46:05.060959] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:12:17.079 23:46:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:17.079 23:46:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:12:17.079 23:46:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:17.079 23:46:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:17.079 [2024-11-26 23:46:05.071480] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:17.079 [2024-11-26 23:46:05.164864] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:12:17.339 [2024-11-26 23:46:05.267397] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:12:17.339 [2024-11-26 23:46:05.275837] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:17.339 [2024-11-26 23:46:05.275882] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:17.339 [2024-11-26 23:46:05.275915] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:12:17.339 [2024-11-26 23:46:05.299716] bdev_raid.c:1974:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 0 raid_ch: 0x60d000002870 00:12:17.339 23:46:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:17.339 23:46:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:12:17.339 23:46:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:17.339 23:46:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:17.339 23:46:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:17.339 23:46:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:17.339 23:46:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:17.339 23:46:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:17.339 23:46:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:17.339 23:46:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:17.339 23:46:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:17.339 23:46:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:17.339 23:46:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:17.339 23:46:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:17.339 23:46:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:17.339 23:46:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:17.339 23:46:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:17.339 "name": "raid_bdev1", 00:12:17.339 "uuid": "1ac4fc07-7ba3-43ad-97f9-b1256b4b254d", 00:12:17.339 "strip_size_kb": 0, 00:12:17.339 "state": "online", 00:12:17.339 "raid_level": "raid1", 00:12:17.339 "superblock": true, 00:12:17.339 "num_base_bdevs": 4, 00:12:17.339 "num_base_bdevs_discovered": 3, 00:12:17.339 "num_base_bdevs_operational": 3, 00:12:17.339 "base_bdevs_list": [ 00:12:17.339 { 00:12:17.339 "name": null, 00:12:17.339 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:17.339 "is_configured": false, 00:12:17.339 "data_offset": 0, 00:12:17.339 "data_size": 63488 00:12:17.339 }, 00:12:17.339 { 00:12:17.339 "name": "BaseBdev2", 00:12:17.339 "uuid": "5197f646-d67b-5e09-af5d-c993bb98c4a5", 00:12:17.339 "is_configured": true, 00:12:17.339 "data_offset": 2048, 00:12:17.339 "data_size": 63488 00:12:17.339 }, 00:12:17.339 { 00:12:17.339 "name": "BaseBdev3", 00:12:17.339 "uuid": "ee6dcd27-62fb-55ac-bd58-5d03fde51b7b", 00:12:17.339 "is_configured": true, 00:12:17.339 "data_offset": 2048, 00:12:17.339 "data_size": 63488 00:12:17.339 }, 00:12:17.339 { 00:12:17.339 "name": "BaseBdev4", 00:12:17.339 "uuid": "5a25f2b0-5c65-5f8f-baeb-76b7b36d0157", 00:12:17.339 "is_configured": true, 00:12:17.339 "data_offset": 2048, 00:12:17.339 "data_size": 63488 00:12:17.339 } 00:12:17.339 ] 00:12:17.339 }' 00:12:17.339 23:46:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:17.339 23:46:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:17.859 129.50 IOPS, 388.50 MiB/s [2024-11-26T23:46:05.991Z] 23:46:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:17.859 23:46:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:17.859 23:46:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:17.859 23:46:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:17.859 23:46:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:17.859 23:46:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:17.859 23:46:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:17.859 23:46:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:17.859 23:46:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:17.859 23:46:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:17.859 23:46:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:17.859 "name": "raid_bdev1", 00:12:17.859 "uuid": "1ac4fc07-7ba3-43ad-97f9-b1256b4b254d", 00:12:17.859 "strip_size_kb": 0, 00:12:17.859 "state": "online", 00:12:17.859 "raid_level": "raid1", 00:12:17.859 "superblock": true, 00:12:17.859 "num_base_bdevs": 4, 00:12:17.859 "num_base_bdevs_discovered": 3, 00:12:17.859 "num_base_bdevs_operational": 3, 00:12:17.859 "base_bdevs_list": [ 00:12:17.859 { 00:12:17.859 "name": null, 00:12:17.859 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:17.859 "is_configured": false, 00:12:17.859 "data_offset": 0, 00:12:17.859 "data_size": 63488 00:12:17.859 }, 00:12:17.859 { 00:12:17.859 "name": "BaseBdev2", 00:12:17.859 "uuid": "5197f646-d67b-5e09-af5d-c993bb98c4a5", 00:12:17.859 "is_configured": true, 00:12:17.859 "data_offset": 2048, 00:12:17.859 "data_size": 63488 00:12:17.859 }, 00:12:17.859 { 00:12:17.859 "name": "BaseBdev3", 00:12:17.859 "uuid": "ee6dcd27-62fb-55ac-bd58-5d03fde51b7b", 00:12:17.859 "is_configured": true, 00:12:17.859 "data_offset": 2048, 00:12:17.859 "data_size": 63488 00:12:17.859 }, 00:12:17.859 { 00:12:17.859 "name": "BaseBdev4", 00:12:17.859 "uuid": "5a25f2b0-5c65-5f8f-baeb-76b7b36d0157", 00:12:17.859 "is_configured": true, 00:12:17.859 "data_offset": 2048, 00:12:17.859 "data_size": 63488 00:12:17.859 } 00:12:17.859 ] 00:12:17.859 }' 00:12:17.859 23:46:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:17.860 23:46:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:17.860 23:46:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:17.860 23:46:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:17.860 23:46:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:12:17.860 23:46:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:17.860 23:46:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:17.860 [2024-11-26 23:46:05.892668] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:17.860 23:46:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:17.860 23:46:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@663 -- # sleep 1 00:12:17.860 [2024-11-26 23:46:05.922883] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002a10 00:12:17.860 [2024-11-26 23:46:05.924825] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:12:18.119 [2024-11-26 23:46:06.055688] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:12:18.119 [2024-11-26 23:46:06.057051] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:12:18.380 [2024-11-26 23:46:06.280050] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:12:18.380 [2024-11-26 23:46:06.280356] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:12:18.641 [2024-11-26 23:46:06.529598] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:12:18.642 [2024-11-26 23:46:06.529947] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:12:18.642 144.67 IOPS, 434.00 MiB/s [2024-11-26T23:46:06.774Z] [2024-11-26 23:46:06.740459] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:12:18.642 [2024-11-26 23:46:06.741167] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:12:18.901 23:46:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:18.901 23:46:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:18.901 23:46:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:18.901 23:46:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:18.901 23:46:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:18.901 23:46:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:18.901 23:46:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:18.901 23:46:06 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:18.901 23:46:06 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:18.901 23:46:06 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:18.901 23:46:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:18.901 "name": "raid_bdev1", 00:12:18.901 "uuid": "1ac4fc07-7ba3-43ad-97f9-b1256b4b254d", 00:12:18.901 "strip_size_kb": 0, 00:12:18.901 "state": "online", 00:12:18.901 "raid_level": "raid1", 00:12:18.901 "superblock": true, 00:12:18.901 "num_base_bdevs": 4, 00:12:18.901 "num_base_bdevs_discovered": 4, 00:12:18.901 "num_base_bdevs_operational": 4, 00:12:18.901 "process": { 00:12:18.901 "type": "rebuild", 00:12:18.901 "target": "spare", 00:12:18.901 "progress": { 00:12:18.901 "blocks": 10240, 00:12:18.901 "percent": 16 00:12:18.901 } 00:12:18.901 }, 00:12:18.901 "base_bdevs_list": [ 00:12:18.901 { 00:12:18.901 "name": "spare", 00:12:18.901 "uuid": "9feca6b8-4d7b-539a-9b6c-0d9b36941f36", 00:12:18.901 "is_configured": true, 00:12:18.901 "data_offset": 2048, 00:12:18.901 "data_size": 63488 00:12:18.901 }, 00:12:18.901 { 00:12:18.901 "name": "BaseBdev2", 00:12:18.901 "uuid": "5197f646-d67b-5e09-af5d-c993bb98c4a5", 00:12:18.901 "is_configured": true, 00:12:18.901 "data_offset": 2048, 00:12:18.901 "data_size": 63488 00:12:18.901 }, 00:12:18.901 { 00:12:18.901 "name": "BaseBdev3", 00:12:18.901 "uuid": "ee6dcd27-62fb-55ac-bd58-5d03fde51b7b", 00:12:18.901 "is_configured": true, 00:12:18.901 "data_offset": 2048, 00:12:18.901 "data_size": 63488 00:12:18.901 }, 00:12:18.901 { 00:12:18.901 "name": "BaseBdev4", 00:12:18.901 "uuid": "5a25f2b0-5c65-5f8f-baeb-76b7b36d0157", 00:12:18.901 "is_configured": true, 00:12:18.901 "data_offset": 2048, 00:12:18.901 "data_size": 63488 00:12:18.901 } 00:12:18.901 ] 00:12:18.901 }' 00:12:18.901 23:46:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:18.901 23:46:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:18.901 23:46:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:19.161 23:46:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:19.161 23:46:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@666 -- # '[' true = true ']' 00:12:19.161 23:46:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@666 -- # '[' = false ']' 00:12:19.161 /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh: line 666: [: =: unary operator expected 00:12:19.161 23:46:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=4 00:12:19.161 23:46:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:12:19.161 23:46:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@693 -- # '[' 4 -gt 2 ']' 00:12:19.161 23:46:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@695 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:12:19.161 23:46:07 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:19.161 23:46:07 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:19.161 [2024-11-26 23:46:07.073873] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:12:19.421 [2024-11-26 23:46:07.304421] bdev_raid.c:1974:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 1 raid_ch: 0x60d000002870 00:12:19.421 [2024-11-26 23:46:07.304460] bdev_raid.c:1974:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 1 raid_ch: 0x60d000002a10 00:12:19.421 [2024-11-26 23:46:07.304515] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 14336 offset_begin: 12288 offset_end: 18432 00:12:19.421 23:46:07 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:19.421 23:46:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@698 -- # base_bdevs[1]= 00:12:19.421 23:46:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@699 -- # (( num_base_bdevs_operational-- )) 00:12:19.421 23:46:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@702 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:19.421 23:46:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:19.421 23:46:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:19.421 23:46:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:19.421 23:46:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:19.421 23:46:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:19.421 23:46:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:19.421 23:46:07 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:19.421 23:46:07 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:19.421 23:46:07 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:19.421 23:46:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:19.421 "name": "raid_bdev1", 00:12:19.421 "uuid": "1ac4fc07-7ba3-43ad-97f9-b1256b4b254d", 00:12:19.421 "strip_size_kb": 0, 00:12:19.421 "state": "online", 00:12:19.421 "raid_level": "raid1", 00:12:19.421 "superblock": true, 00:12:19.421 "num_base_bdevs": 4, 00:12:19.421 "num_base_bdevs_discovered": 3, 00:12:19.421 "num_base_bdevs_operational": 3, 00:12:19.421 "process": { 00:12:19.421 "type": "rebuild", 00:12:19.421 "target": "spare", 00:12:19.421 "progress": { 00:12:19.421 "blocks": 14336, 00:12:19.421 "percent": 22 00:12:19.421 } 00:12:19.421 }, 00:12:19.421 "base_bdevs_list": [ 00:12:19.421 { 00:12:19.421 "name": "spare", 00:12:19.421 "uuid": "9feca6b8-4d7b-539a-9b6c-0d9b36941f36", 00:12:19.421 "is_configured": true, 00:12:19.421 "data_offset": 2048, 00:12:19.421 "data_size": 63488 00:12:19.421 }, 00:12:19.421 { 00:12:19.421 "name": null, 00:12:19.421 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:19.421 "is_configured": false, 00:12:19.421 "data_offset": 0, 00:12:19.421 "data_size": 63488 00:12:19.421 }, 00:12:19.421 { 00:12:19.421 "name": "BaseBdev3", 00:12:19.421 "uuid": "ee6dcd27-62fb-55ac-bd58-5d03fde51b7b", 00:12:19.421 "is_configured": true, 00:12:19.421 "data_offset": 2048, 00:12:19.421 "data_size": 63488 00:12:19.421 }, 00:12:19.421 { 00:12:19.421 "name": "BaseBdev4", 00:12:19.421 "uuid": "5a25f2b0-5c65-5f8f-baeb-76b7b36d0157", 00:12:19.421 "is_configured": true, 00:12:19.421 "data_offset": 2048, 00:12:19.421 "data_size": 63488 00:12:19.421 } 00:12:19.421 ] 00:12:19.421 }' 00:12:19.421 23:46:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:19.421 23:46:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:19.421 23:46:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:19.421 23:46:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:19.421 23:46:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@706 -- # local timeout=399 00:12:19.421 23:46:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:12:19.421 23:46:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:19.421 23:46:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:19.421 23:46:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:19.421 23:46:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:19.421 23:46:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:19.421 23:46:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:19.421 23:46:07 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:19.421 23:46:07 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:19.421 23:46:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:19.421 23:46:07 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:19.421 23:46:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:19.421 "name": "raid_bdev1", 00:12:19.421 "uuid": "1ac4fc07-7ba3-43ad-97f9-b1256b4b254d", 00:12:19.421 "strip_size_kb": 0, 00:12:19.422 "state": "online", 00:12:19.422 "raid_level": "raid1", 00:12:19.422 "superblock": true, 00:12:19.422 "num_base_bdevs": 4, 00:12:19.422 "num_base_bdevs_discovered": 3, 00:12:19.422 "num_base_bdevs_operational": 3, 00:12:19.422 "process": { 00:12:19.422 "type": "rebuild", 00:12:19.422 "target": "spare", 00:12:19.422 "progress": { 00:12:19.422 "blocks": 14336, 00:12:19.422 "percent": 22 00:12:19.422 } 00:12:19.422 }, 00:12:19.422 "base_bdevs_list": [ 00:12:19.422 { 00:12:19.422 "name": "spare", 00:12:19.422 "uuid": "9feca6b8-4d7b-539a-9b6c-0d9b36941f36", 00:12:19.422 "is_configured": true, 00:12:19.422 "data_offset": 2048, 00:12:19.422 "data_size": 63488 00:12:19.422 }, 00:12:19.422 { 00:12:19.422 "name": null, 00:12:19.422 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:19.422 "is_configured": false, 00:12:19.422 "data_offset": 0, 00:12:19.422 "data_size": 63488 00:12:19.422 }, 00:12:19.422 { 00:12:19.422 "name": "BaseBdev3", 00:12:19.422 "uuid": "ee6dcd27-62fb-55ac-bd58-5d03fde51b7b", 00:12:19.422 "is_configured": true, 00:12:19.422 "data_offset": 2048, 00:12:19.422 "data_size": 63488 00:12:19.422 }, 00:12:19.422 { 00:12:19.422 "name": "BaseBdev4", 00:12:19.422 "uuid": "5a25f2b0-5c65-5f8f-baeb-76b7b36d0157", 00:12:19.422 "is_configured": true, 00:12:19.422 "data_offset": 2048, 00:12:19.422 "data_size": 63488 00:12:19.422 } 00:12:19.422 ] 00:12:19.422 }' 00:12:19.422 23:46:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:19.682 23:46:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:19.682 23:46:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:19.682 123.25 IOPS, 369.75 MiB/s [2024-11-26T23:46:07.814Z] 23:46:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:19.682 23:46:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:12:19.682 [2024-11-26 23:46:07.742586] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 20480 offset_begin: 18432 offset_end: 24576 00:12:19.941 [2024-11-26 23:46:07.944445] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 22528 offset_begin: 18432 offset_end: 24576 00:12:20.200 [2024-11-26 23:46:08.153999] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 26624 offset_begin: 24576 offset_end: 30720 00:12:20.200 [2024-11-26 23:46:08.154871] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 26624 offset_begin: 24576 offset_end: 30720 00:12:20.460 [2024-11-26 23:46:08.371405] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 28672 offset_begin: 24576 offset_end: 30720 00:12:20.720 110.00 IOPS, 330.00 MiB/s [2024-11-26T23:46:08.852Z] 23:46:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:12:20.720 23:46:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:20.720 23:46:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:20.720 23:46:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:20.720 23:46:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:20.720 23:46:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:20.720 23:46:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:20.720 23:46:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:20.720 23:46:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:20.720 23:46:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:20.720 23:46:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:20.720 23:46:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:20.720 "name": "raid_bdev1", 00:12:20.720 "uuid": "1ac4fc07-7ba3-43ad-97f9-b1256b4b254d", 00:12:20.720 "strip_size_kb": 0, 00:12:20.720 "state": "online", 00:12:20.720 "raid_level": "raid1", 00:12:20.720 "superblock": true, 00:12:20.720 "num_base_bdevs": 4, 00:12:20.720 "num_base_bdevs_discovered": 3, 00:12:20.720 "num_base_bdevs_operational": 3, 00:12:20.720 "process": { 00:12:20.720 "type": "rebuild", 00:12:20.720 "target": "spare", 00:12:20.720 "progress": { 00:12:20.720 "blocks": 30720, 00:12:20.720 "percent": 48 00:12:20.720 } 00:12:20.720 }, 00:12:20.720 "base_bdevs_list": [ 00:12:20.720 { 00:12:20.720 "name": "spare", 00:12:20.720 "uuid": "9feca6b8-4d7b-539a-9b6c-0d9b36941f36", 00:12:20.720 "is_configured": true, 00:12:20.720 "data_offset": 2048, 00:12:20.720 "data_size": 63488 00:12:20.720 }, 00:12:20.720 { 00:12:20.720 "name": null, 00:12:20.721 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:20.721 "is_configured": false, 00:12:20.721 "data_offset": 0, 00:12:20.721 "data_size": 63488 00:12:20.721 }, 00:12:20.721 { 00:12:20.721 "name": "BaseBdev3", 00:12:20.721 "uuid": "ee6dcd27-62fb-55ac-bd58-5d03fde51b7b", 00:12:20.721 "is_configured": true, 00:12:20.721 "data_offset": 2048, 00:12:20.721 "data_size": 63488 00:12:20.721 }, 00:12:20.721 { 00:12:20.721 "name": "BaseBdev4", 00:12:20.721 "uuid": "5a25f2b0-5c65-5f8f-baeb-76b7b36d0157", 00:12:20.721 "is_configured": true, 00:12:20.721 "data_offset": 2048, 00:12:20.721 "data_size": 63488 00:12:20.721 } 00:12:20.721 ] 00:12:20.721 }' 00:12:20.721 23:46:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:20.721 [2024-11-26 23:46:08.700139] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 32768 offset_begin: 30720 offset_end: 36864 00:12:20.721 23:46:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:20.721 23:46:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:20.721 23:46:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:20.721 23:46:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:12:20.721 [2024-11-26 23:46:08.809976] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 34816 offset_begin: 30720 offset_end: 36864 00:12:21.290 [2024-11-26 23:46:09.154572] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 40960 offset_begin: 36864 offset_end: 43008 00:12:21.290 [2024-11-26 23:46:09.383373] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 45056 offset_begin: 43008 offset_end: 49152 00:12:21.550 [2024-11-26 23:46:09.489476] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 47104 offset_begin: 43008 offset_end: 49152 00:12:21.809 100.83 IOPS, 302.50 MiB/s [2024-11-26T23:46:09.941Z] 23:46:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:12:21.809 23:46:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:21.809 23:46:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:21.809 23:46:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:21.809 23:46:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:21.809 23:46:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:21.809 23:46:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:21.809 23:46:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:21.809 23:46:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:21.809 23:46:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:21.809 [2024-11-26 23:46:09.792495] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 53248 offset_begin: 49152 offset_end: 55296 00:12:21.809 23:46:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:21.809 23:46:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:21.809 "name": "raid_bdev1", 00:12:21.809 "uuid": "1ac4fc07-7ba3-43ad-97f9-b1256b4b254d", 00:12:21.809 "strip_size_kb": 0, 00:12:21.809 "state": "online", 00:12:21.809 "raid_level": "raid1", 00:12:21.809 "superblock": true, 00:12:21.809 "num_base_bdevs": 4, 00:12:21.809 "num_base_bdevs_discovered": 3, 00:12:21.809 "num_base_bdevs_operational": 3, 00:12:21.809 "process": { 00:12:21.809 "type": "rebuild", 00:12:21.809 "target": "spare", 00:12:21.809 "progress": { 00:12:21.809 "blocks": 51200, 00:12:21.809 "percent": 80 00:12:21.809 } 00:12:21.809 }, 00:12:21.809 "base_bdevs_list": [ 00:12:21.809 { 00:12:21.809 "name": "spare", 00:12:21.809 "uuid": "9feca6b8-4d7b-539a-9b6c-0d9b36941f36", 00:12:21.809 "is_configured": true, 00:12:21.809 "data_offset": 2048, 00:12:21.809 "data_size": 63488 00:12:21.809 }, 00:12:21.809 { 00:12:21.809 "name": null, 00:12:21.809 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:21.809 "is_configured": false, 00:12:21.809 "data_offset": 0, 00:12:21.809 "data_size": 63488 00:12:21.809 }, 00:12:21.809 { 00:12:21.809 "name": "BaseBdev3", 00:12:21.809 "uuid": "ee6dcd27-62fb-55ac-bd58-5d03fde51b7b", 00:12:21.809 "is_configured": true, 00:12:21.810 "data_offset": 2048, 00:12:21.810 "data_size": 63488 00:12:21.810 }, 00:12:21.810 { 00:12:21.810 "name": "BaseBdev4", 00:12:21.810 "uuid": "5a25f2b0-5c65-5f8f-baeb-76b7b36d0157", 00:12:21.810 "is_configured": true, 00:12:21.810 "data_offset": 2048, 00:12:21.810 "data_size": 63488 00:12:21.810 } 00:12:21.810 ] 00:12:21.810 }' 00:12:21.810 23:46:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:21.810 23:46:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:21.810 23:46:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:21.810 23:46:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:21.810 23:46:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:12:22.379 [2024-11-26 23:46:10.329865] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:12:22.379 [2024-11-26 23:46:10.434621] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:12:22.379 [2024-11-26 23:46:10.436818] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:22.898 91.86 IOPS, 275.57 MiB/s [2024-11-26T23:46:11.030Z] 23:46:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:12:22.898 23:46:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:22.898 23:46:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:22.898 23:46:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:22.898 23:46:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:22.898 23:46:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:22.898 23:46:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:22.898 23:46:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:22.898 23:46:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:22.898 23:46:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:22.898 23:46:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:22.898 23:46:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:22.898 "name": "raid_bdev1", 00:12:22.898 "uuid": "1ac4fc07-7ba3-43ad-97f9-b1256b4b254d", 00:12:22.898 "strip_size_kb": 0, 00:12:22.898 "state": "online", 00:12:22.898 "raid_level": "raid1", 00:12:22.898 "superblock": true, 00:12:22.898 "num_base_bdevs": 4, 00:12:22.898 "num_base_bdevs_discovered": 3, 00:12:22.898 "num_base_bdevs_operational": 3, 00:12:22.898 "base_bdevs_list": [ 00:12:22.898 { 00:12:22.898 "name": "spare", 00:12:22.898 "uuid": "9feca6b8-4d7b-539a-9b6c-0d9b36941f36", 00:12:22.898 "is_configured": true, 00:12:22.898 "data_offset": 2048, 00:12:22.898 "data_size": 63488 00:12:22.898 }, 00:12:22.898 { 00:12:22.898 "name": null, 00:12:22.898 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:22.898 "is_configured": false, 00:12:22.898 "data_offset": 0, 00:12:22.898 "data_size": 63488 00:12:22.898 }, 00:12:22.898 { 00:12:22.898 "name": "BaseBdev3", 00:12:22.898 "uuid": "ee6dcd27-62fb-55ac-bd58-5d03fde51b7b", 00:12:22.898 "is_configured": true, 00:12:22.898 "data_offset": 2048, 00:12:22.898 "data_size": 63488 00:12:22.898 }, 00:12:22.898 { 00:12:22.898 "name": "BaseBdev4", 00:12:22.898 "uuid": "5a25f2b0-5c65-5f8f-baeb-76b7b36d0157", 00:12:22.898 "is_configured": true, 00:12:22.898 "data_offset": 2048, 00:12:22.898 "data_size": 63488 00:12:22.898 } 00:12:22.898 ] 00:12:22.898 }' 00:12:22.898 23:46:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:22.898 23:46:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:12:22.898 23:46:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:23.159 23:46:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:12:23.159 23:46:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@709 -- # break 00:12:23.159 23:46:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:23.159 23:46:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:23.159 23:46:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:23.159 23:46:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:23.159 23:46:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:23.159 23:46:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:23.159 23:46:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:23.159 23:46:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:23.159 23:46:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:23.159 23:46:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:23.159 23:46:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:23.159 "name": "raid_bdev1", 00:12:23.159 "uuid": "1ac4fc07-7ba3-43ad-97f9-b1256b4b254d", 00:12:23.159 "strip_size_kb": 0, 00:12:23.159 "state": "online", 00:12:23.159 "raid_level": "raid1", 00:12:23.159 "superblock": true, 00:12:23.159 "num_base_bdevs": 4, 00:12:23.159 "num_base_bdevs_discovered": 3, 00:12:23.159 "num_base_bdevs_operational": 3, 00:12:23.159 "base_bdevs_list": [ 00:12:23.159 { 00:12:23.159 "name": "spare", 00:12:23.159 "uuid": "9feca6b8-4d7b-539a-9b6c-0d9b36941f36", 00:12:23.159 "is_configured": true, 00:12:23.159 "data_offset": 2048, 00:12:23.159 "data_size": 63488 00:12:23.159 }, 00:12:23.159 { 00:12:23.159 "name": null, 00:12:23.159 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:23.159 "is_configured": false, 00:12:23.159 "data_offset": 0, 00:12:23.159 "data_size": 63488 00:12:23.159 }, 00:12:23.159 { 00:12:23.159 "name": "BaseBdev3", 00:12:23.159 "uuid": "ee6dcd27-62fb-55ac-bd58-5d03fde51b7b", 00:12:23.159 "is_configured": true, 00:12:23.159 "data_offset": 2048, 00:12:23.159 "data_size": 63488 00:12:23.159 }, 00:12:23.159 { 00:12:23.159 "name": "BaseBdev4", 00:12:23.159 "uuid": "5a25f2b0-5c65-5f8f-baeb-76b7b36d0157", 00:12:23.159 "is_configured": true, 00:12:23.159 "data_offset": 2048, 00:12:23.159 "data_size": 63488 00:12:23.159 } 00:12:23.159 ] 00:12:23.159 }' 00:12:23.159 23:46:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:23.159 23:46:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:23.159 23:46:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:23.159 23:46:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:23.159 23:46:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:12:23.159 23:46:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:23.159 23:46:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:23.159 23:46:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:23.159 23:46:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:23.159 23:46:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:23.159 23:46:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:23.159 23:46:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:23.159 23:46:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:23.159 23:46:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:23.159 23:46:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:23.159 23:46:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:23.159 23:46:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:23.159 23:46:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:23.159 23:46:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:23.159 23:46:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:23.159 "name": "raid_bdev1", 00:12:23.159 "uuid": "1ac4fc07-7ba3-43ad-97f9-b1256b4b254d", 00:12:23.159 "strip_size_kb": 0, 00:12:23.159 "state": "online", 00:12:23.159 "raid_level": "raid1", 00:12:23.159 "superblock": true, 00:12:23.159 "num_base_bdevs": 4, 00:12:23.159 "num_base_bdevs_discovered": 3, 00:12:23.159 "num_base_bdevs_operational": 3, 00:12:23.159 "base_bdevs_list": [ 00:12:23.159 { 00:12:23.159 "name": "spare", 00:12:23.159 "uuid": "9feca6b8-4d7b-539a-9b6c-0d9b36941f36", 00:12:23.159 "is_configured": true, 00:12:23.159 "data_offset": 2048, 00:12:23.159 "data_size": 63488 00:12:23.159 }, 00:12:23.159 { 00:12:23.159 "name": null, 00:12:23.159 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:23.159 "is_configured": false, 00:12:23.159 "data_offset": 0, 00:12:23.159 "data_size": 63488 00:12:23.159 }, 00:12:23.159 { 00:12:23.159 "name": "BaseBdev3", 00:12:23.159 "uuid": "ee6dcd27-62fb-55ac-bd58-5d03fde51b7b", 00:12:23.159 "is_configured": true, 00:12:23.159 "data_offset": 2048, 00:12:23.159 "data_size": 63488 00:12:23.159 }, 00:12:23.159 { 00:12:23.159 "name": "BaseBdev4", 00:12:23.159 "uuid": "5a25f2b0-5c65-5f8f-baeb-76b7b36d0157", 00:12:23.159 "is_configured": true, 00:12:23.159 "data_offset": 2048, 00:12:23.159 "data_size": 63488 00:12:23.159 } 00:12:23.159 ] 00:12:23.159 }' 00:12:23.159 23:46:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:23.159 23:46:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:23.729 85.00 IOPS, 255.00 MiB/s [2024-11-26T23:46:11.861Z] 23:46:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:12:23.729 23:46:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:23.729 23:46:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:23.729 [2024-11-26 23:46:11.638653] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:12:23.729 [2024-11-26 23:46:11.638693] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:12:23.729 00:12:23.729 Latency(us) 00:12:23.729 [2024-11-26T23:46:11.861Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:12:23.729 Job: raid_bdev1 (Core Mask 0x1, workload: randrw, percentage: 50, depth: 2, IO size: 3145728) 00:12:23.729 raid_bdev1 : 8.18 83.64 250.93 0.00 0.00 16009.34 302.28 109436.53 00:12:23.729 [2024-11-26T23:46:11.861Z] =================================================================================================================== 00:12:23.729 [2024-11-26T23:46:11.861Z] Total : 83.64 250.93 0.00 0.00 16009.34 302.28 109436.53 00:12:23.729 [2024-11-26 23:46:11.725918] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:12:23.729 [2024-11-26 23:46:11.725977] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:23.729 [2024-11-26 23:46:11.726066] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:12:23.729 [2024-11-26 23:46:11.726098] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:12:23.729 { 00:12:23.729 "results": [ 00:12:23.729 { 00:12:23.729 "job": "raid_bdev1", 00:12:23.729 "core_mask": "0x1", 00:12:23.729 "workload": "randrw", 00:12:23.729 "percentage": 50, 00:12:23.729 "status": "finished", 00:12:23.729 "queue_depth": 2, 00:12:23.729 "io_size": 3145728, 00:12:23.729 "runtime": 8.177485, 00:12:23.729 "iops": 83.64429895010508, 00:12:23.729 "mibps": 250.93289685031525, 00:12:23.729 "io_failed": 0, 00:12:23.729 "io_timeout": 0, 00:12:23.729 "avg_latency_us": 16009.336132179063, 00:12:23.729 "min_latency_us": 302.2812227074236, 00:12:23.729 "max_latency_us": 109436.5344978166 00:12:23.729 } 00:12:23.729 ], 00:12:23.729 "core_count": 1 00:12:23.729 } 00:12:23.729 23:46:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:23.729 23:46:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:23.729 23:46:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@720 -- # jq length 00:12:23.729 23:46:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:23.729 23:46:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:23.729 23:46:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:23.729 23:46:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:12:23.729 23:46:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:12:23.729 23:46:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@723 -- # '[' true = true ']' 00:12:23.729 23:46:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@725 -- # nbd_start_disks /var/tmp/spdk.sock spare /dev/nbd0 00:12:23.729 23:46:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:12:23.729 23:46:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # bdev_list=('spare') 00:12:23.729 23:46:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:12:23.729 23:46:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:12:23.729 23:46:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:12:23.729 23:46:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@12 -- # local i 00:12:23.729 23:46:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:12:23.729 23:46:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:12:23.729 23:46:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd0 00:12:23.989 /dev/nbd0 00:12:23.989 23:46:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:12:23.989 23:46:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:12:23.989 23:46:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:12:23.989 23:46:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@873 -- # local i 00:12:23.989 23:46:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:12:23.989 23:46:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:12:23.989 23:46:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:12:23.989 23:46:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@877 -- # break 00:12:23.989 23:46:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:12:23.989 23:46:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:12:23.989 23:46:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:12:23.989 1+0 records in 00:12:23.989 1+0 records out 00:12:23.989 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000387647 s, 10.6 MB/s 00:12:23.989 23:46:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:23.989 23:46:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@890 -- # size=4096 00:12:23.989 23:46:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:23.989 23:46:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:12:23.989 23:46:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@893 -- # return 0 00:12:23.989 23:46:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:12:23.989 23:46:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:12:23.989 23:46:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@726 -- # for bdev in "${base_bdevs[@]:1}" 00:12:23.989 23:46:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@727 -- # '[' -z '' ']' 00:12:23.989 23:46:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@728 -- # continue 00:12:23.989 23:46:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@726 -- # for bdev in "${base_bdevs[@]:1}" 00:12:23.989 23:46:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@727 -- # '[' -z BaseBdev3 ']' 00:12:23.989 23:46:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@730 -- # nbd_start_disks /var/tmp/spdk.sock BaseBdev3 /dev/nbd1 00:12:23.989 23:46:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:12:23.989 23:46:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev3') 00:12:23.989 23:46:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:12:23.989 23:46:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd1') 00:12:23.989 23:46:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:12:23.989 23:46:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@12 -- # local i 00:12:23.989 23:46:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:12:23.989 23:46:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:12:23.989 23:46:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev3 /dev/nbd1 00:12:24.250 /dev/nbd1 00:12:24.250 23:46:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:12:24.250 23:46:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:12:24.250 23:46:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:12:24.250 23:46:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@873 -- # local i 00:12:24.250 23:46:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:12:24.250 23:46:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:12:24.250 23:46:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:12:24.250 23:46:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@877 -- # break 00:12:24.250 23:46:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:12:24.250 23:46:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:12:24.250 23:46:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:12:24.250 1+0 records in 00:12:24.250 1+0 records out 00:12:24.250 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000448989 s, 9.1 MB/s 00:12:24.250 23:46:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:24.250 23:46:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@890 -- # size=4096 00:12:24.250 23:46:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:24.250 23:46:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:12:24.250 23:46:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@893 -- # return 0 00:12:24.250 23:46:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:12:24.250 23:46:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:12:24.250 23:46:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@731 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:12:24.250 23:46:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@732 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd1 00:12:24.250 23:46:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:12:24.250 23:46:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd1') 00:12:24.250 23:46:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:12:24.250 23:46:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@51 -- # local i 00:12:24.250 23:46:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:12:24.250 23:46:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:12:24.564 23:46:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:12:24.564 23:46:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:12:24.564 23:46:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:12:24.564 23:46:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:12:24.564 23:46:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:12:24.564 23:46:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:12:24.564 23:46:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@41 -- # break 00:12:24.564 23:46:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@45 -- # return 0 00:12:24.564 23:46:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@726 -- # for bdev in "${base_bdevs[@]:1}" 00:12:24.564 23:46:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@727 -- # '[' -z BaseBdev4 ']' 00:12:24.564 23:46:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@730 -- # nbd_start_disks /var/tmp/spdk.sock BaseBdev4 /dev/nbd1 00:12:24.564 23:46:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:12:24.564 23:46:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev4') 00:12:24.564 23:46:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:12:24.564 23:46:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd1') 00:12:24.564 23:46:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:12:24.564 23:46:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@12 -- # local i 00:12:24.564 23:46:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:12:24.564 23:46:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:12:24.564 23:46:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev4 /dev/nbd1 00:12:24.824 /dev/nbd1 00:12:24.824 23:46:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:12:24.824 23:46:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:12:24.824 23:46:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:12:24.824 23:46:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@873 -- # local i 00:12:24.824 23:46:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:12:24.824 23:46:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:12:24.824 23:46:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:12:24.824 23:46:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@877 -- # break 00:12:24.824 23:46:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:12:24.824 23:46:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:12:24.824 23:46:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:12:24.824 1+0 records in 00:12:24.824 1+0 records out 00:12:24.824 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000468388 s, 8.7 MB/s 00:12:24.824 23:46:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:24.824 23:46:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@890 -- # size=4096 00:12:24.824 23:46:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:24.824 23:46:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:12:24.824 23:46:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@893 -- # return 0 00:12:24.824 23:46:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:12:24.824 23:46:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:12:24.824 23:46:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@731 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:12:24.824 23:46:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@732 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd1 00:12:24.824 23:46:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:12:24.824 23:46:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd1') 00:12:24.824 23:46:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:12:24.824 23:46:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@51 -- # local i 00:12:24.824 23:46:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:12:24.824 23:46:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:12:25.083 23:46:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:12:25.083 23:46:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:12:25.083 23:46:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:12:25.083 23:46:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:12:25.083 23:46:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:12:25.083 23:46:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:12:25.083 23:46:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@41 -- # break 00:12:25.083 23:46:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@45 -- # return 0 00:12:25.083 23:46:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@734 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:12:25.083 23:46:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:12:25.083 23:46:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:12:25.083 23:46:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:12:25.083 23:46:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@51 -- # local i 00:12:25.083 23:46:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:12:25.083 23:46:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:12:25.342 23:46:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:12:25.342 23:46:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:12:25.342 23:46:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:12:25.342 23:46:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:12:25.342 23:46:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:12:25.342 23:46:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:12:25.342 23:46:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@41 -- # break 00:12:25.342 23:46:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@45 -- # return 0 00:12:25.342 23:46:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@743 -- # '[' true = true ']' 00:12:25.342 23:46:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@745 -- # rpc_cmd bdev_passthru_delete spare 00:12:25.342 23:46:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:25.342 23:46:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:25.342 23:46:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:25.342 23:46:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@746 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:12:25.342 23:46:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:25.342 23:46:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:25.342 [2024-11-26 23:46:13.278685] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:12:25.342 [2024-11-26 23:46:13.278747] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:25.342 [2024-11-26 23:46:13.278766] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b180 00:12:25.342 [2024-11-26 23:46:13.278777] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:25.342 [2024-11-26 23:46:13.280914] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:25.342 [2024-11-26 23:46:13.280953] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:12:25.342 [2024-11-26 23:46:13.281029] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:12:25.342 [2024-11-26 23:46:13.281065] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:25.342 [2024-11-26 23:46:13.281204] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:12:25.342 [2024-11-26 23:46:13.281324] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:12:25.342 spare 00:12:25.342 23:46:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:25.342 23:46:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@747 -- # rpc_cmd bdev_wait_for_examine 00:12:25.342 23:46:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:25.342 23:46:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:25.342 [2024-11-26 23:46:13.381235] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001580 00:12:25.342 [2024-11-26 23:46:13.381265] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:12:25.342 [2024-11-26 23:46:13.381570] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000337b0 00:12:25.342 [2024-11-26 23:46:13.381735] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001580 00:12:25.342 [2024-11-26 23:46:13.381752] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001580 00:12:25.342 [2024-11-26 23:46:13.381874] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:25.342 23:46:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:25.342 23:46:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@749 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:12:25.342 23:46:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:25.342 23:46:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:25.342 23:46:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:25.342 23:46:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:25.342 23:46:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:25.342 23:46:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:25.342 23:46:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:25.342 23:46:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:25.342 23:46:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:25.342 23:46:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:25.342 23:46:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:25.342 23:46:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:25.342 23:46:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:25.342 23:46:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:25.342 23:46:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:25.342 "name": "raid_bdev1", 00:12:25.342 "uuid": "1ac4fc07-7ba3-43ad-97f9-b1256b4b254d", 00:12:25.342 "strip_size_kb": 0, 00:12:25.342 "state": "online", 00:12:25.342 "raid_level": "raid1", 00:12:25.342 "superblock": true, 00:12:25.342 "num_base_bdevs": 4, 00:12:25.342 "num_base_bdevs_discovered": 3, 00:12:25.342 "num_base_bdevs_operational": 3, 00:12:25.342 "base_bdevs_list": [ 00:12:25.342 { 00:12:25.342 "name": "spare", 00:12:25.342 "uuid": "9feca6b8-4d7b-539a-9b6c-0d9b36941f36", 00:12:25.342 "is_configured": true, 00:12:25.342 "data_offset": 2048, 00:12:25.342 "data_size": 63488 00:12:25.342 }, 00:12:25.342 { 00:12:25.342 "name": null, 00:12:25.342 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:25.342 "is_configured": false, 00:12:25.342 "data_offset": 2048, 00:12:25.342 "data_size": 63488 00:12:25.342 }, 00:12:25.342 { 00:12:25.342 "name": "BaseBdev3", 00:12:25.342 "uuid": "ee6dcd27-62fb-55ac-bd58-5d03fde51b7b", 00:12:25.342 "is_configured": true, 00:12:25.342 "data_offset": 2048, 00:12:25.342 "data_size": 63488 00:12:25.342 }, 00:12:25.342 { 00:12:25.342 "name": "BaseBdev4", 00:12:25.342 "uuid": "5a25f2b0-5c65-5f8f-baeb-76b7b36d0157", 00:12:25.342 "is_configured": true, 00:12:25.342 "data_offset": 2048, 00:12:25.342 "data_size": 63488 00:12:25.342 } 00:12:25.342 ] 00:12:25.342 }' 00:12:25.342 23:46:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:25.342 23:46:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:25.912 23:46:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@750 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:25.912 23:46:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:25.912 23:46:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:25.912 23:46:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:25.912 23:46:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:25.912 23:46:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:25.912 23:46:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:25.912 23:46:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:25.912 23:46:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:25.912 23:46:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:25.912 23:46:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:25.912 "name": "raid_bdev1", 00:12:25.912 "uuid": "1ac4fc07-7ba3-43ad-97f9-b1256b4b254d", 00:12:25.912 "strip_size_kb": 0, 00:12:25.912 "state": "online", 00:12:25.912 "raid_level": "raid1", 00:12:25.912 "superblock": true, 00:12:25.912 "num_base_bdevs": 4, 00:12:25.912 "num_base_bdevs_discovered": 3, 00:12:25.912 "num_base_bdevs_operational": 3, 00:12:25.912 "base_bdevs_list": [ 00:12:25.912 { 00:12:25.912 "name": "spare", 00:12:25.912 "uuid": "9feca6b8-4d7b-539a-9b6c-0d9b36941f36", 00:12:25.912 "is_configured": true, 00:12:25.912 "data_offset": 2048, 00:12:25.912 "data_size": 63488 00:12:25.912 }, 00:12:25.912 { 00:12:25.912 "name": null, 00:12:25.912 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:25.912 "is_configured": false, 00:12:25.912 "data_offset": 2048, 00:12:25.912 "data_size": 63488 00:12:25.912 }, 00:12:25.912 { 00:12:25.912 "name": "BaseBdev3", 00:12:25.912 "uuid": "ee6dcd27-62fb-55ac-bd58-5d03fde51b7b", 00:12:25.912 "is_configured": true, 00:12:25.912 "data_offset": 2048, 00:12:25.912 "data_size": 63488 00:12:25.912 }, 00:12:25.912 { 00:12:25.912 "name": "BaseBdev4", 00:12:25.912 "uuid": "5a25f2b0-5c65-5f8f-baeb-76b7b36d0157", 00:12:25.912 "is_configured": true, 00:12:25.912 "data_offset": 2048, 00:12:25.912 "data_size": 63488 00:12:25.912 } 00:12:25.912 ] 00:12:25.912 }' 00:12:25.912 23:46:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:25.912 23:46:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:25.912 23:46:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:25.912 23:46:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:25.912 23:46:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@751 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:25.912 23:46:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@751 -- # jq -r '.[].base_bdevs_list[0].name' 00:12:25.912 23:46:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:25.912 23:46:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:25.912 23:46:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:25.912 23:46:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@751 -- # [[ spare == \s\p\a\r\e ]] 00:12:25.912 23:46:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@754 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:12:25.912 23:46:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:25.912 23:46:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:25.912 [2024-11-26 23:46:14.034192] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:25.912 23:46:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:25.912 23:46:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@755 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:12:25.912 23:46:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:25.912 23:46:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:25.912 23:46:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:25.912 23:46:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:25.912 23:46:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:12:26.171 23:46:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:26.171 23:46:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:26.171 23:46:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:26.171 23:46:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:26.171 23:46:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:26.171 23:46:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:26.171 23:46:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:26.171 23:46:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:26.171 23:46:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:26.171 23:46:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:26.171 "name": "raid_bdev1", 00:12:26.171 "uuid": "1ac4fc07-7ba3-43ad-97f9-b1256b4b254d", 00:12:26.171 "strip_size_kb": 0, 00:12:26.171 "state": "online", 00:12:26.171 "raid_level": "raid1", 00:12:26.171 "superblock": true, 00:12:26.171 "num_base_bdevs": 4, 00:12:26.171 "num_base_bdevs_discovered": 2, 00:12:26.171 "num_base_bdevs_operational": 2, 00:12:26.171 "base_bdevs_list": [ 00:12:26.171 { 00:12:26.171 "name": null, 00:12:26.171 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:26.171 "is_configured": false, 00:12:26.172 "data_offset": 0, 00:12:26.172 "data_size": 63488 00:12:26.172 }, 00:12:26.172 { 00:12:26.172 "name": null, 00:12:26.172 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:26.172 "is_configured": false, 00:12:26.172 "data_offset": 2048, 00:12:26.172 "data_size": 63488 00:12:26.172 }, 00:12:26.172 { 00:12:26.172 "name": "BaseBdev3", 00:12:26.172 "uuid": "ee6dcd27-62fb-55ac-bd58-5d03fde51b7b", 00:12:26.172 "is_configured": true, 00:12:26.172 "data_offset": 2048, 00:12:26.172 "data_size": 63488 00:12:26.172 }, 00:12:26.172 { 00:12:26.172 "name": "BaseBdev4", 00:12:26.172 "uuid": "5a25f2b0-5c65-5f8f-baeb-76b7b36d0157", 00:12:26.172 "is_configured": true, 00:12:26.172 "data_offset": 2048, 00:12:26.172 "data_size": 63488 00:12:26.172 } 00:12:26.172 ] 00:12:26.172 }' 00:12:26.172 23:46:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:26.172 23:46:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:26.444 23:46:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@756 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:12:26.444 23:46:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:26.444 23:46:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:26.444 [2024-11-26 23:46:14.525421] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:26.444 [2024-11-26 23:46:14.525633] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (5) smaller than existing raid bdev raid_bdev1 (6) 00:12:26.444 [2024-11-26 23:46:14.525650] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:12:26.444 [2024-11-26 23:46:14.525695] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:26.444 [2024-11-26 23:46:14.530088] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000033880 00:12:26.444 23:46:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:26.444 23:46:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@757 -- # sleep 1 00:12:26.444 [2024-11-26 23:46:14.531927] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:12:27.824 23:46:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@758 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:27.824 23:46:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:27.824 23:46:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:27.824 23:46:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:27.824 23:46:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:27.824 23:46:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:27.824 23:46:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:27.824 23:46:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:27.824 23:46:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:27.824 23:46:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:27.824 23:46:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:27.824 "name": "raid_bdev1", 00:12:27.824 "uuid": "1ac4fc07-7ba3-43ad-97f9-b1256b4b254d", 00:12:27.824 "strip_size_kb": 0, 00:12:27.824 "state": "online", 00:12:27.824 "raid_level": "raid1", 00:12:27.824 "superblock": true, 00:12:27.824 "num_base_bdevs": 4, 00:12:27.824 "num_base_bdevs_discovered": 3, 00:12:27.824 "num_base_bdevs_operational": 3, 00:12:27.824 "process": { 00:12:27.824 "type": "rebuild", 00:12:27.824 "target": "spare", 00:12:27.824 "progress": { 00:12:27.824 "blocks": 20480, 00:12:27.824 "percent": 32 00:12:27.824 } 00:12:27.824 }, 00:12:27.824 "base_bdevs_list": [ 00:12:27.824 { 00:12:27.824 "name": "spare", 00:12:27.824 "uuid": "9feca6b8-4d7b-539a-9b6c-0d9b36941f36", 00:12:27.824 "is_configured": true, 00:12:27.824 "data_offset": 2048, 00:12:27.824 "data_size": 63488 00:12:27.824 }, 00:12:27.824 { 00:12:27.824 "name": null, 00:12:27.824 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:27.824 "is_configured": false, 00:12:27.824 "data_offset": 2048, 00:12:27.824 "data_size": 63488 00:12:27.824 }, 00:12:27.824 { 00:12:27.824 "name": "BaseBdev3", 00:12:27.824 "uuid": "ee6dcd27-62fb-55ac-bd58-5d03fde51b7b", 00:12:27.824 "is_configured": true, 00:12:27.824 "data_offset": 2048, 00:12:27.824 "data_size": 63488 00:12:27.824 }, 00:12:27.824 { 00:12:27.824 "name": "BaseBdev4", 00:12:27.824 "uuid": "5a25f2b0-5c65-5f8f-baeb-76b7b36d0157", 00:12:27.824 "is_configured": true, 00:12:27.824 "data_offset": 2048, 00:12:27.824 "data_size": 63488 00:12:27.824 } 00:12:27.824 ] 00:12:27.824 }' 00:12:27.824 23:46:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:27.824 23:46:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:27.824 23:46:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:27.824 23:46:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:27.824 23:46:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@761 -- # rpc_cmd bdev_passthru_delete spare 00:12:27.824 23:46:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:27.824 23:46:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:27.824 [2024-11-26 23:46:15.692430] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:27.824 [2024-11-26 23:46:15.735839] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:12:27.824 [2024-11-26 23:46:15.735914] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:27.824 [2024-11-26 23:46:15.735928] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:27.824 [2024-11-26 23:46:15.735937] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:12:27.824 23:46:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:27.824 23:46:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@762 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:12:27.824 23:46:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:27.824 23:46:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:27.824 23:46:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:27.824 23:46:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:27.824 23:46:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:12:27.824 23:46:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:27.824 23:46:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:27.824 23:46:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:27.824 23:46:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:27.824 23:46:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:27.824 23:46:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:27.824 23:46:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:27.824 23:46:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:27.824 23:46:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:27.824 23:46:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:27.824 "name": "raid_bdev1", 00:12:27.824 "uuid": "1ac4fc07-7ba3-43ad-97f9-b1256b4b254d", 00:12:27.824 "strip_size_kb": 0, 00:12:27.824 "state": "online", 00:12:27.824 "raid_level": "raid1", 00:12:27.824 "superblock": true, 00:12:27.824 "num_base_bdevs": 4, 00:12:27.824 "num_base_bdevs_discovered": 2, 00:12:27.824 "num_base_bdevs_operational": 2, 00:12:27.824 "base_bdevs_list": [ 00:12:27.824 { 00:12:27.824 "name": null, 00:12:27.824 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:27.824 "is_configured": false, 00:12:27.824 "data_offset": 0, 00:12:27.824 "data_size": 63488 00:12:27.824 }, 00:12:27.824 { 00:12:27.824 "name": null, 00:12:27.824 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:27.824 "is_configured": false, 00:12:27.824 "data_offset": 2048, 00:12:27.824 "data_size": 63488 00:12:27.824 }, 00:12:27.824 { 00:12:27.824 "name": "BaseBdev3", 00:12:27.824 "uuid": "ee6dcd27-62fb-55ac-bd58-5d03fde51b7b", 00:12:27.824 "is_configured": true, 00:12:27.824 "data_offset": 2048, 00:12:27.824 "data_size": 63488 00:12:27.824 }, 00:12:27.824 { 00:12:27.824 "name": "BaseBdev4", 00:12:27.824 "uuid": "5a25f2b0-5c65-5f8f-baeb-76b7b36d0157", 00:12:27.824 "is_configured": true, 00:12:27.824 "data_offset": 2048, 00:12:27.824 "data_size": 63488 00:12:27.824 } 00:12:27.824 ] 00:12:27.824 }' 00:12:27.824 23:46:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:27.824 23:46:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:28.085 23:46:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@763 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:12:28.085 23:46:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:28.085 23:46:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:28.345 [2024-11-26 23:46:16.219467] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:12:28.345 [2024-11-26 23:46:16.219532] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:28.345 [2024-11-26 23:46:16.219556] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ba80 00:12:28.345 [2024-11-26 23:46:16.219567] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:28.345 [2024-11-26 23:46:16.219987] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:28.345 [2024-11-26 23:46:16.220022] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:12:28.345 [2024-11-26 23:46:16.220107] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:12:28.345 [2024-11-26 23:46:16.220123] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (5) smaller than existing raid bdev raid_bdev1 (6) 00:12:28.345 [2024-11-26 23:46:16.220132] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:12:28.345 [2024-11-26 23:46:16.220163] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:28.345 [2024-11-26 23:46:16.224484] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000033950 00:12:28.345 spare 00:12:28.345 23:46:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:28.345 23:46:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@764 -- # sleep 1 00:12:28.345 [2024-11-26 23:46:16.226354] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:12:29.283 23:46:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@765 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:29.283 23:46:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:29.283 23:46:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:29.283 23:46:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:29.283 23:46:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:29.283 23:46:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:29.283 23:46:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:29.283 23:46:17 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:29.283 23:46:17 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:29.283 23:46:17 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:29.283 23:46:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:29.283 "name": "raid_bdev1", 00:12:29.283 "uuid": "1ac4fc07-7ba3-43ad-97f9-b1256b4b254d", 00:12:29.283 "strip_size_kb": 0, 00:12:29.283 "state": "online", 00:12:29.283 "raid_level": "raid1", 00:12:29.283 "superblock": true, 00:12:29.283 "num_base_bdevs": 4, 00:12:29.283 "num_base_bdevs_discovered": 3, 00:12:29.283 "num_base_bdevs_operational": 3, 00:12:29.283 "process": { 00:12:29.283 "type": "rebuild", 00:12:29.283 "target": "spare", 00:12:29.283 "progress": { 00:12:29.283 "blocks": 20480, 00:12:29.283 "percent": 32 00:12:29.283 } 00:12:29.283 }, 00:12:29.283 "base_bdevs_list": [ 00:12:29.283 { 00:12:29.283 "name": "spare", 00:12:29.283 "uuid": "9feca6b8-4d7b-539a-9b6c-0d9b36941f36", 00:12:29.283 "is_configured": true, 00:12:29.283 "data_offset": 2048, 00:12:29.283 "data_size": 63488 00:12:29.283 }, 00:12:29.283 { 00:12:29.283 "name": null, 00:12:29.283 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:29.283 "is_configured": false, 00:12:29.283 "data_offset": 2048, 00:12:29.283 "data_size": 63488 00:12:29.283 }, 00:12:29.283 { 00:12:29.283 "name": "BaseBdev3", 00:12:29.283 "uuid": "ee6dcd27-62fb-55ac-bd58-5d03fde51b7b", 00:12:29.283 "is_configured": true, 00:12:29.283 "data_offset": 2048, 00:12:29.283 "data_size": 63488 00:12:29.283 }, 00:12:29.283 { 00:12:29.283 "name": "BaseBdev4", 00:12:29.283 "uuid": "5a25f2b0-5c65-5f8f-baeb-76b7b36d0157", 00:12:29.283 "is_configured": true, 00:12:29.283 "data_offset": 2048, 00:12:29.283 "data_size": 63488 00:12:29.283 } 00:12:29.283 ] 00:12:29.283 }' 00:12:29.283 23:46:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:29.283 23:46:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:29.283 23:46:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:29.283 23:46:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:29.283 23:46:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@768 -- # rpc_cmd bdev_passthru_delete spare 00:12:29.283 23:46:17 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:29.283 23:46:17 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:29.283 [2024-11-26 23:46:17.390464] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:29.543 [2024-11-26 23:46:17.430138] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:12:29.543 [2024-11-26 23:46:17.430211] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:29.543 [2024-11-26 23:46:17.430228] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:29.543 [2024-11-26 23:46:17.430235] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:12:29.543 23:46:17 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:29.543 23:46:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@769 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:12:29.543 23:46:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:29.543 23:46:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:29.543 23:46:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:29.543 23:46:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:29.543 23:46:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:12:29.543 23:46:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:29.543 23:46:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:29.543 23:46:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:29.543 23:46:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:29.543 23:46:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:29.543 23:46:17 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:29.544 23:46:17 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:29.544 23:46:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:29.544 23:46:17 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:29.544 23:46:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:29.544 "name": "raid_bdev1", 00:12:29.544 "uuid": "1ac4fc07-7ba3-43ad-97f9-b1256b4b254d", 00:12:29.544 "strip_size_kb": 0, 00:12:29.544 "state": "online", 00:12:29.544 "raid_level": "raid1", 00:12:29.544 "superblock": true, 00:12:29.544 "num_base_bdevs": 4, 00:12:29.544 "num_base_bdevs_discovered": 2, 00:12:29.544 "num_base_bdevs_operational": 2, 00:12:29.544 "base_bdevs_list": [ 00:12:29.544 { 00:12:29.544 "name": null, 00:12:29.544 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:29.544 "is_configured": false, 00:12:29.544 "data_offset": 0, 00:12:29.544 "data_size": 63488 00:12:29.544 }, 00:12:29.544 { 00:12:29.544 "name": null, 00:12:29.544 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:29.544 "is_configured": false, 00:12:29.544 "data_offset": 2048, 00:12:29.544 "data_size": 63488 00:12:29.544 }, 00:12:29.544 { 00:12:29.544 "name": "BaseBdev3", 00:12:29.544 "uuid": "ee6dcd27-62fb-55ac-bd58-5d03fde51b7b", 00:12:29.544 "is_configured": true, 00:12:29.544 "data_offset": 2048, 00:12:29.544 "data_size": 63488 00:12:29.544 }, 00:12:29.544 { 00:12:29.544 "name": "BaseBdev4", 00:12:29.544 "uuid": "5a25f2b0-5c65-5f8f-baeb-76b7b36d0157", 00:12:29.544 "is_configured": true, 00:12:29.544 "data_offset": 2048, 00:12:29.544 "data_size": 63488 00:12:29.544 } 00:12:29.544 ] 00:12:29.544 }' 00:12:29.544 23:46:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:29.544 23:46:17 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:29.802 23:46:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@770 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:29.802 23:46:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:29.802 23:46:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:29.802 23:46:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:29.802 23:46:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:29.802 23:46:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:29.802 23:46:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:29.802 23:46:17 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:29.802 23:46:17 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:30.061 23:46:17 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:30.061 23:46:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:30.061 "name": "raid_bdev1", 00:12:30.061 "uuid": "1ac4fc07-7ba3-43ad-97f9-b1256b4b254d", 00:12:30.061 "strip_size_kb": 0, 00:12:30.061 "state": "online", 00:12:30.061 "raid_level": "raid1", 00:12:30.061 "superblock": true, 00:12:30.061 "num_base_bdevs": 4, 00:12:30.061 "num_base_bdevs_discovered": 2, 00:12:30.061 "num_base_bdevs_operational": 2, 00:12:30.061 "base_bdevs_list": [ 00:12:30.061 { 00:12:30.061 "name": null, 00:12:30.061 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:30.061 "is_configured": false, 00:12:30.061 "data_offset": 0, 00:12:30.061 "data_size": 63488 00:12:30.061 }, 00:12:30.061 { 00:12:30.061 "name": null, 00:12:30.061 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:30.061 "is_configured": false, 00:12:30.061 "data_offset": 2048, 00:12:30.061 "data_size": 63488 00:12:30.061 }, 00:12:30.061 { 00:12:30.061 "name": "BaseBdev3", 00:12:30.061 "uuid": "ee6dcd27-62fb-55ac-bd58-5d03fde51b7b", 00:12:30.061 "is_configured": true, 00:12:30.061 "data_offset": 2048, 00:12:30.061 "data_size": 63488 00:12:30.061 }, 00:12:30.061 { 00:12:30.061 "name": "BaseBdev4", 00:12:30.061 "uuid": "5a25f2b0-5c65-5f8f-baeb-76b7b36d0157", 00:12:30.061 "is_configured": true, 00:12:30.061 "data_offset": 2048, 00:12:30.061 "data_size": 63488 00:12:30.061 } 00:12:30.061 ] 00:12:30.061 }' 00:12:30.061 23:46:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:30.061 23:46:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:30.061 23:46:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:30.061 23:46:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:30.061 23:46:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@773 -- # rpc_cmd bdev_passthru_delete BaseBdev1 00:12:30.061 23:46:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:30.061 23:46:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:30.061 23:46:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:30.061 23:46:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@774 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:12:30.061 23:46:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:30.061 23:46:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:30.061 [2024-11-26 23:46:18.077670] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:12:30.061 [2024-11-26 23:46:18.077739] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:30.061 [2024-11-26 23:46:18.077761] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000c080 00:12:30.061 [2024-11-26 23:46:18.077770] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:30.061 [2024-11-26 23:46:18.078152] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:30.061 [2024-11-26 23:46:18.078176] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:12:30.061 [2024-11-26 23:46:18.078241] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:12:30.061 [2024-11-26 23:46:18.078254] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (6) 00:12:30.061 [2024-11-26 23:46:18.078263] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:12:30.061 [2024-11-26 23:46:18.078271] bdev_raid.c:3894:raid_bdev_examine_done: *ERROR*: Failed to examine bdev BaseBdev1: Invalid argument 00:12:30.061 BaseBdev1 00:12:30.061 23:46:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:30.061 23:46:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@775 -- # sleep 1 00:12:31.000 23:46:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@776 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:12:31.000 23:46:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:31.000 23:46:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:31.000 23:46:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:31.000 23:46:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:31.000 23:46:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:12:31.000 23:46:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:31.000 23:46:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:31.000 23:46:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:31.000 23:46:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:31.000 23:46:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:31.000 23:46:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:31.001 23:46:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:31.001 23:46:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:31.001 23:46:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:31.259 23:46:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:31.259 "name": "raid_bdev1", 00:12:31.259 "uuid": "1ac4fc07-7ba3-43ad-97f9-b1256b4b254d", 00:12:31.259 "strip_size_kb": 0, 00:12:31.259 "state": "online", 00:12:31.259 "raid_level": "raid1", 00:12:31.259 "superblock": true, 00:12:31.259 "num_base_bdevs": 4, 00:12:31.259 "num_base_bdevs_discovered": 2, 00:12:31.259 "num_base_bdevs_operational": 2, 00:12:31.259 "base_bdevs_list": [ 00:12:31.259 { 00:12:31.259 "name": null, 00:12:31.259 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:31.259 "is_configured": false, 00:12:31.259 "data_offset": 0, 00:12:31.259 "data_size": 63488 00:12:31.259 }, 00:12:31.259 { 00:12:31.259 "name": null, 00:12:31.259 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:31.259 "is_configured": false, 00:12:31.259 "data_offset": 2048, 00:12:31.259 "data_size": 63488 00:12:31.259 }, 00:12:31.259 { 00:12:31.259 "name": "BaseBdev3", 00:12:31.259 "uuid": "ee6dcd27-62fb-55ac-bd58-5d03fde51b7b", 00:12:31.259 "is_configured": true, 00:12:31.259 "data_offset": 2048, 00:12:31.259 "data_size": 63488 00:12:31.259 }, 00:12:31.259 { 00:12:31.259 "name": "BaseBdev4", 00:12:31.259 "uuid": "5a25f2b0-5c65-5f8f-baeb-76b7b36d0157", 00:12:31.259 "is_configured": true, 00:12:31.259 "data_offset": 2048, 00:12:31.259 "data_size": 63488 00:12:31.259 } 00:12:31.259 ] 00:12:31.259 }' 00:12:31.259 23:46:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:31.259 23:46:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:31.518 23:46:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@777 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:31.518 23:46:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:31.518 23:46:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:31.518 23:46:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:31.518 23:46:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:31.518 23:46:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:31.518 23:46:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:31.518 23:46:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:31.518 23:46:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:31.518 23:46:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:31.518 23:46:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:31.518 "name": "raid_bdev1", 00:12:31.518 "uuid": "1ac4fc07-7ba3-43ad-97f9-b1256b4b254d", 00:12:31.518 "strip_size_kb": 0, 00:12:31.518 "state": "online", 00:12:31.518 "raid_level": "raid1", 00:12:31.518 "superblock": true, 00:12:31.518 "num_base_bdevs": 4, 00:12:31.518 "num_base_bdevs_discovered": 2, 00:12:31.518 "num_base_bdevs_operational": 2, 00:12:31.518 "base_bdevs_list": [ 00:12:31.518 { 00:12:31.518 "name": null, 00:12:31.518 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:31.518 "is_configured": false, 00:12:31.518 "data_offset": 0, 00:12:31.518 "data_size": 63488 00:12:31.518 }, 00:12:31.518 { 00:12:31.518 "name": null, 00:12:31.518 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:31.518 "is_configured": false, 00:12:31.518 "data_offset": 2048, 00:12:31.518 "data_size": 63488 00:12:31.518 }, 00:12:31.518 { 00:12:31.518 "name": "BaseBdev3", 00:12:31.518 "uuid": "ee6dcd27-62fb-55ac-bd58-5d03fde51b7b", 00:12:31.518 "is_configured": true, 00:12:31.518 "data_offset": 2048, 00:12:31.518 "data_size": 63488 00:12:31.518 }, 00:12:31.518 { 00:12:31.518 "name": "BaseBdev4", 00:12:31.518 "uuid": "5a25f2b0-5c65-5f8f-baeb-76b7b36d0157", 00:12:31.518 "is_configured": true, 00:12:31.518 "data_offset": 2048, 00:12:31.518 "data_size": 63488 00:12:31.518 } 00:12:31.518 ] 00:12:31.518 }' 00:12:31.518 23:46:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:31.518 23:46:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:31.518 23:46:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:31.777 23:46:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:31.777 23:46:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@778 -- # NOT rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:12:31.777 23:46:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@652 -- # local es=0 00:12:31.777 23:46:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:12:31.777 23:46:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:12:31.777 23:46:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:12:31.777 23:46:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:12:31.777 23:46:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:12:31.777 23:46:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:12:31.778 23:46:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:31.778 23:46:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:31.778 [2024-11-26 23:46:19.699071] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:12:31.778 [2024-11-26 23:46:19.699233] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (6) 00:12:31.778 [2024-11-26 23:46:19.699256] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:12:31.778 request: 00:12:31.778 { 00:12:31.778 "base_bdev": "BaseBdev1", 00:12:31.778 "raid_bdev": "raid_bdev1", 00:12:31.778 "method": "bdev_raid_add_base_bdev", 00:12:31.778 "req_id": 1 00:12:31.778 } 00:12:31.778 Got JSON-RPC error response 00:12:31.778 response: 00:12:31.778 { 00:12:31.778 "code": -22, 00:12:31.778 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:12:31.778 } 00:12:31.778 23:46:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:12:31.778 23:46:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@655 -- # es=1 00:12:31.778 23:46:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:12:31.778 23:46:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:12:31.778 23:46:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:12:31.778 23:46:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@779 -- # sleep 1 00:12:32.891 23:46:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@780 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:12:32.891 23:46:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:32.891 23:46:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:32.891 23:46:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:32.891 23:46:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:32.891 23:46:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:12:32.891 23:46:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:32.891 23:46:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:32.891 23:46:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:32.891 23:46:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:32.891 23:46:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:32.891 23:46:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:32.892 23:46:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:32.892 23:46:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:32.892 23:46:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:32.892 23:46:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:32.892 "name": "raid_bdev1", 00:12:32.892 "uuid": "1ac4fc07-7ba3-43ad-97f9-b1256b4b254d", 00:12:32.892 "strip_size_kb": 0, 00:12:32.892 "state": "online", 00:12:32.892 "raid_level": "raid1", 00:12:32.892 "superblock": true, 00:12:32.892 "num_base_bdevs": 4, 00:12:32.892 "num_base_bdevs_discovered": 2, 00:12:32.892 "num_base_bdevs_operational": 2, 00:12:32.892 "base_bdevs_list": [ 00:12:32.892 { 00:12:32.892 "name": null, 00:12:32.892 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:32.892 "is_configured": false, 00:12:32.892 "data_offset": 0, 00:12:32.892 "data_size": 63488 00:12:32.892 }, 00:12:32.892 { 00:12:32.892 "name": null, 00:12:32.892 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:32.892 "is_configured": false, 00:12:32.892 "data_offset": 2048, 00:12:32.892 "data_size": 63488 00:12:32.892 }, 00:12:32.892 { 00:12:32.892 "name": "BaseBdev3", 00:12:32.892 "uuid": "ee6dcd27-62fb-55ac-bd58-5d03fde51b7b", 00:12:32.892 "is_configured": true, 00:12:32.892 "data_offset": 2048, 00:12:32.892 "data_size": 63488 00:12:32.892 }, 00:12:32.892 { 00:12:32.892 "name": "BaseBdev4", 00:12:32.892 "uuid": "5a25f2b0-5c65-5f8f-baeb-76b7b36d0157", 00:12:32.892 "is_configured": true, 00:12:32.892 "data_offset": 2048, 00:12:32.892 "data_size": 63488 00:12:32.892 } 00:12:32.892 ] 00:12:32.892 }' 00:12:32.892 23:46:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:32.892 23:46:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:33.151 23:46:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@781 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:33.151 23:46:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:33.151 23:46:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:33.151 23:46:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:33.151 23:46:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:33.151 23:46:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:33.151 23:46:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:33.151 23:46:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:33.151 23:46:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:33.151 23:46:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:33.151 23:46:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:33.151 "name": "raid_bdev1", 00:12:33.151 "uuid": "1ac4fc07-7ba3-43ad-97f9-b1256b4b254d", 00:12:33.151 "strip_size_kb": 0, 00:12:33.151 "state": "online", 00:12:33.151 "raid_level": "raid1", 00:12:33.151 "superblock": true, 00:12:33.151 "num_base_bdevs": 4, 00:12:33.151 "num_base_bdevs_discovered": 2, 00:12:33.151 "num_base_bdevs_operational": 2, 00:12:33.151 "base_bdevs_list": [ 00:12:33.151 { 00:12:33.151 "name": null, 00:12:33.151 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:33.151 "is_configured": false, 00:12:33.151 "data_offset": 0, 00:12:33.151 "data_size": 63488 00:12:33.151 }, 00:12:33.151 { 00:12:33.151 "name": null, 00:12:33.151 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:33.151 "is_configured": false, 00:12:33.151 "data_offset": 2048, 00:12:33.151 "data_size": 63488 00:12:33.151 }, 00:12:33.151 { 00:12:33.151 "name": "BaseBdev3", 00:12:33.151 "uuid": "ee6dcd27-62fb-55ac-bd58-5d03fde51b7b", 00:12:33.151 "is_configured": true, 00:12:33.151 "data_offset": 2048, 00:12:33.151 "data_size": 63488 00:12:33.151 }, 00:12:33.151 { 00:12:33.151 "name": "BaseBdev4", 00:12:33.151 "uuid": "5a25f2b0-5c65-5f8f-baeb-76b7b36d0157", 00:12:33.151 "is_configured": true, 00:12:33.151 "data_offset": 2048, 00:12:33.151 "data_size": 63488 00:12:33.151 } 00:12:33.151 ] 00:12:33.151 }' 00:12:33.151 23:46:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:33.151 23:46:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:33.151 23:46:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:33.411 23:46:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:33.411 23:46:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@784 -- # killprocess 89429 00:12:33.411 23:46:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@954 -- # '[' -z 89429 ']' 00:12:33.411 23:46:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@958 -- # kill -0 89429 00:12:33.411 23:46:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@959 -- # uname 00:12:33.411 23:46:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:12:33.411 23:46:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 89429 00:12:33.411 23:46:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:12:33.411 23:46:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:12:33.411 killing process with pid 89429 00:12:33.411 23:46:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@972 -- # echo 'killing process with pid 89429' 00:12:33.411 23:46:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@973 -- # kill 89429 00:12:33.411 Received shutdown signal, test time was about 17.803296 seconds 00:12:33.411 00:12:33.411 Latency(us) 00:12:33.411 [2024-11-26T23:46:21.543Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:12:33.411 [2024-11-26T23:46:21.543Z] =================================================================================================================== 00:12:33.411 [2024-11-26T23:46:21.543Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:12:33.411 [2024-11-26 23:46:21.329750] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:12:33.411 [2024-11-26 23:46:21.329889] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:12:33.411 23:46:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@978 -- # wait 89429 00:12:33.411 [2024-11-26 23:46:21.329959] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:12:33.411 [2024-11-26 23:46:21.329974] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state offline 00:12:33.411 [2024-11-26 23:46:21.375220] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:12:33.671 23:46:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@786 -- # return 0 00:12:33.671 00:12:33.671 real 0m19.719s 00:12:33.671 user 0m26.368s 00:12:33.671 sys 0m2.492s 00:12:33.671 23:46:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@1130 -- # xtrace_disable 00:12:33.671 23:46:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:33.671 ************************************ 00:12:33.671 END TEST raid_rebuild_test_sb_io 00:12:33.671 ************************************ 00:12:33.671 23:46:21 bdev_raid -- bdev/bdev_raid.sh@985 -- # for n in {3..4} 00:12:33.671 23:46:21 bdev_raid -- bdev/bdev_raid.sh@986 -- # run_test raid5f_state_function_test raid_state_function_test raid5f 3 false 00:12:33.671 23:46:21 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:12:33.671 23:46:21 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:12:33.671 23:46:21 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:12:33.671 ************************************ 00:12:33.671 START TEST raid5f_state_function_test 00:12:33.671 ************************************ 00:12:33.671 23:46:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@1129 -- # raid_state_function_test raid5f 3 false 00:12:33.671 23:46:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=raid5f 00:12:33.671 23:46:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=3 00:12:33.671 23:46:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:12:33.671 23:46:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:12:33.671 23:46:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:12:33.671 23:46:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:12:33.671 23:46:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:12:33.671 23:46:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:12:33.671 23:46:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:12:33.671 23:46:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:12:33.671 23:46:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:12:33.671 23:46:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:12:33.671 23:46:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:12:33.671 23:46:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:12:33.671 23:46:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:12:33.671 23:46:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:12:33.671 23:46:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:12:33.671 23:46:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:12:33.671 23:46:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:12:33.671 23:46:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:12:33.671 23:46:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:12:33.671 23:46:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' raid5f '!=' raid1 ']' 00:12:33.671 23:46:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:12:33.671 23:46:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:12:33.671 23:46:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:12:33.671 23:46:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:12:33.671 23:46:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=90134 00:12:33.672 23:46:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:12:33.672 23:46:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 90134' 00:12:33.672 Process raid pid: 90134 00:12:33.672 23:46:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 90134 00:12:33.672 23:46:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@835 -- # '[' -z 90134 ']' 00:12:33.672 23:46:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:12:33.672 23:46:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:12:33.672 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:12:33.672 23:46:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:12:33.672 23:46:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:12:33.672 23:46:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:12:33.672 [2024-11-26 23:46:21.751218] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:12:33.672 [2024-11-26 23:46:21.751379] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:12:33.932 [2024-11-26 23:46:21.883555] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:12:33.932 [2024-11-26 23:46:21.907615] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:12:33.932 [2024-11-26 23:46:21.950039] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:12:33.932 [2024-11-26 23:46:21.950085] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:12:34.501 23:46:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:12:34.501 23:46:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@868 -- # return 0 00:12:34.501 23:46:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:12:34.501 23:46:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:34.501 23:46:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:12:34.501 [2024-11-26 23:46:22.576105] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:12:34.501 [2024-11-26 23:46:22.576176] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:12:34.501 [2024-11-26 23:46:22.576186] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:12:34.501 [2024-11-26 23:46:22.576195] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:12:34.501 [2024-11-26 23:46:22.576201] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:12:34.501 [2024-11-26 23:46:22.576211] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:12:34.501 23:46:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:34.501 23:46:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:12:34.501 23:46:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:12:34.501 23:46:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:12:34.501 23:46:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:12:34.501 23:46:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:12:34.501 23:46:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:34.501 23:46:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:34.501 23:46:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:34.501 23:46:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:34.501 23:46:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:34.501 23:46:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:34.501 23:46:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:12:34.501 23:46:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:34.501 23:46:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:12:34.501 23:46:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:34.761 23:46:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:34.761 "name": "Existed_Raid", 00:12:34.761 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:34.761 "strip_size_kb": 64, 00:12:34.761 "state": "configuring", 00:12:34.761 "raid_level": "raid5f", 00:12:34.761 "superblock": false, 00:12:34.761 "num_base_bdevs": 3, 00:12:34.761 "num_base_bdevs_discovered": 0, 00:12:34.761 "num_base_bdevs_operational": 3, 00:12:34.761 "base_bdevs_list": [ 00:12:34.761 { 00:12:34.761 "name": "BaseBdev1", 00:12:34.761 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:34.761 "is_configured": false, 00:12:34.761 "data_offset": 0, 00:12:34.761 "data_size": 0 00:12:34.761 }, 00:12:34.761 { 00:12:34.761 "name": "BaseBdev2", 00:12:34.761 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:34.761 "is_configured": false, 00:12:34.761 "data_offset": 0, 00:12:34.761 "data_size": 0 00:12:34.761 }, 00:12:34.761 { 00:12:34.761 "name": "BaseBdev3", 00:12:34.761 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:34.761 "is_configured": false, 00:12:34.761 "data_offset": 0, 00:12:34.761 "data_size": 0 00:12:34.761 } 00:12:34.761 ] 00:12:34.761 }' 00:12:34.761 23:46:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:34.761 23:46:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:12:35.021 23:46:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:12:35.021 23:46:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:35.022 23:46:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:12:35.022 [2024-11-26 23:46:23.023256] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:12:35.022 [2024-11-26 23:46:23.023299] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:12:35.022 23:46:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:35.022 23:46:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:12:35.022 23:46:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:35.022 23:46:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:12:35.022 [2024-11-26 23:46:23.035254] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:12:35.022 [2024-11-26 23:46:23.035299] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:12:35.022 [2024-11-26 23:46:23.035307] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:12:35.022 [2024-11-26 23:46:23.035316] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:12:35.022 [2024-11-26 23:46:23.035322] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:12:35.022 [2024-11-26 23:46:23.035330] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:12:35.022 23:46:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:35.022 23:46:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:12:35.022 23:46:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:35.022 23:46:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:12:35.022 [2024-11-26 23:46:23.055880] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:12:35.022 BaseBdev1 00:12:35.022 23:46:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:35.022 23:46:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:12:35.022 23:46:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:12:35.022 23:46:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:12:35.022 23:46:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:12:35.022 23:46:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:12:35.022 23:46:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:12:35.022 23:46:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:12:35.022 23:46:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:35.022 23:46:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:12:35.022 23:46:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:35.022 23:46:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:12:35.022 23:46:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:35.022 23:46:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:12:35.022 [ 00:12:35.022 { 00:12:35.022 "name": "BaseBdev1", 00:12:35.022 "aliases": [ 00:12:35.022 "b159283f-3797-4f0e-b5d5-044d342ede02" 00:12:35.022 ], 00:12:35.022 "product_name": "Malloc disk", 00:12:35.022 "block_size": 512, 00:12:35.022 "num_blocks": 65536, 00:12:35.022 "uuid": "b159283f-3797-4f0e-b5d5-044d342ede02", 00:12:35.022 "assigned_rate_limits": { 00:12:35.022 "rw_ios_per_sec": 0, 00:12:35.022 "rw_mbytes_per_sec": 0, 00:12:35.022 "r_mbytes_per_sec": 0, 00:12:35.022 "w_mbytes_per_sec": 0 00:12:35.022 }, 00:12:35.022 "claimed": true, 00:12:35.022 "claim_type": "exclusive_write", 00:12:35.022 "zoned": false, 00:12:35.022 "supported_io_types": { 00:12:35.022 "read": true, 00:12:35.022 "write": true, 00:12:35.022 "unmap": true, 00:12:35.022 "flush": true, 00:12:35.022 "reset": true, 00:12:35.022 "nvme_admin": false, 00:12:35.022 "nvme_io": false, 00:12:35.022 "nvme_io_md": false, 00:12:35.022 "write_zeroes": true, 00:12:35.022 "zcopy": true, 00:12:35.022 "get_zone_info": false, 00:12:35.022 "zone_management": false, 00:12:35.022 "zone_append": false, 00:12:35.022 "compare": false, 00:12:35.022 "compare_and_write": false, 00:12:35.022 "abort": true, 00:12:35.022 "seek_hole": false, 00:12:35.022 "seek_data": false, 00:12:35.022 "copy": true, 00:12:35.022 "nvme_iov_md": false 00:12:35.022 }, 00:12:35.022 "memory_domains": [ 00:12:35.022 { 00:12:35.022 "dma_device_id": "system", 00:12:35.022 "dma_device_type": 1 00:12:35.022 }, 00:12:35.022 { 00:12:35.022 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:12:35.022 "dma_device_type": 2 00:12:35.022 } 00:12:35.022 ], 00:12:35.022 "driver_specific": {} 00:12:35.022 } 00:12:35.022 ] 00:12:35.022 23:46:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:35.022 23:46:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:12:35.022 23:46:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:12:35.022 23:46:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:12:35.022 23:46:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:12:35.022 23:46:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:12:35.022 23:46:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:12:35.022 23:46:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:35.022 23:46:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:35.022 23:46:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:35.022 23:46:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:35.022 23:46:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:35.022 23:46:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:35.022 23:46:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:12:35.022 23:46:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:35.022 23:46:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:12:35.022 23:46:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:35.022 23:46:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:35.022 "name": "Existed_Raid", 00:12:35.022 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:35.022 "strip_size_kb": 64, 00:12:35.022 "state": "configuring", 00:12:35.022 "raid_level": "raid5f", 00:12:35.022 "superblock": false, 00:12:35.022 "num_base_bdevs": 3, 00:12:35.022 "num_base_bdevs_discovered": 1, 00:12:35.022 "num_base_bdevs_operational": 3, 00:12:35.022 "base_bdevs_list": [ 00:12:35.022 { 00:12:35.022 "name": "BaseBdev1", 00:12:35.022 "uuid": "b159283f-3797-4f0e-b5d5-044d342ede02", 00:12:35.022 "is_configured": true, 00:12:35.022 "data_offset": 0, 00:12:35.022 "data_size": 65536 00:12:35.022 }, 00:12:35.022 { 00:12:35.022 "name": "BaseBdev2", 00:12:35.022 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:35.022 "is_configured": false, 00:12:35.022 "data_offset": 0, 00:12:35.022 "data_size": 0 00:12:35.022 }, 00:12:35.022 { 00:12:35.022 "name": "BaseBdev3", 00:12:35.022 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:35.022 "is_configured": false, 00:12:35.022 "data_offset": 0, 00:12:35.022 "data_size": 0 00:12:35.022 } 00:12:35.022 ] 00:12:35.022 }' 00:12:35.022 23:46:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:35.022 23:46:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:12:35.593 23:46:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:12:35.593 23:46:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:35.593 23:46:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:12:35.593 [2024-11-26 23:46:23.535062] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:12:35.593 [2024-11-26 23:46:23.535146] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:12:35.593 23:46:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:35.593 23:46:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:12:35.593 23:46:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:35.593 23:46:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:12:35.593 [2024-11-26 23:46:23.547077] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:12:35.593 [2024-11-26 23:46:23.548866] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:12:35.593 [2024-11-26 23:46:23.548907] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:12:35.593 [2024-11-26 23:46:23.548916] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:12:35.593 [2024-11-26 23:46:23.548927] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:12:35.593 23:46:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:35.593 23:46:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:12:35.593 23:46:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:12:35.593 23:46:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:12:35.593 23:46:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:12:35.593 23:46:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:12:35.593 23:46:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:12:35.593 23:46:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:12:35.593 23:46:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:35.593 23:46:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:35.593 23:46:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:35.594 23:46:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:35.594 23:46:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:35.594 23:46:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:12:35.594 23:46:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:35.594 23:46:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:35.594 23:46:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:12:35.594 23:46:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:35.594 23:46:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:35.594 "name": "Existed_Raid", 00:12:35.594 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:35.594 "strip_size_kb": 64, 00:12:35.594 "state": "configuring", 00:12:35.594 "raid_level": "raid5f", 00:12:35.594 "superblock": false, 00:12:35.594 "num_base_bdevs": 3, 00:12:35.594 "num_base_bdevs_discovered": 1, 00:12:35.594 "num_base_bdevs_operational": 3, 00:12:35.594 "base_bdevs_list": [ 00:12:35.594 { 00:12:35.594 "name": "BaseBdev1", 00:12:35.594 "uuid": "b159283f-3797-4f0e-b5d5-044d342ede02", 00:12:35.594 "is_configured": true, 00:12:35.594 "data_offset": 0, 00:12:35.594 "data_size": 65536 00:12:35.594 }, 00:12:35.594 { 00:12:35.594 "name": "BaseBdev2", 00:12:35.594 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:35.594 "is_configured": false, 00:12:35.594 "data_offset": 0, 00:12:35.594 "data_size": 0 00:12:35.594 }, 00:12:35.594 { 00:12:35.594 "name": "BaseBdev3", 00:12:35.594 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:35.594 "is_configured": false, 00:12:35.594 "data_offset": 0, 00:12:35.594 "data_size": 0 00:12:35.594 } 00:12:35.594 ] 00:12:35.594 }' 00:12:35.594 23:46:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:35.594 23:46:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:12:36.164 23:46:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:12:36.164 23:46:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:36.164 23:46:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:12:36.164 [2024-11-26 23:46:24.041036] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:12:36.164 BaseBdev2 00:12:36.164 23:46:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:36.164 23:46:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:12:36.164 23:46:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:12:36.164 23:46:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:12:36.164 23:46:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:12:36.164 23:46:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:12:36.164 23:46:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:12:36.164 23:46:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:12:36.164 23:46:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:36.164 23:46:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:12:36.164 23:46:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:36.164 23:46:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:12:36.164 23:46:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:36.164 23:46:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:12:36.164 [ 00:12:36.164 { 00:12:36.164 "name": "BaseBdev2", 00:12:36.164 "aliases": [ 00:12:36.164 "70351ef2-5c06-4d83-8e05-0589c6533436" 00:12:36.164 ], 00:12:36.164 "product_name": "Malloc disk", 00:12:36.164 "block_size": 512, 00:12:36.164 "num_blocks": 65536, 00:12:36.164 "uuid": "70351ef2-5c06-4d83-8e05-0589c6533436", 00:12:36.164 "assigned_rate_limits": { 00:12:36.164 "rw_ios_per_sec": 0, 00:12:36.164 "rw_mbytes_per_sec": 0, 00:12:36.164 "r_mbytes_per_sec": 0, 00:12:36.164 "w_mbytes_per_sec": 0 00:12:36.164 }, 00:12:36.164 "claimed": true, 00:12:36.164 "claim_type": "exclusive_write", 00:12:36.164 "zoned": false, 00:12:36.164 "supported_io_types": { 00:12:36.164 "read": true, 00:12:36.164 "write": true, 00:12:36.164 "unmap": true, 00:12:36.164 "flush": true, 00:12:36.164 "reset": true, 00:12:36.164 "nvme_admin": false, 00:12:36.164 "nvme_io": false, 00:12:36.164 "nvme_io_md": false, 00:12:36.164 "write_zeroes": true, 00:12:36.164 "zcopy": true, 00:12:36.164 "get_zone_info": false, 00:12:36.164 "zone_management": false, 00:12:36.164 "zone_append": false, 00:12:36.164 "compare": false, 00:12:36.164 "compare_and_write": false, 00:12:36.164 "abort": true, 00:12:36.164 "seek_hole": false, 00:12:36.164 "seek_data": false, 00:12:36.164 "copy": true, 00:12:36.164 "nvme_iov_md": false 00:12:36.164 }, 00:12:36.164 "memory_domains": [ 00:12:36.164 { 00:12:36.164 "dma_device_id": "system", 00:12:36.164 "dma_device_type": 1 00:12:36.164 }, 00:12:36.164 { 00:12:36.164 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:12:36.164 "dma_device_type": 2 00:12:36.164 } 00:12:36.164 ], 00:12:36.164 "driver_specific": {} 00:12:36.164 } 00:12:36.164 ] 00:12:36.164 23:46:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:36.164 23:46:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:12:36.164 23:46:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:12:36.164 23:46:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:12:36.164 23:46:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:12:36.164 23:46:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:12:36.164 23:46:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:12:36.164 23:46:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:12:36.164 23:46:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:12:36.164 23:46:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:36.164 23:46:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:36.164 23:46:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:36.164 23:46:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:36.164 23:46:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:36.164 23:46:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:36.164 23:46:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:12:36.164 23:46:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:36.164 23:46:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:12:36.165 23:46:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:36.165 23:46:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:36.165 "name": "Existed_Raid", 00:12:36.165 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:36.165 "strip_size_kb": 64, 00:12:36.165 "state": "configuring", 00:12:36.165 "raid_level": "raid5f", 00:12:36.165 "superblock": false, 00:12:36.165 "num_base_bdevs": 3, 00:12:36.165 "num_base_bdevs_discovered": 2, 00:12:36.165 "num_base_bdevs_operational": 3, 00:12:36.165 "base_bdevs_list": [ 00:12:36.165 { 00:12:36.165 "name": "BaseBdev1", 00:12:36.165 "uuid": "b159283f-3797-4f0e-b5d5-044d342ede02", 00:12:36.165 "is_configured": true, 00:12:36.165 "data_offset": 0, 00:12:36.165 "data_size": 65536 00:12:36.165 }, 00:12:36.165 { 00:12:36.165 "name": "BaseBdev2", 00:12:36.165 "uuid": "70351ef2-5c06-4d83-8e05-0589c6533436", 00:12:36.165 "is_configured": true, 00:12:36.165 "data_offset": 0, 00:12:36.165 "data_size": 65536 00:12:36.165 }, 00:12:36.165 { 00:12:36.165 "name": "BaseBdev3", 00:12:36.165 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:36.165 "is_configured": false, 00:12:36.165 "data_offset": 0, 00:12:36.165 "data_size": 0 00:12:36.165 } 00:12:36.165 ] 00:12:36.165 }' 00:12:36.165 23:46:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:36.165 23:46:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:12:36.424 23:46:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:12:36.424 23:46:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:36.424 23:46:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:12:36.684 [2024-11-26 23:46:24.559225] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:12:36.684 [2024-11-26 23:46:24.559447] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:12:36.684 [2024-11-26 23:46:24.559490] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 131072, blocklen 512 00:12:36.684 [2024-11-26 23:46:24.560473] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:12:36.684 [2024-11-26 23:46:24.562100] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:12:36.684 [2024-11-26 23:46:24.562142] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:12:36.684 [2024-11-26 23:46:24.562878] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:36.684 BaseBdev3 00:12:36.684 23:46:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:36.684 23:46:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:12:36.684 23:46:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:12:36.684 23:46:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:12:36.684 23:46:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:12:36.684 23:46:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:12:36.684 23:46:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:12:36.684 23:46:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:12:36.684 23:46:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:36.684 23:46:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:12:36.684 23:46:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:36.684 23:46:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:12:36.684 23:46:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:36.684 23:46:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:12:36.684 [ 00:12:36.684 { 00:12:36.684 "name": "BaseBdev3", 00:12:36.684 "aliases": [ 00:12:36.684 "89387ccb-1526-409b-924b-f889fb0721c8" 00:12:36.684 ], 00:12:36.684 "product_name": "Malloc disk", 00:12:36.684 "block_size": 512, 00:12:36.684 "num_blocks": 65536, 00:12:36.684 "uuid": "89387ccb-1526-409b-924b-f889fb0721c8", 00:12:36.684 "assigned_rate_limits": { 00:12:36.684 "rw_ios_per_sec": 0, 00:12:36.684 "rw_mbytes_per_sec": 0, 00:12:36.684 "r_mbytes_per_sec": 0, 00:12:36.684 "w_mbytes_per_sec": 0 00:12:36.684 }, 00:12:36.684 "claimed": true, 00:12:36.684 "claim_type": "exclusive_write", 00:12:36.684 "zoned": false, 00:12:36.684 "supported_io_types": { 00:12:36.684 "read": true, 00:12:36.684 "write": true, 00:12:36.684 "unmap": true, 00:12:36.684 "flush": true, 00:12:36.684 "reset": true, 00:12:36.684 "nvme_admin": false, 00:12:36.684 "nvme_io": false, 00:12:36.684 "nvme_io_md": false, 00:12:36.684 "write_zeroes": true, 00:12:36.684 "zcopy": true, 00:12:36.684 "get_zone_info": false, 00:12:36.684 "zone_management": false, 00:12:36.684 "zone_append": false, 00:12:36.684 "compare": false, 00:12:36.684 "compare_and_write": false, 00:12:36.684 "abort": true, 00:12:36.684 "seek_hole": false, 00:12:36.684 "seek_data": false, 00:12:36.684 "copy": true, 00:12:36.684 "nvme_iov_md": false 00:12:36.684 }, 00:12:36.685 "memory_domains": [ 00:12:36.685 { 00:12:36.685 "dma_device_id": "system", 00:12:36.685 "dma_device_type": 1 00:12:36.685 }, 00:12:36.685 { 00:12:36.685 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:12:36.685 "dma_device_type": 2 00:12:36.685 } 00:12:36.685 ], 00:12:36.685 "driver_specific": {} 00:12:36.685 } 00:12:36.685 ] 00:12:36.685 23:46:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:36.685 23:46:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:12:36.685 23:46:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:12:36.685 23:46:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:12:36.685 23:46:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 3 00:12:36.685 23:46:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:12:36.685 23:46:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:36.685 23:46:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:12:36.685 23:46:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:12:36.685 23:46:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:36.685 23:46:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:36.685 23:46:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:36.685 23:46:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:36.685 23:46:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:36.685 23:46:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:36.685 23:46:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:12:36.685 23:46:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:36.685 23:46:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:12:36.685 23:46:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:36.685 23:46:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:36.685 "name": "Existed_Raid", 00:12:36.685 "uuid": "a9368d2d-9c0f-42f6-a0df-71a8327750c7", 00:12:36.685 "strip_size_kb": 64, 00:12:36.685 "state": "online", 00:12:36.685 "raid_level": "raid5f", 00:12:36.685 "superblock": false, 00:12:36.685 "num_base_bdevs": 3, 00:12:36.685 "num_base_bdevs_discovered": 3, 00:12:36.685 "num_base_bdevs_operational": 3, 00:12:36.685 "base_bdevs_list": [ 00:12:36.685 { 00:12:36.685 "name": "BaseBdev1", 00:12:36.685 "uuid": "b159283f-3797-4f0e-b5d5-044d342ede02", 00:12:36.685 "is_configured": true, 00:12:36.685 "data_offset": 0, 00:12:36.685 "data_size": 65536 00:12:36.685 }, 00:12:36.685 { 00:12:36.685 "name": "BaseBdev2", 00:12:36.685 "uuid": "70351ef2-5c06-4d83-8e05-0589c6533436", 00:12:36.685 "is_configured": true, 00:12:36.685 "data_offset": 0, 00:12:36.685 "data_size": 65536 00:12:36.685 }, 00:12:36.685 { 00:12:36.685 "name": "BaseBdev3", 00:12:36.685 "uuid": "89387ccb-1526-409b-924b-f889fb0721c8", 00:12:36.685 "is_configured": true, 00:12:36.685 "data_offset": 0, 00:12:36.685 "data_size": 65536 00:12:36.685 } 00:12:36.685 ] 00:12:36.685 }' 00:12:36.685 23:46:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:36.685 23:46:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:12:36.945 23:46:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:12:36.945 23:46:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:12:36.945 23:46:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:12:36.945 23:46:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:12:36.945 23:46:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:12:36.945 23:46:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:12:36.945 23:46:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:12:36.945 23:46:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:12:36.945 23:46:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:36.945 23:46:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:12:36.945 [2024-11-26 23:46:25.058631] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:12:37.206 23:46:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:37.206 23:46:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:12:37.206 "name": "Existed_Raid", 00:12:37.206 "aliases": [ 00:12:37.206 "a9368d2d-9c0f-42f6-a0df-71a8327750c7" 00:12:37.206 ], 00:12:37.206 "product_name": "Raid Volume", 00:12:37.206 "block_size": 512, 00:12:37.206 "num_blocks": 131072, 00:12:37.206 "uuid": "a9368d2d-9c0f-42f6-a0df-71a8327750c7", 00:12:37.206 "assigned_rate_limits": { 00:12:37.206 "rw_ios_per_sec": 0, 00:12:37.206 "rw_mbytes_per_sec": 0, 00:12:37.206 "r_mbytes_per_sec": 0, 00:12:37.206 "w_mbytes_per_sec": 0 00:12:37.206 }, 00:12:37.206 "claimed": false, 00:12:37.206 "zoned": false, 00:12:37.206 "supported_io_types": { 00:12:37.206 "read": true, 00:12:37.206 "write": true, 00:12:37.206 "unmap": false, 00:12:37.206 "flush": false, 00:12:37.206 "reset": true, 00:12:37.206 "nvme_admin": false, 00:12:37.206 "nvme_io": false, 00:12:37.206 "nvme_io_md": false, 00:12:37.206 "write_zeroes": true, 00:12:37.206 "zcopy": false, 00:12:37.206 "get_zone_info": false, 00:12:37.206 "zone_management": false, 00:12:37.206 "zone_append": false, 00:12:37.206 "compare": false, 00:12:37.206 "compare_and_write": false, 00:12:37.206 "abort": false, 00:12:37.206 "seek_hole": false, 00:12:37.206 "seek_data": false, 00:12:37.206 "copy": false, 00:12:37.206 "nvme_iov_md": false 00:12:37.206 }, 00:12:37.206 "driver_specific": { 00:12:37.206 "raid": { 00:12:37.206 "uuid": "a9368d2d-9c0f-42f6-a0df-71a8327750c7", 00:12:37.206 "strip_size_kb": 64, 00:12:37.206 "state": "online", 00:12:37.206 "raid_level": "raid5f", 00:12:37.206 "superblock": false, 00:12:37.206 "num_base_bdevs": 3, 00:12:37.206 "num_base_bdevs_discovered": 3, 00:12:37.206 "num_base_bdevs_operational": 3, 00:12:37.206 "base_bdevs_list": [ 00:12:37.206 { 00:12:37.206 "name": "BaseBdev1", 00:12:37.206 "uuid": "b159283f-3797-4f0e-b5d5-044d342ede02", 00:12:37.206 "is_configured": true, 00:12:37.206 "data_offset": 0, 00:12:37.206 "data_size": 65536 00:12:37.206 }, 00:12:37.206 { 00:12:37.206 "name": "BaseBdev2", 00:12:37.206 "uuid": "70351ef2-5c06-4d83-8e05-0589c6533436", 00:12:37.206 "is_configured": true, 00:12:37.206 "data_offset": 0, 00:12:37.206 "data_size": 65536 00:12:37.206 }, 00:12:37.206 { 00:12:37.206 "name": "BaseBdev3", 00:12:37.206 "uuid": "89387ccb-1526-409b-924b-f889fb0721c8", 00:12:37.206 "is_configured": true, 00:12:37.206 "data_offset": 0, 00:12:37.206 "data_size": 65536 00:12:37.206 } 00:12:37.206 ] 00:12:37.206 } 00:12:37.206 } 00:12:37.206 }' 00:12:37.206 23:46:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:12:37.206 23:46:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:12:37.206 BaseBdev2 00:12:37.206 BaseBdev3' 00:12:37.206 23:46:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:12:37.206 23:46:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:12:37.206 23:46:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:12:37.206 23:46:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:12:37.206 23:46:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:12:37.206 23:46:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:37.206 23:46:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:12:37.206 23:46:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:37.207 23:46:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:12:37.207 23:46:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:12:37.207 23:46:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:12:37.207 23:46:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:12:37.207 23:46:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:12:37.207 23:46:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:37.207 23:46:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:12:37.207 23:46:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:37.207 23:46:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:12:37.207 23:46:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:12:37.207 23:46:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:12:37.207 23:46:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:12:37.207 23:46:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:12:37.207 23:46:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:37.207 23:46:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:12:37.207 23:46:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:37.207 23:46:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:12:37.207 23:46:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:12:37.207 23:46:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:12:37.207 23:46:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:37.207 23:46:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:12:37.207 [2024-11-26 23:46:25.310092] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:12:37.207 23:46:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:37.207 23:46:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:12:37.207 23:46:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy raid5f 00:12:37.207 23:46:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:12:37.207 23:46:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@199 -- # return 0 00:12:37.207 23:46:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:12:37.207 23:46:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 2 00:12:37.207 23:46:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:12:37.207 23:46:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:37.207 23:46:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:12:37.207 23:46:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:12:37.207 23:46:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:12:37.207 23:46:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:37.207 23:46:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:37.207 23:46:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:37.207 23:46:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:37.207 23:46:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:37.207 23:46:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:37.207 23:46:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:12:37.207 23:46:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:12:37.468 23:46:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:37.468 23:46:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:37.468 "name": "Existed_Raid", 00:12:37.468 "uuid": "a9368d2d-9c0f-42f6-a0df-71a8327750c7", 00:12:37.468 "strip_size_kb": 64, 00:12:37.468 "state": "online", 00:12:37.468 "raid_level": "raid5f", 00:12:37.468 "superblock": false, 00:12:37.468 "num_base_bdevs": 3, 00:12:37.468 "num_base_bdevs_discovered": 2, 00:12:37.468 "num_base_bdevs_operational": 2, 00:12:37.468 "base_bdevs_list": [ 00:12:37.468 { 00:12:37.468 "name": null, 00:12:37.468 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:37.468 "is_configured": false, 00:12:37.468 "data_offset": 0, 00:12:37.468 "data_size": 65536 00:12:37.468 }, 00:12:37.468 { 00:12:37.468 "name": "BaseBdev2", 00:12:37.468 "uuid": "70351ef2-5c06-4d83-8e05-0589c6533436", 00:12:37.468 "is_configured": true, 00:12:37.468 "data_offset": 0, 00:12:37.468 "data_size": 65536 00:12:37.468 }, 00:12:37.468 { 00:12:37.468 "name": "BaseBdev3", 00:12:37.468 "uuid": "89387ccb-1526-409b-924b-f889fb0721c8", 00:12:37.468 "is_configured": true, 00:12:37.468 "data_offset": 0, 00:12:37.468 "data_size": 65536 00:12:37.468 } 00:12:37.468 ] 00:12:37.468 }' 00:12:37.468 23:46:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:37.468 23:46:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:12:37.728 23:46:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:12:37.728 23:46:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:12:37.728 23:46:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:12:37.728 23:46:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:37.728 23:46:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:37.728 23:46:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:12:37.728 23:46:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:37.728 23:46:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:12:37.728 23:46:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:12:37.728 23:46:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:12:37.729 23:46:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:37.729 23:46:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:12:37.729 [2024-11-26 23:46:25.836309] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:12:37.729 [2024-11-26 23:46:25.836409] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:12:37.729 [2024-11-26 23:46:25.847404] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:12:37.729 23:46:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:37.729 23:46:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:12:37.729 23:46:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:12:37.729 23:46:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:37.729 23:46:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:37.729 23:46:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:12:37.729 23:46:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:12:37.989 23:46:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:37.989 23:46:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:12:37.989 23:46:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:12:37.989 23:46:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:12:37.989 23:46:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:37.989 23:46:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:12:37.989 [2024-11-26 23:46:25.907305] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:12:37.989 [2024-11-26 23:46:25.907405] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:12:37.989 23:46:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:37.989 23:46:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:12:37.989 23:46:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:12:37.989 23:46:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:37.989 23:46:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:37.989 23:46:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:12:37.989 23:46:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:12:37.989 23:46:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:37.989 23:46:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:12:37.989 23:46:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:12:37.989 23:46:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 3 -gt 2 ']' 00:12:37.989 23:46:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:12:37.989 23:46:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:12:37.989 23:46:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:12:37.989 23:46:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:37.989 23:46:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:12:37.989 BaseBdev2 00:12:37.989 23:46:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:37.989 23:46:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:12:37.989 23:46:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:12:37.989 23:46:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:12:37.989 23:46:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:12:37.989 23:46:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:12:37.989 23:46:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:12:37.989 23:46:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:12:37.989 23:46:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:37.989 23:46:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:12:37.989 23:46:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:37.989 23:46:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:12:37.989 23:46:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:37.989 23:46:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:12:37.989 [ 00:12:37.989 { 00:12:37.989 "name": "BaseBdev2", 00:12:37.989 "aliases": [ 00:12:37.989 "3b8f22b1-2ab9-44ba-a5cf-4c8637bb8b91" 00:12:37.989 ], 00:12:37.989 "product_name": "Malloc disk", 00:12:37.989 "block_size": 512, 00:12:37.989 "num_blocks": 65536, 00:12:37.989 "uuid": "3b8f22b1-2ab9-44ba-a5cf-4c8637bb8b91", 00:12:37.989 "assigned_rate_limits": { 00:12:37.989 "rw_ios_per_sec": 0, 00:12:37.989 "rw_mbytes_per_sec": 0, 00:12:37.989 "r_mbytes_per_sec": 0, 00:12:37.989 "w_mbytes_per_sec": 0 00:12:37.989 }, 00:12:37.989 "claimed": false, 00:12:37.989 "zoned": false, 00:12:37.989 "supported_io_types": { 00:12:37.989 "read": true, 00:12:37.989 "write": true, 00:12:37.989 "unmap": true, 00:12:37.989 "flush": true, 00:12:37.989 "reset": true, 00:12:37.989 "nvme_admin": false, 00:12:37.989 "nvme_io": false, 00:12:37.989 "nvme_io_md": false, 00:12:37.990 "write_zeroes": true, 00:12:37.990 "zcopy": true, 00:12:37.990 "get_zone_info": false, 00:12:37.990 "zone_management": false, 00:12:37.990 "zone_append": false, 00:12:37.990 "compare": false, 00:12:37.990 "compare_and_write": false, 00:12:37.990 "abort": true, 00:12:37.990 "seek_hole": false, 00:12:37.990 "seek_data": false, 00:12:37.990 "copy": true, 00:12:37.990 "nvme_iov_md": false 00:12:37.990 }, 00:12:37.990 "memory_domains": [ 00:12:37.990 { 00:12:37.990 "dma_device_id": "system", 00:12:37.990 "dma_device_type": 1 00:12:37.990 }, 00:12:37.990 { 00:12:37.990 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:12:37.990 "dma_device_type": 2 00:12:37.990 } 00:12:37.990 ], 00:12:37.990 "driver_specific": {} 00:12:37.990 } 00:12:37.990 ] 00:12:37.990 23:46:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:37.990 23:46:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:12:37.990 23:46:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:12:37.990 23:46:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:12:37.990 23:46:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:12:37.990 23:46:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:37.990 23:46:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:12:37.990 BaseBdev3 00:12:37.990 23:46:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:37.990 23:46:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:12:37.990 23:46:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:12:37.990 23:46:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:12:37.990 23:46:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:12:37.990 23:46:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:12:37.990 23:46:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:12:37.990 23:46:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:12:37.990 23:46:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:37.990 23:46:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:12:37.990 23:46:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:37.990 23:46:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:12:37.990 23:46:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:37.990 23:46:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:12:37.990 [ 00:12:37.990 { 00:12:37.990 "name": "BaseBdev3", 00:12:37.990 "aliases": [ 00:12:37.990 "c6122897-8df6-4a6a-8d82-eaffb94234d2" 00:12:37.990 ], 00:12:37.990 "product_name": "Malloc disk", 00:12:37.990 "block_size": 512, 00:12:37.990 "num_blocks": 65536, 00:12:37.990 "uuid": "c6122897-8df6-4a6a-8d82-eaffb94234d2", 00:12:37.990 "assigned_rate_limits": { 00:12:37.990 "rw_ios_per_sec": 0, 00:12:37.990 "rw_mbytes_per_sec": 0, 00:12:37.990 "r_mbytes_per_sec": 0, 00:12:37.990 "w_mbytes_per_sec": 0 00:12:37.990 }, 00:12:37.990 "claimed": false, 00:12:37.990 "zoned": false, 00:12:37.990 "supported_io_types": { 00:12:37.990 "read": true, 00:12:37.990 "write": true, 00:12:37.990 "unmap": true, 00:12:37.990 "flush": true, 00:12:37.990 "reset": true, 00:12:37.990 "nvme_admin": false, 00:12:37.990 "nvme_io": false, 00:12:37.990 "nvme_io_md": false, 00:12:37.990 "write_zeroes": true, 00:12:37.990 "zcopy": true, 00:12:37.990 "get_zone_info": false, 00:12:37.990 "zone_management": false, 00:12:37.990 "zone_append": false, 00:12:37.990 "compare": false, 00:12:37.990 "compare_and_write": false, 00:12:37.990 "abort": true, 00:12:37.990 "seek_hole": false, 00:12:37.990 "seek_data": false, 00:12:37.990 "copy": true, 00:12:37.990 "nvme_iov_md": false 00:12:37.990 }, 00:12:37.990 "memory_domains": [ 00:12:37.990 { 00:12:37.990 "dma_device_id": "system", 00:12:37.990 "dma_device_type": 1 00:12:37.990 }, 00:12:37.990 { 00:12:37.990 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:12:37.990 "dma_device_type": 2 00:12:37.990 } 00:12:37.990 ], 00:12:37.990 "driver_specific": {} 00:12:37.990 } 00:12:37.990 ] 00:12:37.990 23:46:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:37.990 23:46:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:12:37.990 23:46:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:12:37.990 23:46:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:12:37.990 23:46:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:12:37.990 23:46:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:37.990 23:46:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:12:37.990 [2024-11-26 23:46:26.082160] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:12:37.990 [2024-11-26 23:46:26.082202] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:12:37.990 [2024-11-26 23:46:26.082237] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:12:37.990 [2024-11-26 23:46:26.084041] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:12:37.990 23:46:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:37.990 23:46:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:12:37.990 23:46:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:12:37.990 23:46:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:12:37.990 23:46:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:12:37.990 23:46:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:12:37.990 23:46:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:37.990 23:46:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:37.990 23:46:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:37.990 23:46:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:37.990 23:46:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:37.990 23:46:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:37.990 23:46:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:12:37.990 23:46:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:37.990 23:46:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:12:37.990 23:46:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:38.249 23:46:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:38.249 "name": "Existed_Raid", 00:12:38.249 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:38.249 "strip_size_kb": 64, 00:12:38.249 "state": "configuring", 00:12:38.249 "raid_level": "raid5f", 00:12:38.249 "superblock": false, 00:12:38.249 "num_base_bdevs": 3, 00:12:38.249 "num_base_bdevs_discovered": 2, 00:12:38.249 "num_base_bdevs_operational": 3, 00:12:38.249 "base_bdevs_list": [ 00:12:38.249 { 00:12:38.249 "name": "BaseBdev1", 00:12:38.249 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:38.249 "is_configured": false, 00:12:38.249 "data_offset": 0, 00:12:38.249 "data_size": 0 00:12:38.249 }, 00:12:38.249 { 00:12:38.249 "name": "BaseBdev2", 00:12:38.249 "uuid": "3b8f22b1-2ab9-44ba-a5cf-4c8637bb8b91", 00:12:38.249 "is_configured": true, 00:12:38.249 "data_offset": 0, 00:12:38.249 "data_size": 65536 00:12:38.249 }, 00:12:38.249 { 00:12:38.249 "name": "BaseBdev3", 00:12:38.249 "uuid": "c6122897-8df6-4a6a-8d82-eaffb94234d2", 00:12:38.249 "is_configured": true, 00:12:38.249 "data_offset": 0, 00:12:38.249 "data_size": 65536 00:12:38.249 } 00:12:38.249 ] 00:12:38.249 }' 00:12:38.249 23:46:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:38.249 23:46:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:12:38.509 23:46:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:12:38.509 23:46:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:38.509 23:46:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:12:38.509 [2024-11-26 23:46:26.581319] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:12:38.509 23:46:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:38.509 23:46:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:12:38.509 23:46:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:12:38.509 23:46:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:12:38.509 23:46:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:12:38.509 23:46:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:12:38.509 23:46:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:38.509 23:46:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:38.509 23:46:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:38.509 23:46:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:38.509 23:46:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:38.509 23:46:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:12:38.509 23:46:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:38.509 23:46:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:38.509 23:46:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:12:38.509 23:46:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:38.509 23:46:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:38.509 "name": "Existed_Raid", 00:12:38.509 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:38.509 "strip_size_kb": 64, 00:12:38.509 "state": "configuring", 00:12:38.509 "raid_level": "raid5f", 00:12:38.509 "superblock": false, 00:12:38.509 "num_base_bdevs": 3, 00:12:38.509 "num_base_bdevs_discovered": 1, 00:12:38.509 "num_base_bdevs_operational": 3, 00:12:38.509 "base_bdevs_list": [ 00:12:38.509 { 00:12:38.509 "name": "BaseBdev1", 00:12:38.509 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:38.509 "is_configured": false, 00:12:38.509 "data_offset": 0, 00:12:38.509 "data_size": 0 00:12:38.509 }, 00:12:38.509 { 00:12:38.509 "name": null, 00:12:38.509 "uuid": "3b8f22b1-2ab9-44ba-a5cf-4c8637bb8b91", 00:12:38.509 "is_configured": false, 00:12:38.509 "data_offset": 0, 00:12:38.509 "data_size": 65536 00:12:38.509 }, 00:12:38.509 { 00:12:38.509 "name": "BaseBdev3", 00:12:38.509 "uuid": "c6122897-8df6-4a6a-8d82-eaffb94234d2", 00:12:38.509 "is_configured": true, 00:12:38.509 "data_offset": 0, 00:12:38.509 "data_size": 65536 00:12:38.509 } 00:12:38.509 ] 00:12:38.509 }' 00:12:38.509 23:46:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:38.509 23:46:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:12:39.081 23:46:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:39.081 23:46:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:39.081 23:46:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:12:39.081 23:46:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:12:39.081 23:46:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:39.081 23:46:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:12:39.081 23:46:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:12:39.081 23:46:27 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:39.081 23:46:27 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:12:39.081 [2024-11-26 23:46:27.035433] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:12:39.081 BaseBdev1 00:12:39.081 23:46:27 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:39.081 23:46:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:12:39.081 23:46:27 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:12:39.081 23:46:27 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:12:39.081 23:46:27 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:12:39.081 23:46:27 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:12:39.081 23:46:27 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:12:39.081 23:46:27 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:12:39.081 23:46:27 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:39.081 23:46:27 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:12:39.081 23:46:27 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:39.081 23:46:27 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:12:39.081 23:46:27 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:39.081 23:46:27 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:12:39.081 [ 00:12:39.081 { 00:12:39.081 "name": "BaseBdev1", 00:12:39.081 "aliases": [ 00:12:39.081 "8d9b68a0-e654-497d-aa87-dc33764743ef" 00:12:39.081 ], 00:12:39.081 "product_name": "Malloc disk", 00:12:39.081 "block_size": 512, 00:12:39.081 "num_blocks": 65536, 00:12:39.081 "uuid": "8d9b68a0-e654-497d-aa87-dc33764743ef", 00:12:39.081 "assigned_rate_limits": { 00:12:39.081 "rw_ios_per_sec": 0, 00:12:39.081 "rw_mbytes_per_sec": 0, 00:12:39.081 "r_mbytes_per_sec": 0, 00:12:39.081 "w_mbytes_per_sec": 0 00:12:39.081 }, 00:12:39.081 "claimed": true, 00:12:39.081 "claim_type": "exclusive_write", 00:12:39.081 "zoned": false, 00:12:39.081 "supported_io_types": { 00:12:39.081 "read": true, 00:12:39.081 "write": true, 00:12:39.081 "unmap": true, 00:12:39.081 "flush": true, 00:12:39.081 "reset": true, 00:12:39.081 "nvme_admin": false, 00:12:39.081 "nvme_io": false, 00:12:39.081 "nvme_io_md": false, 00:12:39.081 "write_zeroes": true, 00:12:39.081 "zcopy": true, 00:12:39.081 "get_zone_info": false, 00:12:39.081 "zone_management": false, 00:12:39.081 "zone_append": false, 00:12:39.081 "compare": false, 00:12:39.081 "compare_and_write": false, 00:12:39.081 "abort": true, 00:12:39.081 "seek_hole": false, 00:12:39.081 "seek_data": false, 00:12:39.081 "copy": true, 00:12:39.081 "nvme_iov_md": false 00:12:39.081 }, 00:12:39.081 "memory_domains": [ 00:12:39.081 { 00:12:39.081 "dma_device_id": "system", 00:12:39.081 "dma_device_type": 1 00:12:39.081 }, 00:12:39.081 { 00:12:39.081 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:12:39.081 "dma_device_type": 2 00:12:39.081 } 00:12:39.081 ], 00:12:39.081 "driver_specific": {} 00:12:39.081 } 00:12:39.081 ] 00:12:39.081 23:46:27 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:39.081 23:46:27 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:12:39.081 23:46:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:12:39.081 23:46:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:12:39.081 23:46:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:12:39.081 23:46:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:12:39.081 23:46:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:12:39.081 23:46:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:39.081 23:46:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:39.081 23:46:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:39.081 23:46:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:39.081 23:46:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:39.081 23:46:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:39.081 23:46:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:12:39.081 23:46:27 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:39.081 23:46:27 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:12:39.081 23:46:27 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:39.081 23:46:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:39.081 "name": "Existed_Raid", 00:12:39.081 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:39.081 "strip_size_kb": 64, 00:12:39.081 "state": "configuring", 00:12:39.081 "raid_level": "raid5f", 00:12:39.081 "superblock": false, 00:12:39.081 "num_base_bdevs": 3, 00:12:39.081 "num_base_bdevs_discovered": 2, 00:12:39.081 "num_base_bdevs_operational": 3, 00:12:39.081 "base_bdevs_list": [ 00:12:39.081 { 00:12:39.081 "name": "BaseBdev1", 00:12:39.081 "uuid": "8d9b68a0-e654-497d-aa87-dc33764743ef", 00:12:39.081 "is_configured": true, 00:12:39.081 "data_offset": 0, 00:12:39.081 "data_size": 65536 00:12:39.081 }, 00:12:39.081 { 00:12:39.081 "name": null, 00:12:39.081 "uuid": "3b8f22b1-2ab9-44ba-a5cf-4c8637bb8b91", 00:12:39.081 "is_configured": false, 00:12:39.081 "data_offset": 0, 00:12:39.081 "data_size": 65536 00:12:39.081 }, 00:12:39.081 { 00:12:39.081 "name": "BaseBdev3", 00:12:39.081 "uuid": "c6122897-8df6-4a6a-8d82-eaffb94234d2", 00:12:39.081 "is_configured": true, 00:12:39.081 "data_offset": 0, 00:12:39.081 "data_size": 65536 00:12:39.081 } 00:12:39.081 ] 00:12:39.081 }' 00:12:39.081 23:46:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:39.081 23:46:27 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:12:39.651 23:46:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:39.651 23:46:27 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:39.651 23:46:27 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:12:39.651 23:46:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:12:39.651 23:46:27 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:39.651 23:46:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:12:39.651 23:46:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:12:39.651 23:46:27 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:39.651 23:46:27 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:12:39.651 [2024-11-26 23:46:27.566574] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:12:39.651 23:46:27 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:39.651 23:46:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:12:39.651 23:46:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:12:39.651 23:46:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:12:39.651 23:46:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:12:39.651 23:46:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:12:39.651 23:46:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:39.651 23:46:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:39.651 23:46:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:39.651 23:46:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:39.651 23:46:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:39.651 23:46:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:39.651 23:46:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:12:39.651 23:46:27 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:39.651 23:46:27 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:12:39.651 23:46:27 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:39.651 23:46:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:39.651 "name": "Existed_Raid", 00:12:39.651 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:39.651 "strip_size_kb": 64, 00:12:39.651 "state": "configuring", 00:12:39.651 "raid_level": "raid5f", 00:12:39.651 "superblock": false, 00:12:39.651 "num_base_bdevs": 3, 00:12:39.651 "num_base_bdevs_discovered": 1, 00:12:39.652 "num_base_bdevs_operational": 3, 00:12:39.652 "base_bdevs_list": [ 00:12:39.652 { 00:12:39.652 "name": "BaseBdev1", 00:12:39.652 "uuid": "8d9b68a0-e654-497d-aa87-dc33764743ef", 00:12:39.652 "is_configured": true, 00:12:39.652 "data_offset": 0, 00:12:39.652 "data_size": 65536 00:12:39.652 }, 00:12:39.652 { 00:12:39.652 "name": null, 00:12:39.652 "uuid": "3b8f22b1-2ab9-44ba-a5cf-4c8637bb8b91", 00:12:39.652 "is_configured": false, 00:12:39.652 "data_offset": 0, 00:12:39.652 "data_size": 65536 00:12:39.652 }, 00:12:39.652 { 00:12:39.652 "name": null, 00:12:39.652 "uuid": "c6122897-8df6-4a6a-8d82-eaffb94234d2", 00:12:39.652 "is_configured": false, 00:12:39.652 "data_offset": 0, 00:12:39.652 "data_size": 65536 00:12:39.652 } 00:12:39.652 ] 00:12:39.652 }' 00:12:39.652 23:46:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:39.652 23:46:27 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:12:39.912 23:46:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:12:39.912 23:46:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:39.912 23:46:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:39.912 23:46:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:12:39.912 23:46:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:40.173 23:46:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:12:40.173 23:46:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:12:40.173 23:46:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:40.173 23:46:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:12:40.173 [2024-11-26 23:46:28.069780] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:12:40.173 23:46:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:40.173 23:46:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:12:40.173 23:46:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:12:40.173 23:46:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:12:40.173 23:46:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:12:40.173 23:46:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:12:40.173 23:46:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:40.173 23:46:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:40.173 23:46:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:40.173 23:46:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:40.173 23:46:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:40.173 23:46:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:12:40.173 23:46:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:40.173 23:46:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:40.173 23:46:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:12:40.173 23:46:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:40.173 23:46:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:40.173 "name": "Existed_Raid", 00:12:40.173 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:40.173 "strip_size_kb": 64, 00:12:40.173 "state": "configuring", 00:12:40.173 "raid_level": "raid5f", 00:12:40.173 "superblock": false, 00:12:40.173 "num_base_bdevs": 3, 00:12:40.173 "num_base_bdevs_discovered": 2, 00:12:40.173 "num_base_bdevs_operational": 3, 00:12:40.173 "base_bdevs_list": [ 00:12:40.173 { 00:12:40.173 "name": "BaseBdev1", 00:12:40.173 "uuid": "8d9b68a0-e654-497d-aa87-dc33764743ef", 00:12:40.173 "is_configured": true, 00:12:40.173 "data_offset": 0, 00:12:40.173 "data_size": 65536 00:12:40.173 }, 00:12:40.173 { 00:12:40.173 "name": null, 00:12:40.173 "uuid": "3b8f22b1-2ab9-44ba-a5cf-4c8637bb8b91", 00:12:40.173 "is_configured": false, 00:12:40.173 "data_offset": 0, 00:12:40.173 "data_size": 65536 00:12:40.173 }, 00:12:40.173 { 00:12:40.173 "name": "BaseBdev3", 00:12:40.173 "uuid": "c6122897-8df6-4a6a-8d82-eaffb94234d2", 00:12:40.173 "is_configured": true, 00:12:40.173 "data_offset": 0, 00:12:40.173 "data_size": 65536 00:12:40.173 } 00:12:40.173 ] 00:12:40.173 }' 00:12:40.173 23:46:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:40.173 23:46:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:12:40.434 23:46:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:40.434 23:46:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:12:40.434 23:46:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:40.434 23:46:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:12:40.434 23:46:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:40.434 23:46:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:12:40.434 23:46:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:12:40.434 23:46:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:40.434 23:46:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:12:40.434 [2024-11-26 23:46:28.560972] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:12:40.695 23:46:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:40.695 23:46:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:12:40.695 23:46:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:12:40.695 23:46:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:12:40.695 23:46:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:12:40.695 23:46:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:12:40.695 23:46:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:40.695 23:46:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:40.695 23:46:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:40.695 23:46:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:40.695 23:46:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:40.695 23:46:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:40.695 23:46:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:12:40.695 23:46:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:40.695 23:46:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:12:40.695 23:46:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:40.695 23:46:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:40.695 "name": "Existed_Raid", 00:12:40.695 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:40.695 "strip_size_kb": 64, 00:12:40.695 "state": "configuring", 00:12:40.695 "raid_level": "raid5f", 00:12:40.695 "superblock": false, 00:12:40.695 "num_base_bdevs": 3, 00:12:40.695 "num_base_bdevs_discovered": 1, 00:12:40.695 "num_base_bdevs_operational": 3, 00:12:40.695 "base_bdevs_list": [ 00:12:40.695 { 00:12:40.695 "name": null, 00:12:40.695 "uuid": "8d9b68a0-e654-497d-aa87-dc33764743ef", 00:12:40.695 "is_configured": false, 00:12:40.695 "data_offset": 0, 00:12:40.695 "data_size": 65536 00:12:40.695 }, 00:12:40.695 { 00:12:40.695 "name": null, 00:12:40.695 "uuid": "3b8f22b1-2ab9-44ba-a5cf-4c8637bb8b91", 00:12:40.695 "is_configured": false, 00:12:40.695 "data_offset": 0, 00:12:40.695 "data_size": 65536 00:12:40.695 }, 00:12:40.695 { 00:12:40.695 "name": "BaseBdev3", 00:12:40.695 "uuid": "c6122897-8df6-4a6a-8d82-eaffb94234d2", 00:12:40.695 "is_configured": true, 00:12:40.695 "data_offset": 0, 00:12:40.695 "data_size": 65536 00:12:40.695 } 00:12:40.695 ] 00:12:40.695 }' 00:12:40.695 23:46:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:40.695 23:46:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:12:40.956 23:46:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:12:40.956 23:46:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:40.956 23:46:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:40.956 23:46:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:12:40.956 23:46:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:40.956 23:46:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:12:40.956 23:46:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:12:40.956 23:46:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:40.956 23:46:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:12:40.956 [2024-11-26 23:46:29.034669] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:12:40.956 23:46:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:40.956 23:46:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:12:40.956 23:46:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:12:40.956 23:46:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:12:40.956 23:46:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:12:40.956 23:46:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:12:40.956 23:46:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:40.956 23:46:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:40.956 23:46:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:40.956 23:46:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:40.956 23:46:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:40.956 23:46:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:12:40.956 23:46:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:40.956 23:46:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:40.956 23:46:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:12:40.956 23:46:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:40.956 23:46:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:40.956 "name": "Existed_Raid", 00:12:40.956 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:40.956 "strip_size_kb": 64, 00:12:40.956 "state": "configuring", 00:12:40.956 "raid_level": "raid5f", 00:12:40.956 "superblock": false, 00:12:40.956 "num_base_bdevs": 3, 00:12:40.956 "num_base_bdevs_discovered": 2, 00:12:40.956 "num_base_bdevs_operational": 3, 00:12:40.956 "base_bdevs_list": [ 00:12:40.956 { 00:12:40.956 "name": null, 00:12:40.956 "uuid": "8d9b68a0-e654-497d-aa87-dc33764743ef", 00:12:40.956 "is_configured": false, 00:12:40.956 "data_offset": 0, 00:12:40.956 "data_size": 65536 00:12:40.956 }, 00:12:40.956 { 00:12:40.956 "name": "BaseBdev2", 00:12:40.956 "uuid": "3b8f22b1-2ab9-44ba-a5cf-4c8637bb8b91", 00:12:40.957 "is_configured": true, 00:12:40.957 "data_offset": 0, 00:12:40.957 "data_size": 65536 00:12:40.957 }, 00:12:40.957 { 00:12:40.957 "name": "BaseBdev3", 00:12:40.957 "uuid": "c6122897-8df6-4a6a-8d82-eaffb94234d2", 00:12:40.957 "is_configured": true, 00:12:40.957 "data_offset": 0, 00:12:40.957 "data_size": 65536 00:12:40.957 } 00:12:40.957 ] 00:12:40.957 }' 00:12:40.957 23:46:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:40.957 23:46:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:12:41.525 23:46:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:41.525 23:46:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:12:41.525 23:46:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:41.525 23:46:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:12:41.525 23:46:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:41.525 23:46:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:12:41.525 23:46:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:12:41.525 23:46:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:41.525 23:46:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:41.525 23:46:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:12:41.525 23:46:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:41.525 23:46:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 8d9b68a0-e654-497d-aa87-dc33764743ef 00:12:41.525 23:46:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:41.525 23:46:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:12:41.525 [2024-11-26 23:46:29.572974] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:12:41.525 [2024-11-26 23:46:29.573082] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:12:41.525 [2024-11-26 23:46:29.573110] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 131072, blocklen 512 00:12:41.525 [2024-11-26 23:46:29.573372] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002870 00:12:41.525 [2024-11-26 23:46:29.573822] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:12:41.525 [2024-11-26 23:46:29.573870] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001c80 00:12:41.525 [2024-11-26 23:46:29.574089] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:41.525 NewBaseBdev 00:12:41.525 23:46:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:41.525 23:46:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:12:41.525 23:46:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:12:41.525 23:46:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:12:41.525 23:46:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:12:41.525 23:46:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:12:41.525 23:46:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:12:41.525 23:46:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:12:41.525 23:46:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:41.525 23:46:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:12:41.525 23:46:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:41.525 23:46:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:12:41.525 23:46:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:41.525 23:46:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:12:41.525 [ 00:12:41.525 { 00:12:41.525 "name": "NewBaseBdev", 00:12:41.525 "aliases": [ 00:12:41.525 "8d9b68a0-e654-497d-aa87-dc33764743ef" 00:12:41.525 ], 00:12:41.525 "product_name": "Malloc disk", 00:12:41.525 "block_size": 512, 00:12:41.525 "num_blocks": 65536, 00:12:41.525 "uuid": "8d9b68a0-e654-497d-aa87-dc33764743ef", 00:12:41.525 "assigned_rate_limits": { 00:12:41.525 "rw_ios_per_sec": 0, 00:12:41.525 "rw_mbytes_per_sec": 0, 00:12:41.525 "r_mbytes_per_sec": 0, 00:12:41.525 "w_mbytes_per_sec": 0 00:12:41.525 }, 00:12:41.525 "claimed": true, 00:12:41.525 "claim_type": "exclusive_write", 00:12:41.525 "zoned": false, 00:12:41.525 "supported_io_types": { 00:12:41.525 "read": true, 00:12:41.525 "write": true, 00:12:41.525 "unmap": true, 00:12:41.525 "flush": true, 00:12:41.525 "reset": true, 00:12:41.525 "nvme_admin": false, 00:12:41.525 "nvme_io": false, 00:12:41.525 "nvme_io_md": false, 00:12:41.525 "write_zeroes": true, 00:12:41.525 "zcopy": true, 00:12:41.525 "get_zone_info": false, 00:12:41.525 "zone_management": false, 00:12:41.525 "zone_append": false, 00:12:41.525 "compare": false, 00:12:41.525 "compare_and_write": false, 00:12:41.525 "abort": true, 00:12:41.525 "seek_hole": false, 00:12:41.525 "seek_data": false, 00:12:41.525 "copy": true, 00:12:41.525 "nvme_iov_md": false 00:12:41.525 }, 00:12:41.525 "memory_domains": [ 00:12:41.525 { 00:12:41.525 "dma_device_id": "system", 00:12:41.525 "dma_device_type": 1 00:12:41.525 }, 00:12:41.525 { 00:12:41.525 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:12:41.525 "dma_device_type": 2 00:12:41.525 } 00:12:41.525 ], 00:12:41.525 "driver_specific": {} 00:12:41.525 } 00:12:41.525 ] 00:12:41.525 23:46:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:41.525 23:46:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:12:41.525 23:46:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 3 00:12:41.525 23:46:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:12:41.525 23:46:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:41.525 23:46:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:12:41.525 23:46:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:12:41.525 23:46:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:41.525 23:46:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:41.525 23:46:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:41.525 23:46:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:41.525 23:46:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:41.525 23:46:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:12:41.525 23:46:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:41.525 23:46:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:41.525 23:46:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:12:41.525 23:46:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:41.526 23:46:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:41.526 "name": "Existed_Raid", 00:12:41.526 "uuid": "00d3e5aa-ea8a-4680-bbbb-09848336a3b6", 00:12:41.526 "strip_size_kb": 64, 00:12:41.526 "state": "online", 00:12:41.526 "raid_level": "raid5f", 00:12:41.526 "superblock": false, 00:12:41.526 "num_base_bdevs": 3, 00:12:41.526 "num_base_bdevs_discovered": 3, 00:12:41.526 "num_base_bdevs_operational": 3, 00:12:41.526 "base_bdevs_list": [ 00:12:41.526 { 00:12:41.526 "name": "NewBaseBdev", 00:12:41.526 "uuid": "8d9b68a0-e654-497d-aa87-dc33764743ef", 00:12:41.526 "is_configured": true, 00:12:41.526 "data_offset": 0, 00:12:41.526 "data_size": 65536 00:12:41.526 }, 00:12:41.526 { 00:12:41.526 "name": "BaseBdev2", 00:12:41.526 "uuid": "3b8f22b1-2ab9-44ba-a5cf-4c8637bb8b91", 00:12:41.526 "is_configured": true, 00:12:41.526 "data_offset": 0, 00:12:41.526 "data_size": 65536 00:12:41.526 }, 00:12:41.526 { 00:12:41.526 "name": "BaseBdev3", 00:12:41.526 "uuid": "c6122897-8df6-4a6a-8d82-eaffb94234d2", 00:12:41.526 "is_configured": true, 00:12:41.526 "data_offset": 0, 00:12:41.526 "data_size": 65536 00:12:41.526 } 00:12:41.526 ] 00:12:41.526 }' 00:12:41.526 23:46:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:41.526 23:46:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:12:42.095 23:46:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:12:42.095 23:46:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:12:42.095 23:46:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:12:42.095 23:46:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:12:42.095 23:46:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:12:42.095 23:46:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:12:42.095 23:46:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:12:42.095 23:46:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:12:42.095 23:46:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:42.095 23:46:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:12:42.095 [2024-11-26 23:46:30.092314] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:12:42.095 23:46:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:42.095 23:46:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:12:42.095 "name": "Existed_Raid", 00:12:42.095 "aliases": [ 00:12:42.095 "00d3e5aa-ea8a-4680-bbbb-09848336a3b6" 00:12:42.095 ], 00:12:42.095 "product_name": "Raid Volume", 00:12:42.095 "block_size": 512, 00:12:42.095 "num_blocks": 131072, 00:12:42.095 "uuid": "00d3e5aa-ea8a-4680-bbbb-09848336a3b6", 00:12:42.095 "assigned_rate_limits": { 00:12:42.095 "rw_ios_per_sec": 0, 00:12:42.095 "rw_mbytes_per_sec": 0, 00:12:42.095 "r_mbytes_per_sec": 0, 00:12:42.095 "w_mbytes_per_sec": 0 00:12:42.095 }, 00:12:42.095 "claimed": false, 00:12:42.095 "zoned": false, 00:12:42.095 "supported_io_types": { 00:12:42.095 "read": true, 00:12:42.095 "write": true, 00:12:42.095 "unmap": false, 00:12:42.095 "flush": false, 00:12:42.095 "reset": true, 00:12:42.095 "nvme_admin": false, 00:12:42.095 "nvme_io": false, 00:12:42.095 "nvme_io_md": false, 00:12:42.095 "write_zeroes": true, 00:12:42.095 "zcopy": false, 00:12:42.095 "get_zone_info": false, 00:12:42.095 "zone_management": false, 00:12:42.095 "zone_append": false, 00:12:42.095 "compare": false, 00:12:42.095 "compare_and_write": false, 00:12:42.095 "abort": false, 00:12:42.095 "seek_hole": false, 00:12:42.095 "seek_data": false, 00:12:42.095 "copy": false, 00:12:42.095 "nvme_iov_md": false 00:12:42.095 }, 00:12:42.095 "driver_specific": { 00:12:42.095 "raid": { 00:12:42.095 "uuid": "00d3e5aa-ea8a-4680-bbbb-09848336a3b6", 00:12:42.095 "strip_size_kb": 64, 00:12:42.095 "state": "online", 00:12:42.095 "raid_level": "raid5f", 00:12:42.096 "superblock": false, 00:12:42.096 "num_base_bdevs": 3, 00:12:42.096 "num_base_bdevs_discovered": 3, 00:12:42.096 "num_base_bdevs_operational": 3, 00:12:42.096 "base_bdevs_list": [ 00:12:42.096 { 00:12:42.096 "name": "NewBaseBdev", 00:12:42.096 "uuid": "8d9b68a0-e654-497d-aa87-dc33764743ef", 00:12:42.096 "is_configured": true, 00:12:42.096 "data_offset": 0, 00:12:42.096 "data_size": 65536 00:12:42.096 }, 00:12:42.096 { 00:12:42.096 "name": "BaseBdev2", 00:12:42.096 "uuid": "3b8f22b1-2ab9-44ba-a5cf-4c8637bb8b91", 00:12:42.096 "is_configured": true, 00:12:42.096 "data_offset": 0, 00:12:42.096 "data_size": 65536 00:12:42.096 }, 00:12:42.096 { 00:12:42.096 "name": "BaseBdev3", 00:12:42.096 "uuid": "c6122897-8df6-4a6a-8d82-eaffb94234d2", 00:12:42.096 "is_configured": true, 00:12:42.096 "data_offset": 0, 00:12:42.096 "data_size": 65536 00:12:42.096 } 00:12:42.096 ] 00:12:42.096 } 00:12:42.096 } 00:12:42.096 }' 00:12:42.096 23:46:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:12:42.096 23:46:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:12:42.096 BaseBdev2 00:12:42.096 BaseBdev3' 00:12:42.096 23:46:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:12:42.096 23:46:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:12:42.096 23:46:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:12:42.096 23:46:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:12:42.096 23:46:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:12:42.096 23:46:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:42.096 23:46:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:12:42.356 23:46:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:42.356 23:46:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:12:42.356 23:46:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:12:42.356 23:46:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:12:42.356 23:46:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:12:42.356 23:46:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:12:42.356 23:46:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:42.356 23:46:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:12:42.356 23:46:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:42.356 23:46:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:12:42.356 23:46:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:12:42.356 23:46:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:12:42.356 23:46:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:12:42.356 23:46:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:12:42.356 23:46:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:42.356 23:46:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:12:42.356 23:46:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:42.356 23:46:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:12:42.356 23:46:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:12:42.356 23:46:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:12:42.356 23:46:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:42.356 23:46:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:12:42.356 [2024-11-26 23:46:30.359647] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:12:42.356 [2024-11-26 23:46:30.359670] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:12:42.356 [2024-11-26 23:46:30.359734] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:12:42.356 [2024-11-26 23:46:30.359961] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:12:42.356 [2024-11-26 23:46:30.359980] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name Existed_Raid, state offline 00:12:42.356 23:46:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:42.356 23:46:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 90134 00:12:42.356 23:46:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@954 -- # '[' -z 90134 ']' 00:12:42.356 23:46:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@958 -- # kill -0 90134 00:12:42.356 23:46:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@959 -- # uname 00:12:42.356 23:46:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:12:42.356 23:46:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 90134 00:12:42.356 23:46:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:12:42.356 23:46:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:12:42.356 23:46:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 90134' 00:12:42.356 killing process with pid 90134 00:12:42.356 23:46:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@973 -- # kill 90134 00:12:42.356 [2024-11-26 23:46:30.408040] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:12:42.356 23:46:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@978 -- # wait 90134 00:12:42.356 [2024-11-26 23:46:30.439326] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:12:42.617 23:46:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:12:42.617 ************************************ 00:12:42.617 END TEST raid5f_state_function_test 00:12:42.617 ************************************ 00:12:42.617 00:12:42.617 real 0m8.993s 00:12:42.617 user 0m15.473s 00:12:42.617 sys 0m1.775s 00:12:42.617 23:46:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:12:42.617 23:46:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:12:42.617 23:46:30 bdev_raid -- bdev/bdev_raid.sh@987 -- # run_test raid5f_state_function_test_sb raid_state_function_test raid5f 3 true 00:12:42.617 23:46:30 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:12:42.617 23:46:30 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:12:42.617 23:46:30 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:12:42.617 ************************************ 00:12:42.617 START TEST raid5f_state_function_test_sb 00:12:42.617 ************************************ 00:12:42.617 23:46:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@1129 -- # raid_state_function_test raid5f 3 true 00:12:42.617 23:46:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=raid5f 00:12:42.617 23:46:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=3 00:12:42.617 23:46:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:12:42.617 23:46:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:12:42.617 23:46:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:12:42.617 23:46:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:12:42.617 23:46:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:12:42.617 23:46:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:12:42.617 23:46:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:12:42.617 23:46:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:12:42.617 23:46:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:12:42.617 23:46:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:12:42.617 23:46:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:12:42.617 23:46:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:12:42.617 23:46:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:12:42.617 23:46:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:12:42.617 23:46:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:12:42.617 23:46:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:12:42.617 23:46:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:12:42.617 23:46:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:12:42.617 23:46:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:12:42.617 23:46:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' raid5f '!=' raid1 ']' 00:12:42.617 23:46:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:12:42.617 23:46:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:12:42.617 23:46:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:12:42.617 23:46:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:12:42.878 23:46:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=90741 00:12:42.878 23:46:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:12:42.879 23:46:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 90741' 00:12:42.879 Process raid pid: 90741 00:12:42.879 23:46:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 90741 00:12:42.879 23:46:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@835 -- # '[' -z 90741 ']' 00:12:42.879 23:46:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:12:42.879 23:46:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@840 -- # local max_retries=100 00:12:42.879 23:46:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:12:42.879 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:12:42.879 23:46:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@844 -- # xtrace_disable 00:12:42.879 23:46:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:42.879 [2024-11-26 23:46:30.819446] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:12:42.879 [2024-11-26 23:46:30.819662] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:12:42.879 [2024-11-26 23:46:30.973445] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:12:42.879 [2024-11-26 23:46:30.997627] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:12:43.138 [2024-11-26 23:46:31.040039] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:12:43.138 [2024-11-26 23:46:31.040065] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:12:43.707 23:46:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:12:43.707 23:46:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@868 -- # return 0 00:12:43.707 23:46:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:12:43.707 23:46:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:43.707 23:46:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:43.707 [2024-11-26 23:46:31.646514] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:12:43.707 [2024-11-26 23:46:31.646565] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:12:43.707 [2024-11-26 23:46:31.646591] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:12:43.707 [2024-11-26 23:46:31.646601] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:12:43.707 [2024-11-26 23:46:31.646607] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:12:43.707 [2024-11-26 23:46:31.646617] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:12:43.707 23:46:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:43.707 23:46:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:12:43.707 23:46:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:12:43.707 23:46:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:12:43.707 23:46:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:12:43.707 23:46:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:12:43.707 23:46:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:43.707 23:46:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:43.707 23:46:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:43.707 23:46:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:43.707 23:46:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:43.707 23:46:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:43.707 23:46:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:12:43.707 23:46:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:43.707 23:46:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:43.707 23:46:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:43.707 23:46:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:43.707 "name": "Existed_Raid", 00:12:43.707 "uuid": "67d874c5-f103-452f-b2c2-b77f37d455d6", 00:12:43.707 "strip_size_kb": 64, 00:12:43.707 "state": "configuring", 00:12:43.707 "raid_level": "raid5f", 00:12:43.707 "superblock": true, 00:12:43.707 "num_base_bdevs": 3, 00:12:43.707 "num_base_bdevs_discovered": 0, 00:12:43.707 "num_base_bdevs_operational": 3, 00:12:43.707 "base_bdevs_list": [ 00:12:43.707 { 00:12:43.707 "name": "BaseBdev1", 00:12:43.707 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:43.707 "is_configured": false, 00:12:43.707 "data_offset": 0, 00:12:43.707 "data_size": 0 00:12:43.707 }, 00:12:43.707 { 00:12:43.707 "name": "BaseBdev2", 00:12:43.707 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:43.707 "is_configured": false, 00:12:43.707 "data_offset": 0, 00:12:43.707 "data_size": 0 00:12:43.707 }, 00:12:43.707 { 00:12:43.707 "name": "BaseBdev3", 00:12:43.707 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:43.707 "is_configured": false, 00:12:43.707 "data_offset": 0, 00:12:43.707 "data_size": 0 00:12:43.707 } 00:12:43.707 ] 00:12:43.707 }' 00:12:43.707 23:46:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:43.707 23:46:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:44.278 23:46:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:12:44.278 23:46:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:44.278 23:46:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:44.278 [2024-11-26 23:46:32.105641] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:12:44.278 [2024-11-26 23:46:32.105676] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:12:44.278 23:46:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:44.278 23:46:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:12:44.278 23:46:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:44.278 23:46:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:44.278 [2024-11-26 23:46:32.117646] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:12:44.278 [2024-11-26 23:46:32.117686] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:12:44.278 [2024-11-26 23:46:32.117694] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:12:44.278 [2024-11-26 23:46:32.117703] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:12:44.278 [2024-11-26 23:46:32.117709] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:12:44.278 [2024-11-26 23:46:32.117717] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:12:44.278 23:46:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:44.278 23:46:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:12:44.278 23:46:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:44.278 23:46:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:44.278 [2024-11-26 23:46:32.138069] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:12:44.278 BaseBdev1 00:12:44.278 23:46:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:44.278 23:46:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:12:44.278 23:46:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:12:44.279 23:46:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:12:44.279 23:46:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:12:44.279 23:46:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:12:44.279 23:46:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:12:44.279 23:46:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:12:44.279 23:46:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:44.279 23:46:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:44.279 23:46:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:44.279 23:46:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:12:44.279 23:46:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:44.279 23:46:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:44.279 [ 00:12:44.279 { 00:12:44.279 "name": "BaseBdev1", 00:12:44.279 "aliases": [ 00:12:44.279 "4fac3b8e-7dfb-407a-8d73-f0ec9b011b4b" 00:12:44.279 ], 00:12:44.279 "product_name": "Malloc disk", 00:12:44.279 "block_size": 512, 00:12:44.279 "num_blocks": 65536, 00:12:44.279 "uuid": "4fac3b8e-7dfb-407a-8d73-f0ec9b011b4b", 00:12:44.279 "assigned_rate_limits": { 00:12:44.279 "rw_ios_per_sec": 0, 00:12:44.279 "rw_mbytes_per_sec": 0, 00:12:44.279 "r_mbytes_per_sec": 0, 00:12:44.279 "w_mbytes_per_sec": 0 00:12:44.279 }, 00:12:44.279 "claimed": true, 00:12:44.279 "claim_type": "exclusive_write", 00:12:44.279 "zoned": false, 00:12:44.279 "supported_io_types": { 00:12:44.279 "read": true, 00:12:44.279 "write": true, 00:12:44.279 "unmap": true, 00:12:44.279 "flush": true, 00:12:44.279 "reset": true, 00:12:44.279 "nvme_admin": false, 00:12:44.279 "nvme_io": false, 00:12:44.279 "nvme_io_md": false, 00:12:44.279 "write_zeroes": true, 00:12:44.279 "zcopy": true, 00:12:44.279 "get_zone_info": false, 00:12:44.279 "zone_management": false, 00:12:44.279 "zone_append": false, 00:12:44.279 "compare": false, 00:12:44.279 "compare_and_write": false, 00:12:44.279 "abort": true, 00:12:44.279 "seek_hole": false, 00:12:44.279 "seek_data": false, 00:12:44.279 "copy": true, 00:12:44.279 "nvme_iov_md": false 00:12:44.279 }, 00:12:44.279 "memory_domains": [ 00:12:44.279 { 00:12:44.279 "dma_device_id": "system", 00:12:44.279 "dma_device_type": 1 00:12:44.279 }, 00:12:44.279 { 00:12:44.279 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:12:44.279 "dma_device_type": 2 00:12:44.279 } 00:12:44.279 ], 00:12:44.279 "driver_specific": {} 00:12:44.279 } 00:12:44.279 ] 00:12:44.279 23:46:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:44.279 23:46:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:12:44.279 23:46:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:12:44.279 23:46:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:12:44.279 23:46:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:12:44.279 23:46:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:12:44.279 23:46:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:12:44.279 23:46:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:44.279 23:46:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:44.279 23:46:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:44.279 23:46:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:44.279 23:46:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:44.279 23:46:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:44.279 23:46:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:12:44.279 23:46:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:44.279 23:46:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:44.279 23:46:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:44.279 23:46:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:44.279 "name": "Existed_Raid", 00:12:44.279 "uuid": "50e7d2b3-d0cd-426e-8ac0-5d07cc1a0084", 00:12:44.279 "strip_size_kb": 64, 00:12:44.279 "state": "configuring", 00:12:44.279 "raid_level": "raid5f", 00:12:44.279 "superblock": true, 00:12:44.279 "num_base_bdevs": 3, 00:12:44.279 "num_base_bdevs_discovered": 1, 00:12:44.279 "num_base_bdevs_operational": 3, 00:12:44.279 "base_bdevs_list": [ 00:12:44.279 { 00:12:44.279 "name": "BaseBdev1", 00:12:44.279 "uuid": "4fac3b8e-7dfb-407a-8d73-f0ec9b011b4b", 00:12:44.279 "is_configured": true, 00:12:44.279 "data_offset": 2048, 00:12:44.279 "data_size": 63488 00:12:44.279 }, 00:12:44.279 { 00:12:44.279 "name": "BaseBdev2", 00:12:44.279 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:44.279 "is_configured": false, 00:12:44.279 "data_offset": 0, 00:12:44.279 "data_size": 0 00:12:44.279 }, 00:12:44.279 { 00:12:44.279 "name": "BaseBdev3", 00:12:44.279 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:44.279 "is_configured": false, 00:12:44.279 "data_offset": 0, 00:12:44.279 "data_size": 0 00:12:44.279 } 00:12:44.279 ] 00:12:44.279 }' 00:12:44.279 23:46:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:44.279 23:46:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:44.539 23:46:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:12:44.539 23:46:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:44.539 23:46:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:44.539 [2024-11-26 23:46:32.601281] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:12:44.539 [2024-11-26 23:46:32.601381] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:12:44.539 23:46:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:44.539 23:46:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:12:44.539 23:46:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:44.539 23:46:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:44.539 [2024-11-26 23:46:32.613298] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:12:44.539 [2024-11-26 23:46:32.615133] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:12:44.539 [2024-11-26 23:46:32.615209] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:12:44.539 [2024-11-26 23:46:32.615253] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:12:44.539 [2024-11-26 23:46:32.615278] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:12:44.539 23:46:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:44.539 23:46:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:12:44.539 23:46:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:12:44.539 23:46:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:12:44.540 23:46:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:12:44.540 23:46:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:12:44.540 23:46:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:12:44.540 23:46:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:12:44.540 23:46:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:44.540 23:46:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:44.540 23:46:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:44.540 23:46:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:44.540 23:46:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:44.540 23:46:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:44.540 23:46:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:12:44.540 23:46:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:44.540 23:46:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:44.540 23:46:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:44.801 23:46:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:44.801 "name": "Existed_Raid", 00:12:44.801 "uuid": "3b971fdc-ef85-425a-b8e3-c1e5ccbc1a00", 00:12:44.801 "strip_size_kb": 64, 00:12:44.801 "state": "configuring", 00:12:44.801 "raid_level": "raid5f", 00:12:44.801 "superblock": true, 00:12:44.801 "num_base_bdevs": 3, 00:12:44.801 "num_base_bdevs_discovered": 1, 00:12:44.801 "num_base_bdevs_operational": 3, 00:12:44.801 "base_bdevs_list": [ 00:12:44.801 { 00:12:44.801 "name": "BaseBdev1", 00:12:44.801 "uuid": "4fac3b8e-7dfb-407a-8d73-f0ec9b011b4b", 00:12:44.801 "is_configured": true, 00:12:44.801 "data_offset": 2048, 00:12:44.801 "data_size": 63488 00:12:44.801 }, 00:12:44.801 { 00:12:44.801 "name": "BaseBdev2", 00:12:44.801 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:44.801 "is_configured": false, 00:12:44.801 "data_offset": 0, 00:12:44.801 "data_size": 0 00:12:44.801 }, 00:12:44.801 { 00:12:44.801 "name": "BaseBdev3", 00:12:44.801 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:44.801 "is_configured": false, 00:12:44.801 "data_offset": 0, 00:12:44.801 "data_size": 0 00:12:44.801 } 00:12:44.801 ] 00:12:44.801 }' 00:12:44.801 23:46:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:44.801 23:46:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:45.071 23:46:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:12:45.071 23:46:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:45.071 23:46:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:45.071 [2024-11-26 23:46:33.035555] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:12:45.071 BaseBdev2 00:12:45.071 23:46:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:45.071 23:46:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:12:45.071 23:46:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:12:45.071 23:46:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:12:45.071 23:46:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:12:45.071 23:46:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:12:45.071 23:46:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:12:45.071 23:46:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:12:45.072 23:46:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:45.072 23:46:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:45.072 23:46:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:45.072 23:46:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:12:45.072 23:46:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:45.072 23:46:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:45.072 [ 00:12:45.072 { 00:12:45.072 "name": "BaseBdev2", 00:12:45.072 "aliases": [ 00:12:45.072 "8a542086-a8eb-4e03-bb85-d7498a4100f2" 00:12:45.072 ], 00:12:45.072 "product_name": "Malloc disk", 00:12:45.072 "block_size": 512, 00:12:45.072 "num_blocks": 65536, 00:12:45.072 "uuid": "8a542086-a8eb-4e03-bb85-d7498a4100f2", 00:12:45.072 "assigned_rate_limits": { 00:12:45.072 "rw_ios_per_sec": 0, 00:12:45.072 "rw_mbytes_per_sec": 0, 00:12:45.072 "r_mbytes_per_sec": 0, 00:12:45.072 "w_mbytes_per_sec": 0 00:12:45.072 }, 00:12:45.072 "claimed": true, 00:12:45.072 "claim_type": "exclusive_write", 00:12:45.072 "zoned": false, 00:12:45.072 "supported_io_types": { 00:12:45.072 "read": true, 00:12:45.072 "write": true, 00:12:45.073 "unmap": true, 00:12:45.073 "flush": true, 00:12:45.073 "reset": true, 00:12:45.073 "nvme_admin": false, 00:12:45.073 "nvme_io": false, 00:12:45.073 "nvme_io_md": false, 00:12:45.073 "write_zeroes": true, 00:12:45.073 "zcopy": true, 00:12:45.074 "get_zone_info": false, 00:12:45.074 "zone_management": false, 00:12:45.074 "zone_append": false, 00:12:45.074 "compare": false, 00:12:45.074 "compare_and_write": false, 00:12:45.074 "abort": true, 00:12:45.074 "seek_hole": false, 00:12:45.074 "seek_data": false, 00:12:45.074 "copy": true, 00:12:45.074 "nvme_iov_md": false 00:12:45.074 }, 00:12:45.074 "memory_domains": [ 00:12:45.074 { 00:12:45.074 "dma_device_id": "system", 00:12:45.074 "dma_device_type": 1 00:12:45.074 }, 00:12:45.074 { 00:12:45.074 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:12:45.074 "dma_device_type": 2 00:12:45.074 } 00:12:45.074 ], 00:12:45.074 "driver_specific": {} 00:12:45.074 } 00:12:45.074 ] 00:12:45.074 23:46:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:45.074 23:46:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:12:45.074 23:46:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:12:45.075 23:46:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:12:45.075 23:46:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:12:45.075 23:46:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:12:45.075 23:46:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:12:45.075 23:46:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:12:45.075 23:46:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:12:45.075 23:46:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:45.075 23:46:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:45.075 23:46:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:45.075 23:46:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:45.075 23:46:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:45.076 23:46:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:45.076 23:46:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:12:45.076 23:46:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:45.076 23:46:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:45.076 23:46:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:45.076 23:46:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:45.076 "name": "Existed_Raid", 00:12:45.076 "uuid": "3b971fdc-ef85-425a-b8e3-c1e5ccbc1a00", 00:12:45.076 "strip_size_kb": 64, 00:12:45.076 "state": "configuring", 00:12:45.076 "raid_level": "raid5f", 00:12:45.076 "superblock": true, 00:12:45.076 "num_base_bdevs": 3, 00:12:45.076 "num_base_bdevs_discovered": 2, 00:12:45.076 "num_base_bdevs_operational": 3, 00:12:45.076 "base_bdevs_list": [ 00:12:45.076 { 00:12:45.076 "name": "BaseBdev1", 00:12:45.076 "uuid": "4fac3b8e-7dfb-407a-8d73-f0ec9b011b4b", 00:12:45.076 "is_configured": true, 00:12:45.076 "data_offset": 2048, 00:12:45.076 "data_size": 63488 00:12:45.076 }, 00:12:45.076 { 00:12:45.076 "name": "BaseBdev2", 00:12:45.076 "uuid": "8a542086-a8eb-4e03-bb85-d7498a4100f2", 00:12:45.076 "is_configured": true, 00:12:45.076 "data_offset": 2048, 00:12:45.076 "data_size": 63488 00:12:45.076 }, 00:12:45.076 { 00:12:45.076 "name": "BaseBdev3", 00:12:45.076 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:45.076 "is_configured": false, 00:12:45.076 "data_offset": 0, 00:12:45.076 "data_size": 0 00:12:45.076 } 00:12:45.076 ] 00:12:45.076 }' 00:12:45.079 23:46:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:45.079 23:46:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:45.650 23:46:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:12:45.650 23:46:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:45.650 23:46:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:45.650 [2024-11-26 23:46:33.539185] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:12:45.650 [2024-11-26 23:46:33.539721] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:12:45.650 [2024-11-26 23:46:33.539848] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:12:45.650 BaseBdev3 00:12:45.650 [2024-11-26 23:46:33.540603] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:12:45.650 23:46:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:45.650 [2024-11-26 23:46:33.541767] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:12:45.650 23:46:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:12:45.650 [2024-11-26 23:46:33.541892] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:12:45.650 23:46:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:12:45.650 [2024-11-26 23:46:33.542303] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:45.650 23:46:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:12:45.650 23:46:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:12:45.650 23:46:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:12:45.650 23:46:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:12:45.650 23:46:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:12:45.650 23:46:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:45.650 23:46:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:45.650 23:46:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:45.650 23:46:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:12:45.650 23:46:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:45.650 23:46:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:45.650 [ 00:12:45.650 { 00:12:45.650 "name": "BaseBdev3", 00:12:45.650 "aliases": [ 00:12:45.650 "905c501e-20b2-4fb5-afbe-18509507c572" 00:12:45.650 ], 00:12:45.650 "product_name": "Malloc disk", 00:12:45.650 "block_size": 512, 00:12:45.650 "num_blocks": 65536, 00:12:45.650 "uuid": "905c501e-20b2-4fb5-afbe-18509507c572", 00:12:45.650 "assigned_rate_limits": { 00:12:45.650 "rw_ios_per_sec": 0, 00:12:45.650 "rw_mbytes_per_sec": 0, 00:12:45.650 "r_mbytes_per_sec": 0, 00:12:45.650 "w_mbytes_per_sec": 0 00:12:45.650 }, 00:12:45.650 "claimed": true, 00:12:45.650 "claim_type": "exclusive_write", 00:12:45.650 "zoned": false, 00:12:45.650 "supported_io_types": { 00:12:45.650 "read": true, 00:12:45.650 "write": true, 00:12:45.650 "unmap": true, 00:12:45.650 "flush": true, 00:12:45.650 "reset": true, 00:12:45.650 "nvme_admin": false, 00:12:45.650 "nvme_io": false, 00:12:45.650 "nvme_io_md": false, 00:12:45.650 "write_zeroes": true, 00:12:45.650 "zcopy": true, 00:12:45.650 "get_zone_info": false, 00:12:45.650 "zone_management": false, 00:12:45.650 "zone_append": false, 00:12:45.650 "compare": false, 00:12:45.650 "compare_and_write": false, 00:12:45.650 "abort": true, 00:12:45.651 "seek_hole": false, 00:12:45.651 "seek_data": false, 00:12:45.651 "copy": true, 00:12:45.651 "nvme_iov_md": false 00:12:45.651 }, 00:12:45.651 "memory_domains": [ 00:12:45.651 { 00:12:45.651 "dma_device_id": "system", 00:12:45.651 "dma_device_type": 1 00:12:45.651 }, 00:12:45.651 { 00:12:45.651 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:12:45.651 "dma_device_type": 2 00:12:45.651 } 00:12:45.651 ], 00:12:45.651 "driver_specific": {} 00:12:45.651 } 00:12:45.651 ] 00:12:45.651 23:46:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:45.651 23:46:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:12:45.651 23:46:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:12:45.651 23:46:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:12:45.651 23:46:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 3 00:12:45.651 23:46:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:12:45.651 23:46:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:45.651 23:46:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:12:45.651 23:46:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:12:45.651 23:46:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:45.651 23:46:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:45.651 23:46:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:45.651 23:46:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:45.651 23:46:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:45.651 23:46:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:45.651 23:46:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:12:45.651 23:46:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:45.651 23:46:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:45.651 23:46:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:45.651 23:46:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:45.651 "name": "Existed_Raid", 00:12:45.651 "uuid": "3b971fdc-ef85-425a-b8e3-c1e5ccbc1a00", 00:12:45.651 "strip_size_kb": 64, 00:12:45.651 "state": "online", 00:12:45.651 "raid_level": "raid5f", 00:12:45.651 "superblock": true, 00:12:45.651 "num_base_bdevs": 3, 00:12:45.651 "num_base_bdevs_discovered": 3, 00:12:45.651 "num_base_bdevs_operational": 3, 00:12:45.651 "base_bdevs_list": [ 00:12:45.651 { 00:12:45.651 "name": "BaseBdev1", 00:12:45.651 "uuid": "4fac3b8e-7dfb-407a-8d73-f0ec9b011b4b", 00:12:45.651 "is_configured": true, 00:12:45.651 "data_offset": 2048, 00:12:45.651 "data_size": 63488 00:12:45.651 }, 00:12:45.651 { 00:12:45.651 "name": "BaseBdev2", 00:12:45.651 "uuid": "8a542086-a8eb-4e03-bb85-d7498a4100f2", 00:12:45.651 "is_configured": true, 00:12:45.651 "data_offset": 2048, 00:12:45.651 "data_size": 63488 00:12:45.651 }, 00:12:45.651 { 00:12:45.651 "name": "BaseBdev3", 00:12:45.651 "uuid": "905c501e-20b2-4fb5-afbe-18509507c572", 00:12:45.651 "is_configured": true, 00:12:45.651 "data_offset": 2048, 00:12:45.651 "data_size": 63488 00:12:45.651 } 00:12:45.651 ] 00:12:45.651 }' 00:12:45.651 23:46:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:45.651 23:46:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:46.222 23:46:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:12:46.222 23:46:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:12:46.222 23:46:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:12:46.222 23:46:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:12:46.222 23:46:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:12:46.222 23:46:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:12:46.222 23:46:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:12:46.222 23:46:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:46.222 23:46:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:46.222 23:46:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:12:46.222 [2024-11-26 23:46:34.051857] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:12:46.222 23:46:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:46.222 23:46:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:12:46.222 "name": "Existed_Raid", 00:12:46.222 "aliases": [ 00:12:46.222 "3b971fdc-ef85-425a-b8e3-c1e5ccbc1a00" 00:12:46.222 ], 00:12:46.222 "product_name": "Raid Volume", 00:12:46.222 "block_size": 512, 00:12:46.222 "num_blocks": 126976, 00:12:46.222 "uuid": "3b971fdc-ef85-425a-b8e3-c1e5ccbc1a00", 00:12:46.222 "assigned_rate_limits": { 00:12:46.222 "rw_ios_per_sec": 0, 00:12:46.222 "rw_mbytes_per_sec": 0, 00:12:46.222 "r_mbytes_per_sec": 0, 00:12:46.222 "w_mbytes_per_sec": 0 00:12:46.222 }, 00:12:46.222 "claimed": false, 00:12:46.222 "zoned": false, 00:12:46.222 "supported_io_types": { 00:12:46.222 "read": true, 00:12:46.222 "write": true, 00:12:46.222 "unmap": false, 00:12:46.222 "flush": false, 00:12:46.222 "reset": true, 00:12:46.222 "nvme_admin": false, 00:12:46.222 "nvme_io": false, 00:12:46.222 "nvme_io_md": false, 00:12:46.222 "write_zeroes": true, 00:12:46.222 "zcopy": false, 00:12:46.222 "get_zone_info": false, 00:12:46.222 "zone_management": false, 00:12:46.222 "zone_append": false, 00:12:46.222 "compare": false, 00:12:46.222 "compare_and_write": false, 00:12:46.222 "abort": false, 00:12:46.222 "seek_hole": false, 00:12:46.222 "seek_data": false, 00:12:46.222 "copy": false, 00:12:46.222 "nvme_iov_md": false 00:12:46.222 }, 00:12:46.222 "driver_specific": { 00:12:46.222 "raid": { 00:12:46.222 "uuid": "3b971fdc-ef85-425a-b8e3-c1e5ccbc1a00", 00:12:46.222 "strip_size_kb": 64, 00:12:46.222 "state": "online", 00:12:46.222 "raid_level": "raid5f", 00:12:46.222 "superblock": true, 00:12:46.222 "num_base_bdevs": 3, 00:12:46.223 "num_base_bdevs_discovered": 3, 00:12:46.223 "num_base_bdevs_operational": 3, 00:12:46.223 "base_bdevs_list": [ 00:12:46.223 { 00:12:46.223 "name": "BaseBdev1", 00:12:46.223 "uuid": "4fac3b8e-7dfb-407a-8d73-f0ec9b011b4b", 00:12:46.223 "is_configured": true, 00:12:46.223 "data_offset": 2048, 00:12:46.223 "data_size": 63488 00:12:46.223 }, 00:12:46.223 { 00:12:46.223 "name": "BaseBdev2", 00:12:46.223 "uuid": "8a542086-a8eb-4e03-bb85-d7498a4100f2", 00:12:46.223 "is_configured": true, 00:12:46.223 "data_offset": 2048, 00:12:46.223 "data_size": 63488 00:12:46.223 }, 00:12:46.223 { 00:12:46.223 "name": "BaseBdev3", 00:12:46.223 "uuid": "905c501e-20b2-4fb5-afbe-18509507c572", 00:12:46.223 "is_configured": true, 00:12:46.223 "data_offset": 2048, 00:12:46.223 "data_size": 63488 00:12:46.223 } 00:12:46.223 ] 00:12:46.223 } 00:12:46.223 } 00:12:46.223 }' 00:12:46.223 23:46:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:12:46.223 23:46:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:12:46.223 BaseBdev2 00:12:46.223 BaseBdev3' 00:12:46.223 23:46:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:12:46.223 23:46:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:12:46.223 23:46:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:12:46.223 23:46:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:12:46.223 23:46:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:12:46.223 23:46:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:46.223 23:46:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:46.223 23:46:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:46.223 23:46:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:12:46.223 23:46:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:12:46.223 23:46:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:12:46.223 23:46:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:12:46.223 23:46:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:46.223 23:46:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:46.223 23:46:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:12:46.223 23:46:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:46.223 23:46:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:12:46.223 23:46:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:12:46.223 23:46:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:12:46.223 23:46:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:12:46.223 23:46:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:46.223 23:46:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:46.223 23:46:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:12:46.223 23:46:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:46.483 23:46:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:12:46.483 23:46:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:12:46.483 23:46:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:12:46.483 23:46:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:46.483 23:46:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:46.483 [2024-11-26 23:46:34.355200] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:12:46.483 23:46:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:46.483 23:46:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:12:46.483 23:46:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy raid5f 00:12:46.483 23:46:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:12:46.483 23:46:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@199 -- # return 0 00:12:46.483 23:46:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:12:46.483 23:46:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 2 00:12:46.483 23:46:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:12:46.483 23:46:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:46.483 23:46:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:12:46.483 23:46:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:12:46.483 23:46:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:12:46.483 23:46:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:46.483 23:46:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:46.483 23:46:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:46.483 23:46:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:46.483 23:46:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:46.483 23:46:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:46.483 23:46:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:46.483 23:46:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:12:46.483 23:46:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:46.483 23:46:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:46.483 "name": "Existed_Raid", 00:12:46.483 "uuid": "3b971fdc-ef85-425a-b8e3-c1e5ccbc1a00", 00:12:46.483 "strip_size_kb": 64, 00:12:46.483 "state": "online", 00:12:46.483 "raid_level": "raid5f", 00:12:46.483 "superblock": true, 00:12:46.483 "num_base_bdevs": 3, 00:12:46.483 "num_base_bdevs_discovered": 2, 00:12:46.483 "num_base_bdevs_operational": 2, 00:12:46.483 "base_bdevs_list": [ 00:12:46.483 { 00:12:46.483 "name": null, 00:12:46.483 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:46.483 "is_configured": false, 00:12:46.483 "data_offset": 0, 00:12:46.483 "data_size": 63488 00:12:46.483 }, 00:12:46.483 { 00:12:46.483 "name": "BaseBdev2", 00:12:46.483 "uuid": "8a542086-a8eb-4e03-bb85-d7498a4100f2", 00:12:46.483 "is_configured": true, 00:12:46.483 "data_offset": 2048, 00:12:46.483 "data_size": 63488 00:12:46.483 }, 00:12:46.483 { 00:12:46.483 "name": "BaseBdev3", 00:12:46.483 "uuid": "905c501e-20b2-4fb5-afbe-18509507c572", 00:12:46.483 "is_configured": true, 00:12:46.483 "data_offset": 2048, 00:12:46.483 "data_size": 63488 00:12:46.483 } 00:12:46.483 ] 00:12:46.483 }' 00:12:46.483 23:46:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:46.483 23:46:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:46.743 23:46:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:12:46.743 23:46:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:12:46.743 23:46:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:46.743 23:46:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:46.743 23:46:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:12:46.743 23:46:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:46.743 23:46:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:46.743 23:46:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:12:46.743 23:46:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:12:46.743 23:46:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:12:46.743 23:46:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:46.743 23:46:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:46.743 [2024-11-26 23:46:34.849530] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:12:46.743 [2024-11-26 23:46:34.849665] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:12:46.743 [2024-11-26 23:46:34.860725] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:12:46.743 23:46:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:46.743 23:46:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:12:46.743 23:46:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:12:46.743 23:46:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:46.743 23:46:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:46.743 23:46:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:46.743 23:46:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:12:47.003 23:46:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:47.003 23:46:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:12:47.003 23:46:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:12:47.003 23:46:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:12:47.003 23:46:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:47.003 23:46:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:47.003 [2024-11-26 23:46:34.920647] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:12:47.003 [2024-11-26 23:46:34.920696] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:12:47.003 23:46:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:47.003 23:46:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:12:47.003 23:46:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:12:47.003 23:46:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:47.003 23:46:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:12:47.003 23:46:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:47.003 23:46:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:47.003 23:46:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:47.003 23:46:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:12:47.003 23:46:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:12:47.003 23:46:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 3 -gt 2 ']' 00:12:47.003 23:46:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:12:47.003 23:46:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:12:47.003 23:46:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:12:47.003 23:46:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:47.003 23:46:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:47.003 BaseBdev2 00:12:47.004 23:46:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:47.004 23:46:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:12:47.004 23:46:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:12:47.004 23:46:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:12:47.004 23:46:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:12:47.004 23:46:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:12:47.004 23:46:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:12:47.004 23:46:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:12:47.004 23:46:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:47.004 23:46:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:47.004 23:46:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:47.004 23:46:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:12:47.004 23:46:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:47.004 23:46:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:47.004 [ 00:12:47.004 { 00:12:47.004 "name": "BaseBdev2", 00:12:47.004 "aliases": [ 00:12:47.004 "d7381a19-da9c-4f3d-8b69-60563848436b" 00:12:47.004 ], 00:12:47.004 "product_name": "Malloc disk", 00:12:47.004 "block_size": 512, 00:12:47.004 "num_blocks": 65536, 00:12:47.004 "uuid": "d7381a19-da9c-4f3d-8b69-60563848436b", 00:12:47.004 "assigned_rate_limits": { 00:12:47.004 "rw_ios_per_sec": 0, 00:12:47.004 "rw_mbytes_per_sec": 0, 00:12:47.004 "r_mbytes_per_sec": 0, 00:12:47.004 "w_mbytes_per_sec": 0 00:12:47.004 }, 00:12:47.004 "claimed": false, 00:12:47.004 "zoned": false, 00:12:47.004 "supported_io_types": { 00:12:47.004 "read": true, 00:12:47.004 "write": true, 00:12:47.004 "unmap": true, 00:12:47.004 "flush": true, 00:12:47.004 "reset": true, 00:12:47.004 "nvme_admin": false, 00:12:47.004 "nvme_io": false, 00:12:47.004 "nvme_io_md": false, 00:12:47.004 "write_zeroes": true, 00:12:47.004 "zcopy": true, 00:12:47.004 "get_zone_info": false, 00:12:47.004 "zone_management": false, 00:12:47.004 "zone_append": false, 00:12:47.004 "compare": false, 00:12:47.004 "compare_and_write": false, 00:12:47.004 "abort": true, 00:12:47.004 "seek_hole": false, 00:12:47.004 "seek_data": false, 00:12:47.004 "copy": true, 00:12:47.004 "nvme_iov_md": false 00:12:47.004 }, 00:12:47.004 "memory_domains": [ 00:12:47.004 { 00:12:47.004 "dma_device_id": "system", 00:12:47.004 "dma_device_type": 1 00:12:47.004 }, 00:12:47.004 { 00:12:47.004 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:12:47.004 "dma_device_type": 2 00:12:47.004 } 00:12:47.004 ], 00:12:47.004 "driver_specific": {} 00:12:47.004 } 00:12:47.004 ] 00:12:47.004 23:46:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:47.004 23:46:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:12:47.004 23:46:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:12:47.004 23:46:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:12:47.004 23:46:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:12:47.004 23:46:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:47.004 23:46:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:47.004 BaseBdev3 00:12:47.004 23:46:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:47.004 23:46:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:12:47.004 23:46:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:12:47.004 23:46:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:12:47.004 23:46:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:12:47.004 23:46:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:12:47.004 23:46:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:12:47.004 23:46:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:12:47.004 23:46:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:47.004 23:46:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:47.004 23:46:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:47.004 23:46:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:12:47.004 23:46:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:47.004 23:46:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:47.004 [ 00:12:47.004 { 00:12:47.004 "name": "BaseBdev3", 00:12:47.004 "aliases": [ 00:12:47.004 "766bfc6b-ad40-48a8-9988-3eb63a768cfa" 00:12:47.004 ], 00:12:47.004 "product_name": "Malloc disk", 00:12:47.004 "block_size": 512, 00:12:47.004 "num_blocks": 65536, 00:12:47.004 "uuid": "766bfc6b-ad40-48a8-9988-3eb63a768cfa", 00:12:47.004 "assigned_rate_limits": { 00:12:47.004 "rw_ios_per_sec": 0, 00:12:47.004 "rw_mbytes_per_sec": 0, 00:12:47.004 "r_mbytes_per_sec": 0, 00:12:47.004 "w_mbytes_per_sec": 0 00:12:47.004 }, 00:12:47.004 "claimed": false, 00:12:47.004 "zoned": false, 00:12:47.004 "supported_io_types": { 00:12:47.004 "read": true, 00:12:47.004 "write": true, 00:12:47.004 "unmap": true, 00:12:47.004 "flush": true, 00:12:47.004 "reset": true, 00:12:47.004 "nvme_admin": false, 00:12:47.004 "nvme_io": false, 00:12:47.004 "nvme_io_md": false, 00:12:47.004 "write_zeroes": true, 00:12:47.004 "zcopy": true, 00:12:47.004 "get_zone_info": false, 00:12:47.004 "zone_management": false, 00:12:47.004 "zone_append": false, 00:12:47.004 "compare": false, 00:12:47.004 "compare_and_write": false, 00:12:47.004 "abort": true, 00:12:47.004 "seek_hole": false, 00:12:47.004 "seek_data": false, 00:12:47.004 "copy": true, 00:12:47.004 "nvme_iov_md": false 00:12:47.004 }, 00:12:47.004 "memory_domains": [ 00:12:47.004 { 00:12:47.004 "dma_device_id": "system", 00:12:47.004 "dma_device_type": 1 00:12:47.004 }, 00:12:47.004 { 00:12:47.004 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:12:47.004 "dma_device_type": 2 00:12:47.004 } 00:12:47.004 ], 00:12:47.004 "driver_specific": {} 00:12:47.004 } 00:12:47.004 ] 00:12:47.004 23:46:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:47.004 23:46:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:12:47.004 23:46:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:12:47.004 23:46:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:12:47.004 23:46:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:12:47.004 23:46:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:47.004 23:46:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:47.004 [2024-11-26 23:46:35.095915] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:12:47.004 [2024-11-26 23:46:35.095997] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:12:47.004 [2024-11-26 23:46:35.096053] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:12:47.004 [2024-11-26 23:46:35.097805] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:12:47.004 23:46:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:47.004 23:46:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:12:47.004 23:46:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:12:47.004 23:46:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:12:47.004 23:46:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:12:47.004 23:46:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:12:47.004 23:46:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:47.004 23:46:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:47.004 23:46:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:47.004 23:46:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:47.004 23:46:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:47.004 23:46:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:47.004 23:46:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:12:47.004 23:46:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:47.004 23:46:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:47.004 23:46:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:47.264 23:46:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:47.264 "name": "Existed_Raid", 00:12:47.264 "uuid": "8bfc6758-a9fd-4421-8720-35579347545a", 00:12:47.264 "strip_size_kb": 64, 00:12:47.264 "state": "configuring", 00:12:47.264 "raid_level": "raid5f", 00:12:47.264 "superblock": true, 00:12:47.264 "num_base_bdevs": 3, 00:12:47.264 "num_base_bdevs_discovered": 2, 00:12:47.264 "num_base_bdevs_operational": 3, 00:12:47.264 "base_bdevs_list": [ 00:12:47.264 { 00:12:47.264 "name": "BaseBdev1", 00:12:47.264 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:47.264 "is_configured": false, 00:12:47.264 "data_offset": 0, 00:12:47.264 "data_size": 0 00:12:47.264 }, 00:12:47.264 { 00:12:47.264 "name": "BaseBdev2", 00:12:47.264 "uuid": "d7381a19-da9c-4f3d-8b69-60563848436b", 00:12:47.264 "is_configured": true, 00:12:47.264 "data_offset": 2048, 00:12:47.264 "data_size": 63488 00:12:47.264 }, 00:12:47.264 { 00:12:47.264 "name": "BaseBdev3", 00:12:47.264 "uuid": "766bfc6b-ad40-48a8-9988-3eb63a768cfa", 00:12:47.264 "is_configured": true, 00:12:47.264 "data_offset": 2048, 00:12:47.264 "data_size": 63488 00:12:47.264 } 00:12:47.264 ] 00:12:47.264 }' 00:12:47.264 23:46:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:47.264 23:46:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:47.525 23:46:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:12:47.525 23:46:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:47.525 23:46:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:47.525 [2024-11-26 23:46:35.559153] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:12:47.525 23:46:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:47.525 23:46:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:12:47.525 23:46:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:12:47.525 23:46:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:12:47.525 23:46:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:12:47.525 23:46:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:12:47.525 23:46:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:47.525 23:46:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:47.525 23:46:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:47.525 23:46:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:47.525 23:46:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:47.525 23:46:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:47.525 23:46:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:47.525 23:46:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:47.525 23:46:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:12:47.525 23:46:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:47.525 23:46:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:47.525 "name": "Existed_Raid", 00:12:47.525 "uuid": "8bfc6758-a9fd-4421-8720-35579347545a", 00:12:47.525 "strip_size_kb": 64, 00:12:47.525 "state": "configuring", 00:12:47.525 "raid_level": "raid5f", 00:12:47.525 "superblock": true, 00:12:47.525 "num_base_bdevs": 3, 00:12:47.525 "num_base_bdevs_discovered": 1, 00:12:47.525 "num_base_bdevs_operational": 3, 00:12:47.525 "base_bdevs_list": [ 00:12:47.525 { 00:12:47.525 "name": "BaseBdev1", 00:12:47.525 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:47.525 "is_configured": false, 00:12:47.525 "data_offset": 0, 00:12:47.525 "data_size": 0 00:12:47.525 }, 00:12:47.525 { 00:12:47.525 "name": null, 00:12:47.525 "uuid": "d7381a19-da9c-4f3d-8b69-60563848436b", 00:12:47.525 "is_configured": false, 00:12:47.525 "data_offset": 0, 00:12:47.525 "data_size": 63488 00:12:47.525 }, 00:12:47.525 { 00:12:47.525 "name": "BaseBdev3", 00:12:47.525 "uuid": "766bfc6b-ad40-48a8-9988-3eb63a768cfa", 00:12:47.525 "is_configured": true, 00:12:47.525 "data_offset": 2048, 00:12:47.525 "data_size": 63488 00:12:47.525 } 00:12:47.525 ] 00:12:47.525 }' 00:12:47.525 23:46:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:47.525 23:46:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:48.101 23:46:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:48.101 23:46:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:12:48.101 23:46:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:48.101 23:46:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:48.101 23:46:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:48.101 23:46:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:12:48.101 23:46:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:12:48.101 23:46:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:48.101 23:46:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:48.101 [2024-11-26 23:46:36.061299] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:12:48.101 BaseBdev1 00:12:48.101 23:46:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:48.101 23:46:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:12:48.101 23:46:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:12:48.101 23:46:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:12:48.101 23:46:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:12:48.101 23:46:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:12:48.101 23:46:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:12:48.101 23:46:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:12:48.101 23:46:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:48.101 23:46:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:48.101 23:46:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:48.101 23:46:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:12:48.101 23:46:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:48.101 23:46:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:48.101 [ 00:12:48.101 { 00:12:48.101 "name": "BaseBdev1", 00:12:48.101 "aliases": [ 00:12:48.102 "d2dcd3d4-d10b-4ae0-b1ed-06083f748bac" 00:12:48.102 ], 00:12:48.102 "product_name": "Malloc disk", 00:12:48.102 "block_size": 512, 00:12:48.102 "num_blocks": 65536, 00:12:48.102 "uuid": "d2dcd3d4-d10b-4ae0-b1ed-06083f748bac", 00:12:48.102 "assigned_rate_limits": { 00:12:48.102 "rw_ios_per_sec": 0, 00:12:48.102 "rw_mbytes_per_sec": 0, 00:12:48.102 "r_mbytes_per_sec": 0, 00:12:48.102 "w_mbytes_per_sec": 0 00:12:48.102 }, 00:12:48.102 "claimed": true, 00:12:48.102 "claim_type": "exclusive_write", 00:12:48.102 "zoned": false, 00:12:48.102 "supported_io_types": { 00:12:48.102 "read": true, 00:12:48.102 "write": true, 00:12:48.102 "unmap": true, 00:12:48.102 "flush": true, 00:12:48.102 "reset": true, 00:12:48.102 "nvme_admin": false, 00:12:48.102 "nvme_io": false, 00:12:48.102 "nvme_io_md": false, 00:12:48.102 "write_zeroes": true, 00:12:48.102 "zcopy": true, 00:12:48.102 "get_zone_info": false, 00:12:48.102 "zone_management": false, 00:12:48.102 "zone_append": false, 00:12:48.102 "compare": false, 00:12:48.102 "compare_and_write": false, 00:12:48.102 "abort": true, 00:12:48.102 "seek_hole": false, 00:12:48.102 "seek_data": false, 00:12:48.102 "copy": true, 00:12:48.102 "nvme_iov_md": false 00:12:48.102 }, 00:12:48.102 "memory_domains": [ 00:12:48.102 { 00:12:48.102 "dma_device_id": "system", 00:12:48.102 "dma_device_type": 1 00:12:48.102 }, 00:12:48.102 { 00:12:48.102 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:12:48.102 "dma_device_type": 2 00:12:48.102 } 00:12:48.102 ], 00:12:48.102 "driver_specific": {} 00:12:48.102 } 00:12:48.102 ] 00:12:48.102 23:46:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:48.102 23:46:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:12:48.102 23:46:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:12:48.102 23:46:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:12:48.102 23:46:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:12:48.102 23:46:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:12:48.102 23:46:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:12:48.102 23:46:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:48.102 23:46:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:48.102 23:46:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:48.102 23:46:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:48.102 23:46:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:48.102 23:46:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:48.102 23:46:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:12:48.102 23:46:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:48.102 23:46:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:48.102 23:46:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:48.102 23:46:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:48.102 "name": "Existed_Raid", 00:12:48.102 "uuid": "8bfc6758-a9fd-4421-8720-35579347545a", 00:12:48.102 "strip_size_kb": 64, 00:12:48.102 "state": "configuring", 00:12:48.102 "raid_level": "raid5f", 00:12:48.102 "superblock": true, 00:12:48.102 "num_base_bdevs": 3, 00:12:48.102 "num_base_bdevs_discovered": 2, 00:12:48.102 "num_base_bdevs_operational": 3, 00:12:48.102 "base_bdevs_list": [ 00:12:48.102 { 00:12:48.102 "name": "BaseBdev1", 00:12:48.102 "uuid": "d2dcd3d4-d10b-4ae0-b1ed-06083f748bac", 00:12:48.102 "is_configured": true, 00:12:48.102 "data_offset": 2048, 00:12:48.102 "data_size": 63488 00:12:48.102 }, 00:12:48.102 { 00:12:48.102 "name": null, 00:12:48.102 "uuid": "d7381a19-da9c-4f3d-8b69-60563848436b", 00:12:48.102 "is_configured": false, 00:12:48.102 "data_offset": 0, 00:12:48.102 "data_size": 63488 00:12:48.102 }, 00:12:48.102 { 00:12:48.102 "name": "BaseBdev3", 00:12:48.102 "uuid": "766bfc6b-ad40-48a8-9988-3eb63a768cfa", 00:12:48.102 "is_configured": true, 00:12:48.102 "data_offset": 2048, 00:12:48.102 "data_size": 63488 00:12:48.102 } 00:12:48.102 ] 00:12:48.102 }' 00:12:48.102 23:46:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:48.102 23:46:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:48.674 23:46:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:12:48.674 23:46:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:48.674 23:46:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:48.674 23:46:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:48.674 23:46:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:48.674 23:46:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:12:48.674 23:46:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:12:48.674 23:46:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:48.674 23:46:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:48.674 [2024-11-26 23:46:36.576448] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:12:48.674 23:46:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:48.674 23:46:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:12:48.674 23:46:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:12:48.674 23:46:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:12:48.674 23:46:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:12:48.674 23:46:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:12:48.674 23:46:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:48.675 23:46:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:48.675 23:46:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:48.675 23:46:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:48.675 23:46:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:48.675 23:46:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:48.675 23:46:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:48.675 23:46:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:48.675 23:46:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:12:48.675 23:46:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:48.675 23:46:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:48.675 "name": "Existed_Raid", 00:12:48.675 "uuid": "8bfc6758-a9fd-4421-8720-35579347545a", 00:12:48.675 "strip_size_kb": 64, 00:12:48.675 "state": "configuring", 00:12:48.675 "raid_level": "raid5f", 00:12:48.675 "superblock": true, 00:12:48.675 "num_base_bdevs": 3, 00:12:48.675 "num_base_bdevs_discovered": 1, 00:12:48.675 "num_base_bdevs_operational": 3, 00:12:48.675 "base_bdevs_list": [ 00:12:48.675 { 00:12:48.675 "name": "BaseBdev1", 00:12:48.675 "uuid": "d2dcd3d4-d10b-4ae0-b1ed-06083f748bac", 00:12:48.675 "is_configured": true, 00:12:48.675 "data_offset": 2048, 00:12:48.675 "data_size": 63488 00:12:48.675 }, 00:12:48.675 { 00:12:48.675 "name": null, 00:12:48.675 "uuid": "d7381a19-da9c-4f3d-8b69-60563848436b", 00:12:48.675 "is_configured": false, 00:12:48.675 "data_offset": 0, 00:12:48.675 "data_size": 63488 00:12:48.675 }, 00:12:48.675 { 00:12:48.675 "name": null, 00:12:48.675 "uuid": "766bfc6b-ad40-48a8-9988-3eb63a768cfa", 00:12:48.675 "is_configured": false, 00:12:48.675 "data_offset": 0, 00:12:48.675 "data_size": 63488 00:12:48.675 } 00:12:48.675 ] 00:12:48.675 }' 00:12:48.675 23:46:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:48.675 23:46:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:48.935 23:46:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:48.935 23:46:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:48.935 23:46:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:48.935 23:46:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:12:48.936 23:46:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:48.936 23:46:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:12:48.936 23:46:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:12:48.936 23:46:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:48.936 23:46:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:48.936 [2024-11-26 23:46:37.043656] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:12:48.936 23:46:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:48.936 23:46:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:12:48.936 23:46:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:12:48.936 23:46:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:12:48.936 23:46:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:12:48.936 23:46:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:12:48.936 23:46:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:48.936 23:46:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:48.936 23:46:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:48.936 23:46:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:48.936 23:46:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:48.936 23:46:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:48.936 23:46:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:12:48.936 23:46:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:48.936 23:46:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:49.195 23:46:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:49.195 23:46:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:49.195 "name": "Existed_Raid", 00:12:49.195 "uuid": "8bfc6758-a9fd-4421-8720-35579347545a", 00:12:49.195 "strip_size_kb": 64, 00:12:49.195 "state": "configuring", 00:12:49.195 "raid_level": "raid5f", 00:12:49.195 "superblock": true, 00:12:49.195 "num_base_bdevs": 3, 00:12:49.195 "num_base_bdevs_discovered": 2, 00:12:49.195 "num_base_bdevs_operational": 3, 00:12:49.195 "base_bdevs_list": [ 00:12:49.195 { 00:12:49.195 "name": "BaseBdev1", 00:12:49.195 "uuid": "d2dcd3d4-d10b-4ae0-b1ed-06083f748bac", 00:12:49.195 "is_configured": true, 00:12:49.195 "data_offset": 2048, 00:12:49.195 "data_size": 63488 00:12:49.195 }, 00:12:49.195 { 00:12:49.195 "name": null, 00:12:49.195 "uuid": "d7381a19-da9c-4f3d-8b69-60563848436b", 00:12:49.195 "is_configured": false, 00:12:49.195 "data_offset": 0, 00:12:49.195 "data_size": 63488 00:12:49.195 }, 00:12:49.195 { 00:12:49.195 "name": "BaseBdev3", 00:12:49.195 "uuid": "766bfc6b-ad40-48a8-9988-3eb63a768cfa", 00:12:49.195 "is_configured": true, 00:12:49.195 "data_offset": 2048, 00:12:49.195 "data_size": 63488 00:12:49.195 } 00:12:49.195 ] 00:12:49.195 }' 00:12:49.195 23:46:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:49.195 23:46:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:49.454 23:46:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:49.454 23:46:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:12:49.454 23:46:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:49.454 23:46:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:49.454 23:46:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:49.454 23:46:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:12:49.454 23:46:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:12:49.454 23:46:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:49.454 23:46:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:49.454 [2024-11-26 23:46:37.522877] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:12:49.454 23:46:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:49.454 23:46:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:12:49.454 23:46:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:12:49.454 23:46:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:12:49.454 23:46:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:12:49.454 23:46:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:12:49.454 23:46:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:49.454 23:46:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:49.454 23:46:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:49.455 23:46:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:49.455 23:46:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:49.455 23:46:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:49.455 23:46:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:12:49.455 23:46:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:49.455 23:46:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:49.455 23:46:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:49.715 23:46:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:49.715 "name": "Existed_Raid", 00:12:49.715 "uuid": "8bfc6758-a9fd-4421-8720-35579347545a", 00:12:49.715 "strip_size_kb": 64, 00:12:49.715 "state": "configuring", 00:12:49.715 "raid_level": "raid5f", 00:12:49.715 "superblock": true, 00:12:49.715 "num_base_bdevs": 3, 00:12:49.715 "num_base_bdevs_discovered": 1, 00:12:49.715 "num_base_bdevs_operational": 3, 00:12:49.715 "base_bdevs_list": [ 00:12:49.715 { 00:12:49.715 "name": null, 00:12:49.715 "uuid": "d2dcd3d4-d10b-4ae0-b1ed-06083f748bac", 00:12:49.715 "is_configured": false, 00:12:49.715 "data_offset": 0, 00:12:49.715 "data_size": 63488 00:12:49.715 }, 00:12:49.715 { 00:12:49.715 "name": null, 00:12:49.715 "uuid": "d7381a19-da9c-4f3d-8b69-60563848436b", 00:12:49.715 "is_configured": false, 00:12:49.715 "data_offset": 0, 00:12:49.715 "data_size": 63488 00:12:49.715 }, 00:12:49.715 { 00:12:49.715 "name": "BaseBdev3", 00:12:49.715 "uuid": "766bfc6b-ad40-48a8-9988-3eb63a768cfa", 00:12:49.715 "is_configured": true, 00:12:49.715 "data_offset": 2048, 00:12:49.715 "data_size": 63488 00:12:49.715 } 00:12:49.715 ] 00:12:49.715 }' 00:12:49.715 23:46:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:49.715 23:46:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:49.976 23:46:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:12:49.976 23:46:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:49.976 23:46:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:49.976 23:46:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:49.976 23:46:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:49.976 23:46:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:12:49.976 23:46:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:12:49.976 23:46:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:49.976 23:46:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:49.976 [2024-11-26 23:46:38.048494] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:12:49.976 23:46:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:49.976 23:46:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:12:49.976 23:46:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:12:49.976 23:46:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:12:49.976 23:46:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:12:49.976 23:46:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:12:49.976 23:46:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:49.976 23:46:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:49.976 23:46:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:49.976 23:46:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:49.976 23:46:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:49.976 23:46:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:49.976 23:46:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:12:49.976 23:46:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:49.976 23:46:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:49.976 23:46:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:49.976 23:46:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:49.976 "name": "Existed_Raid", 00:12:49.976 "uuid": "8bfc6758-a9fd-4421-8720-35579347545a", 00:12:49.976 "strip_size_kb": 64, 00:12:49.976 "state": "configuring", 00:12:49.976 "raid_level": "raid5f", 00:12:49.976 "superblock": true, 00:12:49.976 "num_base_bdevs": 3, 00:12:49.976 "num_base_bdevs_discovered": 2, 00:12:49.976 "num_base_bdevs_operational": 3, 00:12:49.976 "base_bdevs_list": [ 00:12:49.976 { 00:12:49.976 "name": null, 00:12:49.976 "uuid": "d2dcd3d4-d10b-4ae0-b1ed-06083f748bac", 00:12:49.976 "is_configured": false, 00:12:49.976 "data_offset": 0, 00:12:49.976 "data_size": 63488 00:12:49.976 }, 00:12:49.976 { 00:12:49.976 "name": "BaseBdev2", 00:12:49.976 "uuid": "d7381a19-da9c-4f3d-8b69-60563848436b", 00:12:49.976 "is_configured": true, 00:12:49.976 "data_offset": 2048, 00:12:49.976 "data_size": 63488 00:12:49.976 }, 00:12:49.976 { 00:12:49.976 "name": "BaseBdev3", 00:12:49.976 "uuid": "766bfc6b-ad40-48a8-9988-3eb63a768cfa", 00:12:49.976 "is_configured": true, 00:12:49.976 "data_offset": 2048, 00:12:49.976 "data_size": 63488 00:12:49.976 } 00:12:49.976 ] 00:12:49.976 }' 00:12:49.976 23:46:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:50.237 23:46:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:50.497 23:46:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:50.498 23:46:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:50.498 23:46:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:50.498 23:46:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:12:50.498 23:46:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:50.498 23:46:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:12:50.498 23:46:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:12:50.498 23:46:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:50.498 23:46:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:50.498 23:46:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:50.498 23:46:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:50.498 23:46:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u d2dcd3d4-d10b-4ae0-b1ed-06083f748bac 00:12:50.498 23:46:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:50.498 23:46:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:50.498 [2024-11-26 23:46:38.534771] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:12:50.498 [2024-11-26 23:46:38.535015] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:12:50.498 [2024-11-26 23:46:38.535065] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:12:50.498 [2024-11-26 23:46:38.535324] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002870 00:12:50.498 NewBaseBdev 00:12:50.498 [2024-11-26 23:46:38.535751] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:12:50.498 [2024-11-26 23:46:38.535801] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001c80 00:12:50.498 [2024-11-26 23:46:38.535944] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:50.498 23:46:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:50.498 23:46:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:12:50.498 23:46:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:12:50.498 23:46:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:12:50.498 23:46:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:12:50.498 23:46:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:12:50.498 23:46:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:12:50.498 23:46:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:12:50.498 23:46:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:50.498 23:46:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:50.498 23:46:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:50.498 23:46:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:12:50.498 23:46:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:50.498 23:46:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:50.498 [ 00:12:50.498 { 00:12:50.498 "name": "NewBaseBdev", 00:12:50.498 "aliases": [ 00:12:50.498 "d2dcd3d4-d10b-4ae0-b1ed-06083f748bac" 00:12:50.498 ], 00:12:50.498 "product_name": "Malloc disk", 00:12:50.498 "block_size": 512, 00:12:50.498 "num_blocks": 65536, 00:12:50.498 "uuid": "d2dcd3d4-d10b-4ae0-b1ed-06083f748bac", 00:12:50.498 "assigned_rate_limits": { 00:12:50.498 "rw_ios_per_sec": 0, 00:12:50.498 "rw_mbytes_per_sec": 0, 00:12:50.498 "r_mbytes_per_sec": 0, 00:12:50.498 "w_mbytes_per_sec": 0 00:12:50.498 }, 00:12:50.498 "claimed": true, 00:12:50.498 "claim_type": "exclusive_write", 00:12:50.498 "zoned": false, 00:12:50.498 "supported_io_types": { 00:12:50.498 "read": true, 00:12:50.498 "write": true, 00:12:50.498 "unmap": true, 00:12:50.498 "flush": true, 00:12:50.498 "reset": true, 00:12:50.498 "nvme_admin": false, 00:12:50.498 "nvme_io": false, 00:12:50.498 "nvme_io_md": false, 00:12:50.498 "write_zeroes": true, 00:12:50.498 "zcopy": true, 00:12:50.498 "get_zone_info": false, 00:12:50.498 "zone_management": false, 00:12:50.498 "zone_append": false, 00:12:50.498 "compare": false, 00:12:50.498 "compare_and_write": false, 00:12:50.498 "abort": true, 00:12:50.498 "seek_hole": false, 00:12:50.498 "seek_data": false, 00:12:50.498 "copy": true, 00:12:50.498 "nvme_iov_md": false 00:12:50.498 }, 00:12:50.498 "memory_domains": [ 00:12:50.498 { 00:12:50.498 "dma_device_id": "system", 00:12:50.498 "dma_device_type": 1 00:12:50.498 }, 00:12:50.498 { 00:12:50.498 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:12:50.498 "dma_device_type": 2 00:12:50.498 } 00:12:50.498 ], 00:12:50.498 "driver_specific": {} 00:12:50.498 } 00:12:50.498 ] 00:12:50.498 23:46:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:50.498 23:46:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:12:50.498 23:46:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 3 00:12:50.498 23:46:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:12:50.498 23:46:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:50.498 23:46:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:12:50.498 23:46:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:12:50.498 23:46:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:50.498 23:46:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:50.498 23:46:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:50.498 23:46:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:50.498 23:46:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:50.498 23:46:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:50.498 23:46:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:12:50.498 23:46:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:50.498 23:46:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:50.498 23:46:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:50.498 23:46:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:50.498 "name": "Existed_Raid", 00:12:50.498 "uuid": "8bfc6758-a9fd-4421-8720-35579347545a", 00:12:50.498 "strip_size_kb": 64, 00:12:50.498 "state": "online", 00:12:50.498 "raid_level": "raid5f", 00:12:50.498 "superblock": true, 00:12:50.498 "num_base_bdevs": 3, 00:12:50.498 "num_base_bdevs_discovered": 3, 00:12:50.498 "num_base_bdevs_operational": 3, 00:12:50.498 "base_bdevs_list": [ 00:12:50.498 { 00:12:50.498 "name": "NewBaseBdev", 00:12:50.498 "uuid": "d2dcd3d4-d10b-4ae0-b1ed-06083f748bac", 00:12:50.498 "is_configured": true, 00:12:50.498 "data_offset": 2048, 00:12:50.498 "data_size": 63488 00:12:50.498 }, 00:12:50.498 { 00:12:50.498 "name": "BaseBdev2", 00:12:50.498 "uuid": "d7381a19-da9c-4f3d-8b69-60563848436b", 00:12:50.498 "is_configured": true, 00:12:50.498 "data_offset": 2048, 00:12:50.498 "data_size": 63488 00:12:50.498 }, 00:12:50.498 { 00:12:50.498 "name": "BaseBdev3", 00:12:50.498 "uuid": "766bfc6b-ad40-48a8-9988-3eb63a768cfa", 00:12:50.498 "is_configured": true, 00:12:50.498 "data_offset": 2048, 00:12:50.498 "data_size": 63488 00:12:50.498 } 00:12:50.498 ] 00:12:50.498 }' 00:12:50.759 23:46:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:50.759 23:46:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:51.018 23:46:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:12:51.018 23:46:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:12:51.018 23:46:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:12:51.018 23:46:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:12:51.018 23:46:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:12:51.018 23:46:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:12:51.018 23:46:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:12:51.018 23:46:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:12:51.018 23:46:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:51.018 23:46:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:51.018 [2024-11-26 23:46:39.018247] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:12:51.018 23:46:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:51.018 23:46:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:12:51.018 "name": "Existed_Raid", 00:12:51.018 "aliases": [ 00:12:51.018 "8bfc6758-a9fd-4421-8720-35579347545a" 00:12:51.018 ], 00:12:51.018 "product_name": "Raid Volume", 00:12:51.018 "block_size": 512, 00:12:51.018 "num_blocks": 126976, 00:12:51.018 "uuid": "8bfc6758-a9fd-4421-8720-35579347545a", 00:12:51.018 "assigned_rate_limits": { 00:12:51.018 "rw_ios_per_sec": 0, 00:12:51.018 "rw_mbytes_per_sec": 0, 00:12:51.018 "r_mbytes_per_sec": 0, 00:12:51.018 "w_mbytes_per_sec": 0 00:12:51.018 }, 00:12:51.018 "claimed": false, 00:12:51.018 "zoned": false, 00:12:51.018 "supported_io_types": { 00:12:51.018 "read": true, 00:12:51.018 "write": true, 00:12:51.018 "unmap": false, 00:12:51.018 "flush": false, 00:12:51.018 "reset": true, 00:12:51.018 "nvme_admin": false, 00:12:51.018 "nvme_io": false, 00:12:51.018 "nvme_io_md": false, 00:12:51.018 "write_zeroes": true, 00:12:51.018 "zcopy": false, 00:12:51.018 "get_zone_info": false, 00:12:51.018 "zone_management": false, 00:12:51.018 "zone_append": false, 00:12:51.018 "compare": false, 00:12:51.018 "compare_and_write": false, 00:12:51.018 "abort": false, 00:12:51.018 "seek_hole": false, 00:12:51.018 "seek_data": false, 00:12:51.018 "copy": false, 00:12:51.018 "nvme_iov_md": false 00:12:51.018 }, 00:12:51.018 "driver_specific": { 00:12:51.018 "raid": { 00:12:51.018 "uuid": "8bfc6758-a9fd-4421-8720-35579347545a", 00:12:51.018 "strip_size_kb": 64, 00:12:51.018 "state": "online", 00:12:51.018 "raid_level": "raid5f", 00:12:51.018 "superblock": true, 00:12:51.018 "num_base_bdevs": 3, 00:12:51.018 "num_base_bdevs_discovered": 3, 00:12:51.018 "num_base_bdevs_operational": 3, 00:12:51.018 "base_bdevs_list": [ 00:12:51.018 { 00:12:51.018 "name": "NewBaseBdev", 00:12:51.018 "uuid": "d2dcd3d4-d10b-4ae0-b1ed-06083f748bac", 00:12:51.018 "is_configured": true, 00:12:51.018 "data_offset": 2048, 00:12:51.018 "data_size": 63488 00:12:51.018 }, 00:12:51.018 { 00:12:51.018 "name": "BaseBdev2", 00:12:51.018 "uuid": "d7381a19-da9c-4f3d-8b69-60563848436b", 00:12:51.018 "is_configured": true, 00:12:51.018 "data_offset": 2048, 00:12:51.018 "data_size": 63488 00:12:51.018 }, 00:12:51.018 { 00:12:51.018 "name": "BaseBdev3", 00:12:51.018 "uuid": "766bfc6b-ad40-48a8-9988-3eb63a768cfa", 00:12:51.018 "is_configured": true, 00:12:51.018 "data_offset": 2048, 00:12:51.018 "data_size": 63488 00:12:51.018 } 00:12:51.018 ] 00:12:51.018 } 00:12:51.018 } 00:12:51.018 }' 00:12:51.018 23:46:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:12:51.018 23:46:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:12:51.018 BaseBdev2 00:12:51.018 BaseBdev3' 00:12:51.018 23:46:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:12:51.277 23:46:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:12:51.277 23:46:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:12:51.277 23:46:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:12:51.277 23:46:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:51.277 23:46:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:51.277 23:46:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:12:51.277 23:46:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:51.277 23:46:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:12:51.277 23:46:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:12:51.277 23:46:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:12:51.277 23:46:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:12:51.277 23:46:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:12:51.277 23:46:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:51.277 23:46:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:51.277 23:46:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:51.277 23:46:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:12:51.277 23:46:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:12:51.277 23:46:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:12:51.277 23:46:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:12:51.277 23:46:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:51.277 23:46:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:51.277 23:46:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:12:51.277 23:46:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:51.277 23:46:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:12:51.278 23:46:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:12:51.278 23:46:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:12:51.278 23:46:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:51.278 23:46:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:51.278 [2024-11-26 23:46:39.301574] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:12:51.278 [2024-11-26 23:46:39.301597] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:12:51.278 [2024-11-26 23:46:39.301666] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:12:51.278 [2024-11-26 23:46:39.301906] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:12:51.278 [2024-11-26 23:46:39.301921] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name Existed_Raid, state offline 00:12:51.278 23:46:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:51.278 23:46:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 90741 00:12:51.278 23:46:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@954 -- # '[' -z 90741 ']' 00:12:51.278 23:46:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@958 -- # kill -0 90741 00:12:51.278 23:46:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@959 -- # uname 00:12:51.278 23:46:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:12:51.278 23:46:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 90741 00:12:51.278 23:46:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:12:51.278 killing process with pid 90741 00:12:51.278 23:46:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:12:51.278 23:46:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@972 -- # echo 'killing process with pid 90741' 00:12:51.278 23:46:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@973 -- # kill 90741 00:12:51.278 [2024-11-26 23:46:39.349544] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:12:51.278 23:46:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@978 -- # wait 90741 00:12:51.278 [2024-11-26 23:46:39.379083] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:12:51.537 ************************************ 00:12:51.537 END TEST raid5f_state_function_test_sb 00:12:51.537 23:46:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:12:51.537 00:12:51.537 real 0m8.861s 00:12:51.537 user 0m15.155s 00:12:51.537 sys 0m1.819s 00:12:51.537 23:46:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@1130 -- # xtrace_disable 00:12:51.537 23:46:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:51.537 ************************************ 00:12:51.537 23:46:39 bdev_raid -- bdev/bdev_raid.sh@988 -- # run_test raid5f_superblock_test raid_superblock_test raid5f 3 00:12:51.537 23:46:39 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:12:51.537 23:46:39 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:12:51.537 23:46:39 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:12:51.797 ************************************ 00:12:51.797 START TEST raid5f_superblock_test 00:12:51.797 ************************************ 00:12:51.797 23:46:39 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@1129 -- # raid_superblock_test raid5f 3 00:12:51.797 23:46:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=raid5f 00:12:51.797 23:46:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=3 00:12:51.797 23:46:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:12:51.797 23:46:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:12:51.797 23:46:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:12:51.797 23:46:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:12:51.797 23:46:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:12:51.797 23:46:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:12:51.797 23:46:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:12:51.797 23:46:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:12:51.797 23:46:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:12:51.797 23:46:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:12:51.797 23:46:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:12:51.797 23:46:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' raid5f '!=' raid1 ']' 00:12:51.797 23:46:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@405 -- # strip_size=64 00:12:51.797 23:46:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@406 -- # strip_size_create_arg='-z 64' 00:12:51.797 23:46:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=91345 00:12:51.797 23:46:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:12:51.797 23:46:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 91345 00:12:51.797 23:46:39 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@835 -- # '[' -z 91345 ']' 00:12:51.797 23:46:39 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:12:51.797 23:46:39 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:12:51.797 23:46:39 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:12:51.797 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:12:51.797 23:46:39 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:12:51.797 23:46:39 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:51.797 [2024-11-26 23:46:39.748998] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:12:51.797 [2024-11-26 23:46:39.749208] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid91345 ] 00:12:51.797 [2024-11-26 23:46:39.880180] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:12:51.797 [2024-11-26 23:46:39.904334] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:12:52.057 [2024-11-26 23:46:39.946038] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:12:52.057 [2024-11-26 23:46:39.946151] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:12:52.657 23:46:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:12:52.657 23:46:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@868 -- # return 0 00:12:52.657 23:46:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:12:52.657 23:46:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:12:52.657 23:46:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:12:52.657 23:46:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:12:52.657 23:46:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:12:52.657 23:46:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:12:52.657 23:46:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:12:52.657 23:46:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:12:52.657 23:46:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:12:52.657 23:46:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:52.657 23:46:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:52.657 malloc1 00:12:52.657 23:46:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:52.657 23:46:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:12:52.657 23:46:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:52.657 23:46:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:52.657 [2024-11-26 23:46:40.585155] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:12:52.657 [2024-11-26 23:46:40.585225] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:52.657 [2024-11-26 23:46:40.585265] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:12:52.657 [2024-11-26 23:46:40.585279] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:52.657 [2024-11-26 23:46:40.587428] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:52.657 [2024-11-26 23:46:40.587521] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:12:52.657 pt1 00:12:52.657 23:46:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:52.657 23:46:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:12:52.657 23:46:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:12:52.657 23:46:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:12:52.657 23:46:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:12:52.657 23:46:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:12:52.657 23:46:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:12:52.657 23:46:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:12:52.657 23:46:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:12:52.657 23:46:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:12:52.657 23:46:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:52.657 23:46:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:52.657 malloc2 00:12:52.657 23:46:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:52.657 23:46:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:12:52.657 23:46:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:52.657 23:46:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:52.658 [2024-11-26 23:46:40.613466] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:12:52.658 [2024-11-26 23:46:40.613585] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:52.658 [2024-11-26 23:46:40.613622] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:12:52.658 [2024-11-26 23:46:40.613657] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:52.658 [2024-11-26 23:46:40.615666] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:52.658 [2024-11-26 23:46:40.615737] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:12:52.658 pt2 00:12:52.658 23:46:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:52.658 23:46:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:12:52.658 23:46:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:12:52.658 23:46:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc3 00:12:52.658 23:46:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt3 00:12:52.658 23:46:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000003 00:12:52.658 23:46:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:12:52.658 23:46:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:12:52.658 23:46:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:12:52.658 23:46:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc3 00:12:52.658 23:46:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:52.658 23:46:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:52.658 malloc3 00:12:52.658 23:46:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:52.658 23:46:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:12:52.658 23:46:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:52.658 23:46:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:52.658 [2024-11-26 23:46:40.645715] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:12:52.658 [2024-11-26 23:46:40.645820] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:52.658 [2024-11-26 23:46:40.645853] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:12:52.658 [2024-11-26 23:46:40.645881] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:52.658 [2024-11-26 23:46:40.647913] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:52.658 [2024-11-26 23:46:40.647983] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:12:52.658 pt3 00:12:52.658 23:46:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:52.658 23:46:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:12:52.658 23:46:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:12:52.658 23:46:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''pt1 pt2 pt3'\''' -n raid_bdev1 -s 00:12:52.658 23:46:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:52.658 23:46:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:52.658 [2024-11-26 23:46:40.657737] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:12:52.658 [2024-11-26 23:46:40.659547] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:12:52.658 [2024-11-26 23:46:40.659603] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:12:52.658 [2024-11-26 23:46:40.659760] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:12:52.658 [2024-11-26 23:46:40.659776] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:12:52.658 [2024-11-26 23:46:40.660037] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:12:52.658 [2024-11-26 23:46:40.660478] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:12:52.658 [2024-11-26 23:46:40.660497] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:12:52.658 [2024-11-26 23:46:40.660613] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:52.658 23:46:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:52.658 23:46:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:12:52.658 23:46:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:52.658 23:46:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:52.658 23:46:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:12:52.658 23:46:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:12:52.658 23:46:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:52.658 23:46:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:52.658 23:46:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:52.658 23:46:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:52.658 23:46:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:52.658 23:46:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:52.658 23:46:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:52.658 23:46:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:52.658 23:46:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:52.658 23:46:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:52.658 23:46:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:52.658 "name": "raid_bdev1", 00:12:52.658 "uuid": "15cf1b13-20c0-43d0-be7b-47a73e2c0de8", 00:12:52.658 "strip_size_kb": 64, 00:12:52.658 "state": "online", 00:12:52.658 "raid_level": "raid5f", 00:12:52.658 "superblock": true, 00:12:52.658 "num_base_bdevs": 3, 00:12:52.658 "num_base_bdevs_discovered": 3, 00:12:52.658 "num_base_bdevs_operational": 3, 00:12:52.658 "base_bdevs_list": [ 00:12:52.658 { 00:12:52.658 "name": "pt1", 00:12:52.658 "uuid": "00000000-0000-0000-0000-000000000001", 00:12:52.658 "is_configured": true, 00:12:52.658 "data_offset": 2048, 00:12:52.658 "data_size": 63488 00:12:52.658 }, 00:12:52.658 { 00:12:52.658 "name": "pt2", 00:12:52.658 "uuid": "00000000-0000-0000-0000-000000000002", 00:12:52.658 "is_configured": true, 00:12:52.658 "data_offset": 2048, 00:12:52.658 "data_size": 63488 00:12:52.658 }, 00:12:52.658 { 00:12:52.658 "name": "pt3", 00:12:52.658 "uuid": "00000000-0000-0000-0000-000000000003", 00:12:52.658 "is_configured": true, 00:12:52.658 "data_offset": 2048, 00:12:52.658 "data_size": 63488 00:12:52.658 } 00:12:52.658 ] 00:12:52.658 }' 00:12:52.658 23:46:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:52.658 23:46:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:53.311 23:46:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:12:53.311 23:46:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:12:53.312 23:46:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:12:53.312 23:46:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:12:53.312 23:46:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:12:53.312 23:46:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:12:53.312 23:46:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:12:53.312 23:46:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:53.312 23:46:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:53.312 23:46:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:12:53.312 [2024-11-26 23:46:41.101376] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:12:53.312 23:46:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:53.312 23:46:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:12:53.312 "name": "raid_bdev1", 00:12:53.312 "aliases": [ 00:12:53.312 "15cf1b13-20c0-43d0-be7b-47a73e2c0de8" 00:12:53.312 ], 00:12:53.312 "product_name": "Raid Volume", 00:12:53.312 "block_size": 512, 00:12:53.312 "num_blocks": 126976, 00:12:53.312 "uuid": "15cf1b13-20c0-43d0-be7b-47a73e2c0de8", 00:12:53.312 "assigned_rate_limits": { 00:12:53.312 "rw_ios_per_sec": 0, 00:12:53.312 "rw_mbytes_per_sec": 0, 00:12:53.312 "r_mbytes_per_sec": 0, 00:12:53.312 "w_mbytes_per_sec": 0 00:12:53.312 }, 00:12:53.312 "claimed": false, 00:12:53.312 "zoned": false, 00:12:53.312 "supported_io_types": { 00:12:53.312 "read": true, 00:12:53.312 "write": true, 00:12:53.312 "unmap": false, 00:12:53.312 "flush": false, 00:12:53.312 "reset": true, 00:12:53.312 "nvme_admin": false, 00:12:53.312 "nvme_io": false, 00:12:53.312 "nvme_io_md": false, 00:12:53.312 "write_zeroes": true, 00:12:53.312 "zcopy": false, 00:12:53.312 "get_zone_info": false, 00:12:53.312 "zone_management": false, 00:12:53.312 "zone_append": false, 00:12:53.312 "compare": false, 00:12:53.312 "compare_and_write": false, 00:12:53.312 "abort": false, 00:12:53.312 "seek_hole": false, 00:12:53.312 "seek_data": false, 00:12:53.312 "copy": false, 00:12:53.312 "nvme_iov_md": false 00:12:53.312 }, 00:12:53.312 "driver_specific": { 00:12:53.312 "raid": { 00:12:53.312 "uuid": "15cf1b13-20c0-43d0-be7b-47a73e2c0de8", 00:12:53.312 "strip_size_kb": 64, 00:12:53.312 "state": "online", 00:12:53.312 "raid_level": "raid5f", 00:12:53.312 "superblock": true, 00:12:53.312 "num_base_bdevs": 3, 00:12:53.312 "num_base_bdevs_discovered": 3, 00:12:53.312 "num_base_bdevs_operational": 3, 00:12:53.312 "base_bdevs_list": [ 00:12:53.312 { 00:12:53.312 "name": "pt1", 00:12:53.312 "uuid": "00000000-0000-0000-0000-000000000001", 00:12:53.312 "is_configured": true, 00:12:53.312 "data_offset": 2048, 00:12:53.312 "data_size": 63488 00:12:53.312 }, 00:12:53.312 { 00:12:53.312 "name": "pt2", 00:12:53.312 "uuid": "00000000-0000-0000-0000-000000000002", 00:12:53.312 "is_configured": true, 00:12:53.312 "data_offset": 2048, 00:12:53.312 "data_size": 63488 00:12:53.312 }, 00:12:53.312 { 00:12:53.312 "name": "pt3", 00:12:53.312 "uuid": "00000000-0000-0000-0000-000000000003", 00:12:53.312 "is_configured": true, 00:12:53.312 "data_offset": 2048, 00:12:53.312 "data_size": 63488 00:12:53.312 } 00:12:53.312 ] 00:12:53.312 } 00:12:53.312 } 00:12:53.312 }' 00:12:53.312 23:46:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:12:53.312 23:46:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:12:53.312 pt2 00:12:53.312 pt3' 00:12:53.312 23:46:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:12:53.312 23:46:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:12:53.312 23:46:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:12:53.312 23:46:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:12:53.312 23:46:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:53.312 23:46:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:12:53.312 23:46:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:53.312 23:46:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:53.312 23:46:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:12:53.312 23:46:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:12:53.312 23:46:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:12:53.312 23:46:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:12:53.312 23:46:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:12:53.312 23:46:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:53.312 23:46:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:53.312 23:46:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:53.312 23:46:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:12:53.312 23:46:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:12:53.312 23:46:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:12:53.312 23:46:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:12:53.313 23:46:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:12:53.313 23:46:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:53.313 23:46:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:53.313 23:46:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:53.313 23:46:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:12:53.313 23:46:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:12:53.313 23:46:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:12:53.313 23:46:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:12:53.313 23:46:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:53.313 23:46:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:53.313 [2024-11-26 23:46:41.348910] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:12:53.313 23:46:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:53.313 23:46:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=15cf1b13-20c0-43d0-be7b-47a73e2c0de8 00:12:53.313 23:46:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z 15cf1b13-20c0-43d0-be7b-47a73e2c0de8 ']' 00:12:53.313 23:46:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:12:53.313 23:46:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:53.313 23:46:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:53.313 [2024-11-26 23:46:41.392657] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:12:53.313 [2024-11-26 23:46:41.392718] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:12:53.313 [2024-11-26 23:46:41.392828] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:12:53.313 [2024-11-26 23:46:41.392923] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:12:53.313 [2024-11-26 23:46:41.392966] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:12:53.313 23:46:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:53.313 23:46:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:12:53.313 23:46:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:53.313 23:46:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:53.313 23:46:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:53.313 23:46:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:53.313 23:46:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:12:53.313 23:46:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:12:53.313 23:46:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:12:53.313 23:46:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:12:53.313 23:46:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:53.313 23:46:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:53.313 23:46:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:53.313 23:46:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:12:53.313 23:46:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:12:53.313 23:46:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:53.313 23:46:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:53.574 23:46:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:53.574 23:46:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:12:53.574 23:46:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt3 00:12:53.574 23:46:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:53.574 23:46:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:53.574 23:46:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:53.574 23:46:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:12:53.574 23:46:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:12:53.574 23:46:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:53.574 23:46:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:53.574 23:46:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:53.574 23:46:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:12:53.574 23:46:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:12:53.574 23:46:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@652 -- # local es=0 00:12:53.574 23:46:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:12:53.574 23:46:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:12:53.574 23:46:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:12:53.574 23:46:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:12:53.575 23:46:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:12:53.575 23:46:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:12:53.575 23:46:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:53.575 23:46:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:53.575 [2024-11-26 23:46:41.504495] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:12:53.575 [2024-11-26 23:46:41.506383] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:12:53.575 [2024-11-26 23:46:41.506432] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc3 is claimed 00:12:53.575 [2024-11-26 23:46:41.506482] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:12:53.575 [2024-11-26 23:46:41.506529] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:12:53.575 [2024-11-26 23:46:41.506549] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc3 00:12:53.575 [2024-11-26 23:46:41.506562] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:12:53.575 [2024-11-26 23:46:41.506592] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state configuring 00:12:53.575 request: 00:12:53.575 { 00:12:53.575 "name": "raid_bdev1", 00:12:53.575 "raid_level": "raid5f", 00:12:53.575 "base_bdevs": [ 00:12:53.575 "malloc1", 00:12:53.575 "malloc2", 00:12:53.575 "malloc3" 00:12:53.575 ], 00:12:53.575 "strip_size_kb": 64, 00:12:53.575 "superblock": false, 00:12:53.575 "method": "bdev_raid_create", 00:12:53.575 "req_id": 1 00:12:53.575 } 00:12:53.575 Got JSON-RPC error response 00:12:53.575 response: 00:12:53.575 { 00:12:53.575 "code": -17, 00:12:53.575 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:12:53.575 } 00:12:53.575 23:46:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:12:53.575 23:46:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@655 -- # es=1 00:12:53.575 23:46:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:12:53.575 23:46:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:12:53.575 23:46:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:12:53.575 23:46:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:53.575 23:46:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:53.575 23:46:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:53.575 23:46:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:12:53.575 23:46:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:53.575 23:46:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:12:53.575 23:46:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:12:53.575 23:46:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:12:53.575 23:46:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:53.575 23:46:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:53.575 [2024-11-26 23:46:41.568352] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:12:53.575 [2024-11-26 23:46:41.568396] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:53.575 [2024-11-26 23:46:41.568426] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:12:53.575 [2024-11-26 23:46:41.568436] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:53.575 [2024-11-26 23:46:41.570479] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:53.575 [2024-11-26 23:46:41.570516] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:12:53.575 [2024-11-26 23:46:41.570592] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:12:53.575 [2024-11-26 23:46:41.570644] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:12:53.575 pt1 00:12:53.575 23:46:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:53.575 23:46:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid5f 64 3 00:12:53.575 23:46:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:53.575 23:46:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:12:53.575 23:46:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:12:53.575 23:46:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:12:53.575 23:46:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:53.575 23:46:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:53.575 23:46:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:53.575 23:46:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:53.575 23:46:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:53.575 23:46:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:53.575 23:46:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:53.575 23:46:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:53.575 23:46:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:53.575 23:46:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:53.575 23:46:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:53.575 "name": "raid_bdev1", 00:12:53.575 "uuid": "15cf1b13-20c0-43d0-be7b-47a73e2c0de8", 00:12:53.575 "strip_size_kb": 64, 00:12:53.575 "state": "configuring", 00:12:53.575 "raid_level": "raid5f", 00:12:53.575 "superblock": true, 00:12:53.575 "num_base_bdevs": 3, 00:12:53.575 "num_base_bdevs_discovered": 1, 00:12:53.575 "num_base_bdevs_operational": 3, 00:12:53.575 "base_bdevs_list": [ 00:12:53.575 { 00:12:53.575 "name": "pt1", 00:12:53.575 "uuid": "00000000-0000-0000-0000-000000000001", 00:12:53.575 "is_configured": true, 00:12:53.575 "data_offset": 2048, 00:12:53.575 "data_size": 63488 00:12:53.575 }, 00:12:53.575 { 00:12:53.575 "name": null, 00:12:53.575 "uuid": "00000000-0000-0000-0000-000000000002", 00:12:53.575 "is_configured": false, 00:12:53.575 "data_offset": 2048, 00:12:53.575 "data_size": 63488 00:12:53.575 }, 00:12:53.575 { 00:12:53.575 "name": null, 00:12:53.575 "uuid": "00000000-0000-0000-0000-000000000003", 00:12:53.575 "is_configured": false, 00:12:53.575 "data_offset": 2048, 00:12:53.575 "data_size": 63488 00:12:53.575 } 00:12:53.575 ] 00:12:53.575 }' 00:12:53.575 23:46:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:53.575 23:46:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:54.146 23:46:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 3 -gt 2 ']' 00:12:54.146 23:46:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@472 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:12:54.146 23:46:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:54.146 23:46:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:54.146 [2024-11-26 23:46:41.999615] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:12:54.146 [2024-11-26 23:46:41.999674] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:54.146 [2024-11-26 23:46:41.999694] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:12:54.146 [2024-11-26 23:46:41.999706] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:54.146 [2024-11-26 23:46:42.000091] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:54.146 [2024-11-26 23:46:42.000123] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:12:54.146 [2024-11-26 23:46:42.000189] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:12:54.146 [2024-11-26 23:46:42.000219] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:12:54.146 pt2 00:12:54.146 23:46:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:54.146 23:46:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@473 -- # rpc_cmd bdev_passthru_delete pt2 00:12:54.146 23:46:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:54.146 23:46:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:54.146 [2024-11-26 23:46:42.011604] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt2 00:12:54.146 23:46:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:54.146 23:46:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@474 -- # verify_raid_bdev_state raid_bdev1 configuring raid5f 64 3 00:12:54.146 23:46:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:54.146 23:46:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:12:54.146 23:46:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:12:54.146 23:46:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:12:54.146 23:46:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:54.146 23:46:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:54.146 23:46:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:54.146 23:46:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:54.146 23:46:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:54.146 23:46:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:54.146 23:46:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:54.146 23:46:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:54.146 23:46:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:54.146 23:46:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:54.146 23:46:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:54.146 "name": "raid_bdev1", 00:12:54.146 "uuid": "15cf1b13-20c0-43d0-be7b-47a73e2c0de8", 00:12:54.146 "strip_size_kb": 64, 00:12:54.146 "state": "configuring", 00:12:54.146 "raid_level": "raid5f", 00:12:54.146 "superblock": true, 00:12:54.146 "num_base_bdevs": 3, 00:12:54.146 "num_base_bdevs_discovered": 1, 00:12:54.146 "num_base_bdevs_operational": 3, 00:12:54.146 "base_bdevs_list": [ 00:12:54.146 { 00:12:54.146 "name": "pt1", 00:12:54.146 "uuid": "00000000-0000-0000-0000-000000000001", 00:12:54.146 "is_configured": true, 00:12:54.146 "data_offset": 2048, 00:12:54.146 "data_size": 63488 00:12:54.146 }, 00:12:54.146 { 00:12:54.146 "name": null, 00:12:54.146 "uuid": "00000000-0000-0000-0000-000000000002", 00:12:54.146 "is_configured": false, 00:12:54.146 "data_offset": 0, 00:12:54.146 "data_size": 63488 00:12:54.146 }, 00:12:54.146 { 00:12:54.146 "name": null, 00:12:54.146 "uuid": "00000000-0000-0000-0000-000000000003", 00:12:54.146 "is_configured": false, 00:12:54.146 "data_offset": 2048, 00:12:54.146 "data_size": 63488 00:12:54.146 } 00:12:54.146 ] 00:12:54.146 }' 00:12:54.146 23:46:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:54.146 23:46:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:54.407 23:46:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:12:54.407 23:46:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:12:54.407 23:46:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:12:54.407 23:46:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:54.407 23:46:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:54.407 [2024-11-26 23:46:42.398929] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:12:54.407 [2024-11-26 23:46:42.398982] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:54.407 [2024-11-26 23:46:42.399015] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:12:54.407 [2024-11-26 23:46:42.399023] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:54.407 [2024-11-26 23:46:42.399395] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:54.407 [2024-11-26 23:46:42.399420] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:12:54.407 [2024-11-26 23:46:42.399487] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:12:54.407 [2024-11-26 23:46:42.399506] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:12:54.407 pt2 00:12:54.407 23:46:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:54.407 23:46:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:12:54.407 23:46:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:12:54.407 23:46:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:12:54.407 23:46:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:54.407 23:46:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:54.407 [2024-11-26 23:46:42.410905] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:12:54.407 [2024-11-26 23:46:42.410949] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:54.407 [2024-11-26 23:46:42.410966] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:12:54.407 [2024-11-26 23:46:42.410973] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:54.407 [2024-11-26 23:46:42.411291] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:54.407 [2024-11-26 23:46:42.411315] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:12:54.407 [2024-11-26 23:46:42.411379] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:12:54.407 [2024-11-26 23:46:42.411397] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:12:54.407 [2024-11-26 23:46:42.411490] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:12:54.407 [2024-11-26 23:46:42.411516] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:12:54.407 [2024-11-26 23:46:42.411738] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:12:54.407 [2024-11-26 23:46:42.412111] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:12:54.407 [2024-11-26 23:46:42.412131] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:12:54.407 [2024-11-26 23:46:42.412228] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:54.407 pt3 00:12:54.407 23:46:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:54.407 23:46:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:12:54.407 23:46:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:12:54.407 23:46:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:12:54.407 23:46:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:54.407 23:46:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:54.407 23:46:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:12:54.407 23:46:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:12:54.407 23:46:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:54.407 23:46:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:54.407 23:46:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:54.407 23:46:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:54.407 23:46:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:54.407 23:46:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:54.407 23:46:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:54.407 23:46:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:54.407 23:46:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:54.407 23:46:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:54.407 23:46:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:54.407 "name": "raid_bdev1", 00:12:54.407 "uuid": "15cf1b13-20c0-43d0-be7b-47a73e2c0de8", 00:12:54.407 "strip_size_kb": 64, 00:12:54.407 "state": "online", 00:12:54.407 "raid_level": "raid5f", 00:12:54.407 "superblock": true, 00:12:54.407 "num_base_bdevs": 3, 00:12:54.407 "num_base_bdevs_discovered": 3, 00:12:54.407 "num_base_bdevs_operational": 3, 00:12:54.407 "base_bdevs_list": [ 00:12:54.407 { 00:12:54.407 "name": "pt1", 00:12:54.407 "uuid": "00000000-0000-0000-0000-000000000001", 00:12:54.407 "is_configured": true, 00:12:54.407 "data_offset": 2048, 00:12:54.407 "data_size": 63488 00:12:54.407 }, 00:12:54.407 { 00:12:54.407 "name": "pt2", 00:12:54.407 "uuid": "00000000-0000-0000-0000-000000000002", 00:12:54.407 "is_configured": true, 00:12:54.407 "data_offset": 2048, 00:12:54.407 "data_size": 63488 00:12:54.407 }, 00:12:54.407 { 00:12:54.407 "name": "pt3", 00:12:54.407 "uuid": "00000000-0000-0000-0000-000000000003", 00:12:54.407 "is_configured": true, 00:12:54.407 "data_offset": 2048, 00:12:54.407 "data_size": 63488 00:12:54.407 } 00:12:54.407 ] 00:12:54.407 }' 00:12:54.407 23:46:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:54.407 23:46:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:54.977 23:46:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:12:54.977 23:46:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:12:54.977 23:46:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:12:54.977 23:46:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:12:54.977 23:46:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:12:54.977 23:46:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:12:54.977 23:46:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:12:54.977 23:46:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:12:54.977 23:46:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:54.977 23:46:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:54.977 [2024-11-26 23:46:42.850659] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:12:54.977 23:46:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:54.977 23:46:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:12:54.977 "name": "raid_bdev1", 00:12:54.977 "aliases": [ 00:12:54.977 "15cf1b13-20c0-43d0-be7b-47a73e2c0de8" 00:12:54.977 ], 00:12:54.977 "product_name": "Raid Volume", 00:12:54.977 "block_size": 512, 00:12:54.977 "num_blocks": 126976, 00:12:54.977 "uuid": "15cf1b13-20c0-43d0-be7b-47a73e2c0de8", 00:12:54.977 "assigned_rate_limits": { 00:12:54.977 "rw_ios_per_sec": 0, 00:12:54.977 "rw_mbytes_per_sec": 0, 00:12:54.977 "r_mbytes_per_sec": 0, 00:12:54.977 "w_mbytes_per_sec": 0 00:12:54.977 }, 00:12:54.977 "claimed": false, 00:12:54.977 "zoned": false, 00:12:54.977 "supported_io_types": { 00:12:54.977 "read": true, 00:12:54.977 "write": true, 00:12:54.977 "unmap": false, 00:12:54.977 "flush": false, 00:12:54.977 "reset": true, 00:12:54.977 "nvme_admin": false, 00:12:54.977 "nvme_io": false, 00:12:54.977 "nvme_io_md": false, 00:12:54.977 "write_zeroes": true, 00:12:54.977 "zcopy": false, 00:12:54.977 "get_zone_info": false, 00:12:54.977 "zone_management": false, 00:12:54.977 "zone_append": false, 00:12:54.977 "compare": false, 00:12:54.977 "compare_and_write": false, 00:12:54.977 "abort": false, 00:12:54.977 "seek_hole": false, 00:12:54.977 "seek_data": false, 00:12:54.977 "copy": false, 00:12:54.978 "nvme_iov_md": false 00:12:54.978 }, 00:12:54.978 "driver_specific": { 00:12:54.978 "raid": { 00:12:54.978 "uuid": "15cf1b13-20c0-43d0-be7b-47a73e2c0de8", 00:12:54.978 "strip_size_kb": 64, 00:12:54.978 "state": "online", 00:12:54.978 "raid_level": "raid5f", 00:12:54.978 "superblock": true, 00:12:54.978 "num_base_bdevs": 3, 00:12:54.978 "num_base_bdevs_discovered": 3, 00:12:54.978 "num_base_bdevs_operational": 3, 00:12:54.978 "base_bdevs_list": [ 00:12:54.978 { 00:12:54.978 "name": "pt1", 00:12:54.978 "uuid": "00000000-0000-0000-0000-000000000001", 00:12:54.978 "is_configured": true, 00:12:54.978 "data_offset": 2048, 00:12:54.978 "data_size": 63488 00:12:54.978 }, 00:12:54.978 { 00:12:54.978 "name": "pt2", 00:12:54.978 "uuid": "00000000-0000-0000-0000-000000000002", 00:12:54.978 "is_configured": true, 00:12:54.978 "data_offset": 2048, 00:12:54.978 "data_size": 63488 00:12:54.978 }, 00:12:54.978 { 00:12:54.978 "name": "pt3", 00:12:54.978 "uuid": "00000000-0000-0000-0000-000000000003", 00:12:54.978 "is_configured": true, 00:12:54.978 "data_offset": 2048, 00:12:54.978 "data_size": 63488 00:12:54.978 } 00:12:54.978 ] 00:12:54.978 } 00:12:54.978 } 00:12:54.978 }' 00:12:54.978 23:46:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:12:54.978 23:46:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:12:54.978 pt2 00:12:54.978 pt3' 00:12:54.978 23:46:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:12:54.978 23:46:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:12:54.978 23:46:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:12:54.978 23:46:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:12:54.978 23:46:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:12:54.978 23:46:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:54.978 23:46:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:54.978 23:46:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:54.978 23:46:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:12:54.978 23:46:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:12:54.978 23:46:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:12:54.978 23:46:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:12:54.978 23:46:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:12:54.978 23:46:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:54.978 23:46:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:54.978 23:46:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:54.978 23:46:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:12:54.978 23:46:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:12:54.978 23:46:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:12:54.978 23:46:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:12:54.978 23:46:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:12:54.978 23:46:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:54.978 23:46:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:54.978 23:46:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:55.237 23:46:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:12:55.237 23:46:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:12:55.237 23:46:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:12:55.237 23:46:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:12:55.237 23:46:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:55.237 23:46:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:55.237 [2024-11-26 23:46:43.118592] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:12:55.237 23:46:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:55.237 23:46:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' 15cf1b13-20c0-43d0-be7b-47a73e2c0de8 '!=' 15cf1b13-20c0-43d0-be7b-47a73e2c0de8 ']' 00:12:55.237 23:46:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy raid5f 00:12:55.237 23:46:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:12:55.237 23:46:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@199 -- # return 0 00:12:55.237 23:46:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@493 -- # rpc_cmd bdev_passthru_delete pt1 00:12:55.237 23:46:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:55.237 23:46:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:55.237 [2024-11-26 23:46:43.154475] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt1 00:12:55.237 23:46:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:55.237 23:46:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@496 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:12:55.237 23:46:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:55.237 23:46:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:55.237 23:46:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:12:55.237 23:46:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:12:55.237 23:46:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:12:55.237 23:46:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:55.237 23:46:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:55.237 23:46:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:55.237 23:46:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:55.237 23:46:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:55.237 23:46:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:55.237 23:46:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:55.237 23:46:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:55.237 23:46:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:55.237 23:46:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:55.237 "name": "raid_bdev1", 00:12:55.237 "uuid": "15cf1b13-20c0-43d0-be7b-47a73e2c0de8", 00:12:55.237 "strip_size_kb": 64, 00:12:55.237 "state": "online", 00:12:55.237 "raid_level": "raid5f", 00:12:55.237 "superblock": true, 00:12:55.237 "num_base_bdevs": 3, 00:12:55.237 "num_base_bdevs_discovered": 2, 00:12:55.237 "num_base_bdevs_operational": 2, 00:12:55.237 "base_bdevs_list": [ 00:12:55.237 { 00:12:55.237 "name": null, 00:12:55.237 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:55.237 "is_configured": false, 00:12:55.237 "data_offset": 0, 00:12:55.237 "data_size": 63488 00:12:55.237 }, 00:12:55.237 { 00:12:55.237 "name": "pt2", 00:12:55.237 "uuid": "00000000-0000-0000-0000-000000000002", 00:12:55.237 "is_configured": true, 00:12:55.237 "data_offset": 2048, 00:12:55.237 "data_size": 63488 00:12:55.237 }, 00:12:55.237 { 00:12:55.237 "name": "pt3", 00:12:55.237 "uuid": "00000000-0000-0000-0000-000000000003", 00:12:55.237 "is_configured": true, 00:12:55.237 "data_offset": 2048, 00:12:55.237 "data_size": 63488 00:12:55.237 } 00:12:55.237 ] 00:12:55.237 }' 00:12:55.237 23:46:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:55.237 23:46:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:55.496 23:46:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@499 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:12:55.496 23:46:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:55.496 23:46:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:55.496 [2024-11-26 23:46:43.614443] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:12:55.496 [2024-11-26 23:46:43.614473] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:12:55.496 [2024-11-26 23:46:43.614531] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:12:55.496 [2024-11-26 23:46:43.614584] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:12:55.496 [2024-11-26 23:46:43.614592] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:12:55.496 23:46:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:55.496 23:46:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@500 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:55.496 23:46:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@500 -- # jq -r '.[]' 00:12:55.496 23:46:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:55.496 23:46:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:55.756 23:46:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:55.756 23:46:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@500 -- # raid_bdev= 00:12:55.756 23:46:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@501 -- # '[' -n '' ']' 00:12:55.756 23:46:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i = 1 )) 00:12:55.756 23:46:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:12:55.756 23:46:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt2 00:12:55.756 23:46:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:55.756 23:46:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:55.756 23:46:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:55.756 23:46:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:12:55.756 23:46:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:12:55.756 23:46:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt3 00:12:55.756 23:46:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:55.756 23:46:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:55.756 23:46:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:55.756 23:46:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:12:55.756 23:46:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:12:55.756 23:46:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i = 1 )) 00:12:55.756 23:46:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:12:55.756 23:46:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@512 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:12:55.756 23:46:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:55.756 23:46:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:55.756 [2024-11-26 23:46:43.694456] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:12:55.756 [2024-11-26 23:46:43.694500] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:55.756 [2024-11-26 23:46:43.694517] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009980 00:12:55.756 [2024-11-26 23:46:43.694525] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:55.756 [2024-11-26 23:46:43.696657] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:55.756 [2024-11-26 23:46:43.696689] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:12:55.756 [2024-11-26 23:46:43.696772] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:12:55.756 [2024-11-26 23:46:43.696801] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:12:55.756 pt2 00:12:55.756 23:46:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:55.756 23:46:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@515 -- # verify_raid_bdev_state raid_bdev1 configuring raid5f 64 2 00:12:55.757 23:46:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:55.757 23:46:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:12:55.757 23:46:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:12:55.757 23:46:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:12:55.757 23:46:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:12:55.757 23:46:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:55.757 23:46:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:55.757 23:46:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:55.757 23:46:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:55.757 23:46:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:55.757 23:46:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:55.757 23:46:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:55.757 23:46:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:55.757 23:46:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:55.757 23:46:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:55.757 "name": "raid_bdev1", 00:12:55.757 "uuid": "15cf1b13-20c0-43d0-be7b-47a73e2c0de8", 00:12:55.757 "strip_size_kb": 64, 00:12:55.757 "state": "configuring", 00:12:55.757 "raid_level": "raid5f", 00:12:55.757 "superblock": true, 00:12:55.757 "num_base_bdevs": 3, 00:12:55.757 "num_base_bdevs_discovered": 1, 00:12:55.757 "num_base_bdevs_operational": 2, 00:12:55.757 "base_bdevs_list": [ 00:12:55.757 { 00:12:55.757 "name": null, 00:12:55.757 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:55.757 "is_configured": false, 00:12:55.757 "data_offset": 2048, 00:12:55.757 "data_size": 63488 00:12:55.757 }, 00:12:55.757 { 00:12:55.757 "name": "pt2", 00:12:55.757 "uuid": "00000000-0000-0000-0000-000000000002", 00:12:55.757 "is_configured": true, 00:12:55.757 "data_offset": 2048, 00:12:55.757 "data_size": 63488 00:12:55.757 }, 00:12:55.757 { 00:12:55.757 "name": null, 00:12:55.757 "uuid": "00000000-0000-0000-0000-000000000003", 00:12:55.757 "is_configured": false, 00:12:55.757 "data_offset": 2048, 00:12:55.757 "data_size": 63488 00:12:55.757 } 00:12:55.757 ] 00:12:55.757 }' 00:12:55.757 23:46:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:55.757 23:46:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:56.017 23:46:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i++ )) 00:12:56.017 23:46:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:12:56.017 23:46:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@519 -- # i=2 00:12:56.017 23:46:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@520 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:12:56.017 23:46:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:56.017 23:46:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:56.017 [2024-11-26 23:46:44.094466] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:12:56.017 [2024-11-26 23:46:44.094518] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:56.017 [2024-11-26 23:46:44.094542] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009f80 00:12:56.017 [2024-11-26 23:46:44.094552] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:56.017 [2024-11-26 23:46:44.094902] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:56.017 [2024-11-26 23:46:44.094925] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:12:56.017 [2024-11-26 23:46:44.094987] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:12:56.017 [2024-11-26 23:46:44.095010] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:12:56.017 [2024-11-26 23:46:44.095094] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:12:56.017 [2024-11-26 23:46:44.095108] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:12:56.017 [2024-11-26 23:46:44.095335] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:12:56.017 [2024-11-26 23:46:44.095817] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:12:56.017 [2024-11-26 23:46:44.095841] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001c80 00:12:56.017 [2024-11-26 23:46:44.096055] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:56.017 pt3 00:12:56.017 23:46:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:56.017 23:46:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@523 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:12:56.017 23:46:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:56.017 23:46:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:56.017 23:46:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:12:56.017 23:46:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:12:56.017 23:46:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:12:56.017 23:46:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:56.017 23:46:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:56.017 23:46:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:56.017 23:46:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:56.017 23:46:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:56.017 23:46:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:56.017 23:46:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:56.017 23:46:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:56.017 23:46:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:56.277 23:46:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:56.277 "name": "raid_bdev1", 00:12:56.277 "uuid": "15cf1b13-20c0-43d0-be7b-47a73e2c0de8", 00:12:56.277 "strip_size_kb": 64, 00:12:56.277 "state": "online", 00:12:56.277 "raid_level": "raid5f", 00:12:56.277 "superblock": true, 00:12:56.277 "num_base_bdevs": 3, 00:12:56.277 "num_base_bdevs_discovered": 2, 00:12:56.277 "num_base_bdevs_operational": 2, 00:12:56.277 "base_bdevs_list": [ 00:12:56.277 { 00:12:56.277 "name": null, 00:12:56.277 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:56.277 "is_configured": false, 00:12:56.277 "data_offset": 2048, 00:12:56.277 "data_size": 63488 00:12:56.277 }, 00:12:56.277 { 00:12:56.277 "name": "pt2", 00:12:56.277 "uuid": "00000000-0000-0000-0000-000000000002", 00:12:56.277 "is_configured": true, 00:12:56.277 "data_offset": 2048, 00:12:56.277 "data_size": 63488 00:12:56.277 }, 00:12:56.277 { 00:12:56.277 "name": "pt3", 00:12:56.277 "uuid": "00000000-0000-0000-0000-000000000003", 00:12:56.277 "is_configured": true, 00:12:56.277 "data_offset": 2048, 00:12:56.277 "data_size": 63488 00:12:56.277 } 00:12:56.277 ] 00:12:56.277 }' 00:12:56.277 23:46:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:56.277 23:46:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:56.537 23:46:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@526 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:12:56.537 23:46:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:56.537 23:46:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:56.537 [2024-11-26 23:46:44.542447] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:12:56.537 [2024-11-26 23:46:44.542476] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:12:56.537 [2024-11-26 23:46:44.542539] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:12:56.537 [2024-11-26 23:46:44.542591] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:12:56.537 [2024-11-26 23:46:44.542602] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name raid_bdev1, state offline 00:12:56.537 23:46:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:56.537 23:46:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@527 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:56.537 23:46:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:56.537 23:46:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:56.537 23:46:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@527 -- # jq -r '.[]' 00:12:56.537 23:46:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:56.537 23:46:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@527 -- # raid_bdev= 00:12:56.537 23:46:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@528 -- # '[' -n '' ']' 00:12:56.537 23:46:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@532 -- # '[' 3 -gt 2 ']' 00:12:56.537 23:46:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@534 -- # i=2 00:12:56.537 23:46:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@535 -- # rpc_cmd bdev_passthru_delete pt3 00:12:56.537 23:46:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:56.537 23:46:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:56.537 23:46:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:56.537 23:46:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@540 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:12:56.537 23:46:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:56.537 23:46:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:56.537 [2024-11-26 23:46:44.614461] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:12:56.537 [2024-11-26 23:46:44.614514] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:56.537 [2024-11-26 23:46:44.614529] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:12:56.537 [2024-11-26 23:46:44.614540] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:56.537 [2024-11-26 23:46:44.616640] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:56.537 [2024-11-26 23:46:44.616677] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:12:56.537 [2024-11-26 23:46:44.616736] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:12:56.537 [2024-11-26 23:46:44.616769] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:12:56.537 [2024-11-26 23:46:44.616862] bdev_raid.c:3685:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev pt2 (4) greater than existing raid bdev raid_bdev1 (2) 00:12:56.537 [2024-11-26 23:46:44.616878] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:12:56.537 [2024-11-26 23:46:44.616892] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002000 name raid_bdev1, state configuring 00:12:56.537 [2024-11-26 23:46:44.616937] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:12:56.537 pt1 00:12:56.537 23:46:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:56.537 23:46:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@542 -- # '[' 3 -gt 2 ']' 00:12:56.537 23:46:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@545 -- # verify_raid_bdev_state raid_bdev1 configuring raid5f 64 2 00:12:56.537 23:46:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:56.537 23:46:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:12:56.537 23:46:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:12:56.537 23:46:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:12:56.537 23:46:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:12:56.537 23:46:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:56.537 23:46:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:56.537 23:46:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:56.537 23:46:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:56.537 23:46:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:56.537 23:46:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:56.537 23:46:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:56.537 23:46:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:56.537 23:46:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:56.795 23:46:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:56.795 "name": "raid_bdev1", 00:12:56.795 "uuid": "15cf1b13-20c0-43d0-be7b-47a73e2c0de8", 00:12:56.795 "strip_size_kb": 64, 00:12:56.795 "state": "configuring", 00:12:56.795 "raid_level": "raid5f", 00:12:56.795 "superblock": true, 00:12:56.795 "num_base_bdevs": 3, 00:12:56.795 "num_base_bdevs_discovered": 1, 00:12:56.795 "num_base_bdevs_operational": 2, 00:12:56.795 "base_bdevs_list": [ 00:12:56.795 { 00:12:56.795 "name": null, 00:12:56.795 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:56.795 "is_configured": false, 00:12:56.795 "data_offset": 2048, 00:12:56.795 "data_size": 63488 00:12:56.795 }, 00:12:56.795 { 00:12:56.795 "name": "pt2", 00:12:56.795 "uuid": "00000000-0000-0000-0000-000000000002", 00:12:56.795 "is_configured": true, 00:12:56.795 "data_offset": 2048, 00:12:56.795 "data_size": 63488 00:12:56.795 }, 00:12:56.795 { 00:12:56.795 "name": null, 00:12:56.795 "uuid": "00000000-0000-0000-0000-000000000003", 00:12:56.795 "is_configured": false, 00:12:56.795 "data_offset": 2048, 00:12:56.795 "data_size": 63488 00:12:56.795 } 00:12:56.795 ] 00:12:56.795 }' 00:12:56.795 23:46:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:56.795 23:46:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:57.054 23:46:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@546 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:12:57.054 23:46:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@546 -- # rpc_cmd bdev_raid_get_bdevs configuring 00:12:57.054 23:46:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:57.054 23:46:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:57.054 23:46:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:57.054 23:46:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@546 -- # [[ false == \f\a\l\s\e ]] 00:12:57.054 23:46:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@549 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:12:57.055 23:46:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:57.055 23:46:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:57.055 [2024-11-26 23:46:45.122471] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:12:57.055 [2024-11-26 23:46:45.122537] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:57.055 [2024-11-26 23:46:45.122555] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a880 00:12:57.055 [2024-11-26 23:46:45.122565] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:57.055 [2024-11-26 23:46:45.122980] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:57.055 [2024-11-26 23:46:45.123012] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:12:57.055 [2024-11-26 23:46:45.123086] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:12:57.055 [2024-11-26 23:46:45.123122] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:12:57.055 [2024-11-26 23:46:45.123219] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000002380 00:12:57.055 [2024-11-26 23:46:45.123251] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:12:57.055 [2024-11-26 23:46:45.123521] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000026d0 00:12:57.055 [2024-11-26 23:46:45.124011] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000002380 00:12:57.055 [2024-11-26 23:46:45.124030] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000002380 00:12:57.055 [2024-11-26 23:46:45.124191] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:57.055 pt3 00:12:57.055 23:46:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:57.055 23:46:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@554 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:12:57.055 23:46:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:57.055 23:46:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:57.055 23:46:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:12:57.055 23:46:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:12:57.055 23:46:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:12:57.055 23:46:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:57.055 23:46:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:57.055 23:46:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:57.055 23:46:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:57.055 23:46:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:57.055 23:46:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:57.055 23:46:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:57.055 23:46:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:57.055 23:46:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:57.055 23:46:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:57.055 "name": "raid_bdev1", 00:12:57.055 "uuid": "15cf1b13-20c0-43d0-be7b-47a73e2c0de8", 00:12:57.055 "strip_size_kb": 64, 00:12:57.055 "state": "online", 00:12:57.055 "raid_level": "raid5f", 00:12:57.055 "superblock": true, 00:12:57.055 "num_base_bdevs": 3, 00:12:57.055 "num_base_bdevs_discovered": 2, 00:12:57.055 "num_base_bdevs_operational": 2, 00:12:57.055 "base_bdevs_list": [ 00:12:57.055 { 00:12:57.055 "name": null, 00:12:57.055 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:57.055 "is_configured": false, 00:12:57.055 "data_offset": 2048, 00:12:57.055 "data_size": 63488 00:12:57.055 }, 00:12:57.055 { 00:12:57.055 "name": "pt2", 00:12:57.055 "uuid": "00000000-0000-0000-0000-000000000002", 00:12:57.055 "is_configured": true, 00:12:57.055 "data_offset": 2048, 00:12:57.055 "data_size": 63488 00:12:57.055 }, 00:12:57.055 { 00:12:57.055 "name": "pt3", 00:12:57.055 "uuid": "00000000-0000-0000-0000-000000000003", 00:12:57.055 "is_configured": true, 00:12:57.055 "data_offset": 2048, 00:12:57.055 "data_size": 63488 00:12:57.055 } 00:12:57.055 ] 00:12:57.055 }' 00:12:57.055 23:46:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:57.055 23:46:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:57.625 23:46:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@555 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:12:57.625 23:46:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@555 -- # rpc_cmd bdev_raid_get_bdevs online 00:12:57.625 23:46:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:57.625 23:46:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:57.625 23:46:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:57.625 23:46:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@555 -- # [[ false == \f\a\l\s\e ]] 00:12:57.625 23:46:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@558 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:12:57.625 23:46:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:57.625 23:46:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:57.625 23:46:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@558 -- # jq -r '.[] | .uuid' 00:12:57.625 [2024-11-26 23:46:45.634659] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:12:57.625 23:46:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:57.625 23:46:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@558 -- # '[' 15cf1b13-20c0-43d0-be7b-47a73e2c0de8 '!=' 15cf1b13-20c0-43d0-be7b-47a73e2c0de8 ']' 00:12:57.625 23:46:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 91345 00:12:57.625 23:46:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@954 -- # '[' -z 91345 ']' 00:12:57.625 23:46:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@958 -- # kill -0 91345 00:12:57.625 23:46:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@959 -- # uname 00:12:57.625 23:46:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:12:57.625 23:46:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 91345 00:12:57.625 killing process with pid 91345 00:12:57.625 23:46:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:12:57.625 23:46:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:12:57.625 23:46:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 91345' 00:12:57.625 23:46:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@973 -- # kill 91345 00:12:57.625 [2024-11-26 23:46:45.720780] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:12:57.625 [2024-11-26 23:46:45.720844] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:12:57.625 [2024-11-26 23:46:45.720900] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:12:57.625 [2024-11-26 23:46:45.720909] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002380 name raid_bdev1, state offline 00:12:57.625 23:46:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@978 -- # wait 91345 00:12:57.625 [2024-11-26 23:46:45.753297] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:12:57.886 23:46:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:12:57.886 00:12:57.886 real 0m6.295s 00:12:57.886 user 0m10.583s 00:12:57.886 sys 0m1.361s 00:12:57.886 23:46:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:12:57.886 23:46:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:57.886 ************************************ 00:12:57.886 END TEST raid5f_superblock_test 00:12:57.886 ************************************ 00:12:58.145 23:46:46 bdev_raid -- bdev/bdev_raid.sh@989 -- # '[' true = true ']' 00:12:58.145 23:46:46 bdev_raid -- bdev/bdev_raid.sh@990 -- # run_test raid5f_rebuild_test raid_rebuild_test raid5f 3 false false true 00:12:58.145 23:46:46 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:12:58.145 23:46:46 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:12:58.145 23:46:46 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:12:58.145 ************************************ 00:12:58.145 START TEST raid5f_rebuild_test 00:12:58.145 ************************************ 00:12:58.145 23:46:46 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@1129 -- # raid_rebuild_test raid5f 3 false false true 00:12:58.145 23:46:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@569 -- # local raid_level=raid5f 00:12:58.145 23:46:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=3 00:12:58.145 23:46:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@571 -- # local superblock=false 00:12:58.145 23:46:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:12:58.145 23:46:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@573 -- # local verify=true 00:12:58.145 23:46:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:12:58.145 23:46:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:58.145 23:46:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:12:58.145 23:46:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:58.145 23:46:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:58.145 23:46:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:12:58.145 23:46:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:58.145 23:46:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:58.145 23:46:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev3 00:12:58.145 23:46:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:58.145 23:46:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:58.145 23:46:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:12:58.145 23:46:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:12:58.145 23:46:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:12:58.145 23:46:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@576 -- # local strip_size 00:12:58.145 23:46:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@577 -- # local create_arg 00:12:58.145 23:46:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:12:58.145 23:46:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@579 -- # local data_offset 00:12:58.145 23:46:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@581 -- # '[' raid5f '!=' raid1 ']' 00:12:58.145 23:46:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@582 -- # '[' false = true ']' 00:12:58.145 23:46:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@586 -- # strip_size=64 00:12:58.145 23:46:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@587 -- # create_arg+=' -z 64' 00:12:58.145 23:46:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@592 -- # '[' false = true ']' 00:12:58.145 23:46:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@597 -- # raid_pid=91772 00:12:58.145 23:46:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:12:58.145 23:46:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@598 -- # waitforlisten 91772 00:12:58.145 23:46:46 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@835 -- # '[' -z 91772 ']' 00:12:58.145 23:46:46 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:12:58.145 23:46:46 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:12:58.145 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:12:58.145 23:46:46 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:12:58.145 23:46:46 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:12:58.145 23:46:46 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:58.145 I/O size of 3145728 is greater than zero copy threshold (65536). 00:12:58.146 Zero copy mechanism will not be used. 00:12:58.146 [2024-11-26 23:46:46.130076] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:12:58.146 [2024-11-26 23:46:46.130215] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid91772 ] 00:12:58.406 [2024-11-26 23:46:46.285508] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:12:58.406 [2024-11-26 23:46:46.309842] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:12:58.406 [2024-11-26 23:46:46.352020] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:12:58.406 [2024-11-26 23:46:46.352058] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:12:58.978 23:46:46 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:12:58.978 23:46:46 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@868 -- # return 0 00:12:58.978 23:46:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:12:58.978 23:46:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:12:58.978 23:46:46 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:58.978 23:46:46 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:58.978 BaseBdev1_malloc 00:12:58.978 23:46:46 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:58.978 23:46:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:12:58.978 23:46:46 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:58.978 23:46:46 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:58.978 [2024-11-26 23:46:46.963536] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:12:58.978 [2024-11-26 23:46:46.963589] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:58.978 [2024-11-26 23:46:46.963619] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:12:58.978 [2024-11-26 23:46:46.963633] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:58.978 [2024-11-26 23:46:46.965661] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:58.978 [2024-11-26 23:46:46.965699] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:12:58.978 BaseBdev1 00:12:58.978 23:46:46 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:58.978 23:46:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:12:58.978 23:46:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:12:58.978 23:46:46 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:58.978 23:46:46 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:58.978 BaseBdev2_malloc 00:12:58.978 23:46:46 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:58.978 23:46:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:12:58.978 23:46:46 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:58.978 23:46:46 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:58.978 [2024-11-26 23:46:46.991873] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:12:58.978 [2024-11-26 23:46:46.991924] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:58.978 [2024-11-26 23:46:46.991948] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:12:58.978 [2024-11-26 23:46:46.991956] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:58.978 [2024-11-26 23:46:46.993933] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:58.978 [2024-11-26 23:46:46.993969] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:12:58.978 BaseBdev2 00:12:58.978 23:46:46 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:58.978 23:46:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:12:58.978 23:46:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:12:58.978 23:46:46 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:58.978 23:46:46 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:58.978 BaseBdev3_malloc 00:12:58.978 23:46:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:58.978 23:46:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev3_malloc -p BaseBdev3 00:12:58.978 23:46:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:58.978 23:46:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:58.978 [2024-11-26 23:46:47.020307] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev3_malloc 00:12:58.978 [2024-11-26 23:46:47.020392] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:58.978 [2024-11-26 23:46:47.020416] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:12:58.978 [2024-11-26 23:46:47.020425] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:58.978 [2024-11-26 23:46:47.022436] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:58.978 [2024-11-26 23:46:47.022466] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:12:58.978 BaseBdev3 00:12:58.978 23:46:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:58.978 23:46:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:12:58.978 23:46:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:58.978 23:46:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:58.978 spare_malloc 00:12:58.978 23:46:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:58.978 23:46:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:12:58.978 23:46:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:58.978 23:46:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:58.978 spare_delay 00:12:58.978 23:46:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:58.978 23:46:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:12:58.978 23:46:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:58.978 23:46:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:58.979 [2024-11-26 23:46:47.081117] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:12:58.979 [2024-11-26 23:46:47.081176] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:58.979 [2024-11-26 23:46:47.081204] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:12:58.979 [2024-11-26 23:46:47.081214] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:58.979 [2024-11-26 23:46:47.083549] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:58.979 [2024-11-26 23:46:47.083590] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:12:58.979 spare 00:12:58.979 23:46:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:58.979 23:46:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n raid_bdev1 00:12:58.979 23:46:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:58.979 23:46:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:58.979 [2024-11-26 23:46:47.093134] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:12:58.979 [2024-11-26 23:46:47.094944] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:12:58.979 [2024-11-26 23:46:47.095021] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:12:58.979 [2024-11-26 23:46:47.095100] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:12:58.979 [2024-11-26 23:46:47.095113] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 131072, blocklen 512 00:12:58.979 [2024-11-26 23:46:47.095378] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:12:58.979 [2024-11-26 23:46:47.095765] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:12:58.979 [2024-11-26 23:46:47.095783] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:12:58.979 [2024-11-26 23:46:47.095898] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:58.979 23:46:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:58.979 23:46:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:12:58.979 23:46:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:58.979 23:46:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:58.979 23:46:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:12:58.979 23:46:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:12:58.979 23:46:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:58.979 23:46:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:58.979 23:46:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:58.979 23:46:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:58.979 23:46:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:58.979 23:46:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:58.979 23:46:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:58.979 23:46:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:59.245 23:46:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:59.245 23:46:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:59.245 23:46:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:59.245 "name": "raid_bdev1", 00:12:59.245 "uuid": "f838604d-143f-41dc-a204-8b95e2c8d556", 00:12:59.245 "strip_size_kb": 64, 00:12:59.245 "state": "online", 00:12:59.245 "raid_level": "raid5f", 00:12:59.245 "superblock": false, 00:12:59.245 "num_base_bdevs": 3, 00:12:59.245 "num_base_bdevs_discovered": 3, 00:12:59.245 "num_base_bdevs_operational": 3, 00:12:59.245 "base_bdevs_list": [ 00:12:59.245 { 00:12:59.245 "name": "BaseBdev1", 00:12:59.245 "uuid": "b7315b4a-cc51-5186-8f53-00e8d585b6b2", 00:12:59.245 "is_configured": true, 00:12:59.245 "data_offset": 0, 00:12:59.245 "data_size": 65536 00:12:59.245 }, 00:12:59.245 { 00:12:59.245 "name": "BaseBdev2", 00:12:59.245 "uuid": "4c93a001-9bcf-521b-91e3-47ab5cd76e5e", 00:12:59.245 "is_configured": true, 00:12:59.245 "data_offset": 0, 00:12:59.245 "data_size": 65536 00:12:59.245 }, 00:12:59.245 { 00:12:59.245 "name": "BaseBdev3", 00:12:59.245 "uuid": "32f26f6f-bc03-5aad-a3d0-6487083c18fa", 00:12:59.245 "is_configured": true, 00:12:59.245 "data_offset": 0, 00:12:59.246 "data_size": 65536 00:12:59.246 } 00:12:59.246 ] 00:12:59.246 }' 00:12:59.246 23:46:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:59.246 23:46:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:59.508 23:46:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:12:59.508 23:46:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:59.508 23:46:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:59.508 23:46:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:12:59.508 [2024-11-26 23:46:47.580574] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:12:59.508 23:46:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:59.508 23:46:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=131072 00:12:59.508 23:46:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:12:59.508 23:46:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:59.508 23:46:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:59.508 23:46:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:59.508 23:46:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:59.508 23:46:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@619 -- # data_offset=0 00:12:59.508 23:46:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:12:59.508 23:46:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:12:59.768 23:46:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:12:59.768 23:46:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:12:59.768 23:46:47 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:12:59.768 23:46:47 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:12:59.768 23:46:47 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@10 -- # local bdev_list 00:12:59.768 23:46:47 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:12:59.768 23:46:47 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@11 -- # local nbd_list 00:12:59.768 23:46:47 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@12 -- # local i 00:12:59.768 23:46:47 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:12:59.768 23:46:47 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:12:59.768 23:46:47 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:12:59.768 [2024-11-26 23:46:47.820038] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000026d0 00:12:59.768 /dev/nbd0 00:12:59.768 23:46:47 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:12:59.768 23:46:47 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:12:59.768 23:46:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:12:59.768 23:46:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@873 -- # local i 00:12:59.768 23:46:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:12:59.768 23:46:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:12:59.768 23:46:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:12:59.768 23:46:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@877 -- # break 00:12:59.768 23:46:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:12:59.768 23:46:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:12:59.768 23:46:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:12:59.768 1+0 records in 00:12:59.768 1+0 records out 00:12:59.768 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000343412 s, 11.9 MB/s 00:12:59.768 23:46:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:59.768 23:46:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@890 -- # size=4096 00:12:59.768 23:46:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:00.027 23:46:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:13:00.027 23:46:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@893 -- # return 0 00:13:00.027 23:46:47 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:13:00.027 23:46:47 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:13:00.027 23:46:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@629 -- # '[' raid5f = raid5f ']' 00:13:00.027 23:46:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@630 -- # write_unit_size=256 00:13:00.027 23:46:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@631 -- # echo 128 00:13:00.027 23:46:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=131072 count=512 oflag=direct 00:13:00.287 512+0 records in 00:13:00.287 512+0 records out 00:13:00.287 67108864 bytes (67 MB, 64 MiB) copied, 0.287593 s, 233 MB/s 00:13:00.287 23:46:48 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:13:00.287 23:46:48 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:13:00.287 23:46:48 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:13:00.287 23:46:48 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@50 -- # local nbd_list 00:13:00.287 23:46:48 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@51 -- # local i 00:13:00.287 23:46:48 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:13:00.287 23:46:48 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:13:00.287 23:46:48 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:13:00.287 [2024-11-26 23:46:48.402059] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:00.287 23:46:48 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:13:00.287 23:46:48 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:13:00.287 23:46:48 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:13:00.287 23:46:48 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:13:00.287 23:46:48 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:13:00.287 23:46:48 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:13:00.287 23:46:48 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:13:00.287 23:46:48 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:13:00.287 23:46:48 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:00.287 23:46:48 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:00.546 [2024-11-26 23:46:48.418125] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:13:00.546 23:46:48 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:00.546 23:46:48 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:13:00.546 23:46:48 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:00.546 23:46:48 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:00.546 23:46:48 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:00.546 23:46:48 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:00.546 23:46:48 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:00.546 23:46:48 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:00.546 23:46:48 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:00.546 23:46:48 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:00.546 23:46:48 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:00.546 23:46:48 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:00.546 23:46:48 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:00.546 23:46:48 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:00.546 23:46:48 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:00.546 23:46:48 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:00.546 23:46:48 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:00.546 "name": "raid_bdev1", 00:13:00.546 "uuid": "f838604d-143f-41dc-a204-8b95e2c8d556", 00:13:00.547 "strip_size_kb": 64, 00:13:00.547 "state": "online", 00:13:00.547 "raid_level": "raid5f", 00:13:00.547 "superblock": false, 00:13:00.547 "num_base_bdevs": 3, 00:13:00.547 "num_base_bdevs_discovered": 2, 00:13:00.547 "num_base_bdevs_operational": 2, 00:13:00.547 "base_bdevs_list": [ 00:13:00.547 { 00:13:00.547 "name": null, 00:13:00.547 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:00.547 "is_configured": false, 00:13:00.547 "data_offset": 0, 00:13:00.547 "data_size": 65536 00:13:00.547 }, 00:13:00.547 { 00:13:00.547 "name": "BaseBdev2", 00:13:00.547 "uuid": "4c93a001-9bcf-521b-91e3-47ab5cd76e5e", 00:13:00.547 "is_configured": true, 00:13:00.547 "data_offset": 0, 00:13:00.547 "data_size": 65536 00:13:00.547 }, 00:13:00.547 { 00:13:00.547 "name": "BaseBdev3", 00:13:00.547 "uuid": "32f26f6f-bc03-5aad-a3d0-6487083c18fa", 00:13:00.547 "is_configured": true, 00:13:00.547 "data_offset": 0, 00:13:00.547 "data_size": 65536 00:13:00.547 } 00:13:00.547 ] 00:13:00.547 }' 00:13:00.547 23:46:48 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:00.547 23:46:48 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:00.806 23:46:48 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:13:00.806 23:46:48 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:00.806 23:46:48 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:00.806 [2024-11-26 23:46:48.833395] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:13:00.806 [2024-11-26 23:46:48.838005] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000027cd0 00:13:00.806 23:46:48 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:00.806 23:46:48 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@647 -- # sleep 1 00:13:00.806 [2024-11-26 23:46:48.840122] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:13:01.762 23:46:49 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:01.762 23:46:49 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:01.762 23:46:49 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:01.762 23:46:49 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:01.762 23:46:49 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:01.762 23:46:49 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:01.762 23:46:49 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:01.762 23:46:49 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:01.762 23:46:49 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:01.762 23:46:49 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:02.022 23:46:49 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:02.022 "name": "raid_bdev1", 00:13:02.022 "uuid": "f838604d-143f-41dc-a204-8b95e2c8d556", 00:13:02.022 "strip_size_kb": 64, 00:13:02.022 "state": "online", 00:13:02.022 "raid_level": "raid5f", 00:13:02.022 "superblock": false, 00:13:02.022 "num_base_bdevs": 3, 00:13:02.022 "num_base_bdevs_discovered": 3, 00:13:02.022 "num_base_bdevs_operational": 3, 00:13:02.022 "process": { 00:13:02.022 "type": "rebuild", 00:13:02.022 "target": "spare", 00:13:02.022 "progress": { 00:13:02.022 "blocks": 20480, 00:13:02.022 "percent": 15 00:13:02.022 } 00:13:02.022 }, 00:13:02.022 "base_bdevs_list": [ 00:13:02.022 { 00:13:02.022 "name": "spare", 00:13:02.022 "uuid": "023771d5-72b8-5e19-a7f6-df9a242e54b2", 00:13:02.022 "is_configured": true, 00:13:02.022 "data_offset": 0, 00:13:02.022 "data_size": 65536 00:13:02.022 }, 00:13:02.022 { 00:13:02.022 "name": "BaseBdev2", 00:13:02.022 "uuid": "4c93a001-9bcf-521b-91e3-47ab5cd76e5e", 00:13:02.022 "is_configured": true, 00:13:02.022 "data_offset": 0, 00:13:02.022 "data_size": 65536 00:13:02.022 }, 00:13:02.022 { 00:13:02.022 "name": "BaseBdev3", 00:13:02.022 "uuid": "32f26f6f-bc03-5aad-a3d0-6487083c18fa", 00:13:02.022 "is_configured": true, 00:13:02.022 "data_offset": 0, 00:13:02.022 "data_size": 65536 00:13:02.022 } 00:13:02.022 ] 00:13:02.022 }' 00:13:02.022 23:46:49 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:02.022 23:46:49 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:02.022 23:46:49 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:02.022 23:46:49 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:02.022 23:46:49 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:13:02.022 23:46:49 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:02.022 23:46:49 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:02.022 [2024-11-26 23:46:49.988182] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:02.022 [2024-11-26 23:46:50.047770] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:13:02.022 [2024-11-26 23:46:50.047897] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:02.022 [2024-11-26 23:46:50.047939] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:02.022 [2024-11-26 23:46:50.047965] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:13:02.022 23:46:50 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:02.023 23:46:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:13:02.023 23:46:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:02.023 23:46:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:02.023 23:46:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:02.023 23:46:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:02.023 23:46:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:02.023 23:46:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:02.023 23:46:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:02.023 23:46:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:02.023 23:46:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:02.023 23:46:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:02.023 23:46:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:02.023 23:46:50 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:02.023 23:46:50 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:02.023 23:46:50 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:02.023 23:46:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:02.023 "name": "raid_bdev1", 00:13:02.023 "uuid": "f838604d-143f-41dc-a204-8b95e2c8d556", 00:13:02.023 "strip_size_kb": 64, 00:13:02.023 "state": "online", 00:13:02.023 "raid_level": "raid5f", 00:13:02.023 "superblock": false, 00:13:02.023 "num_base_bdevs": 3, 00:13:02.023 "num_base_bdevs_discovered": 2, 00:13:02.023 "num_base_bdevs_operational": 2, 00:13:02.023 "base_bdevs_list": [ 00:13:02.023 { 00:13:02.023 "name": null, 00:13:02.023 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:02.023 "is_configured": false, 00:13:02.023 "data_offset": 0, 00:13:02.023 "data_size": 65536 00:13:02.023 }, 00:13:02.023 { 00:13:02.023 "name": "BaseBdev2", 00:13:02.023 "uuid": "4c93a001-9bcf-521b-91e3-47ab5cd76e5e", 00:13:02.023 "is_configured": true, 00:13:02.023 "data_offset": 0, 00:13:02.023 "data_size": 65536 00:13:02.023 }, 00:13:02.023 { 00:13:02.023 "name": "BaseBdev3", 00:13:02.023 "uuid": "32f26f6f-bc03-5aad-a3d0-6487083c18fa", 00:13:02.023 "is_configured": true, 00:13:02.023 "data_offset": 0, 00:13:02.023 "data_size": 65536 00:13:02.023 } 00:13:02.023 ] 00:13:02.023 }' 00:13:02.023 23:46:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:02.023 23:46:50 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:02.593 23:46:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:13:02.593 23:46:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:02.593 23:46:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:13:02.593 23:46:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=none 00:13:02.593 23:46:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:02.593 23:46:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:02.593 23:46:50 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:02.593 23:46:50 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:02.593 23:46:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:02.593 23:46:50 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:02.593 23:46:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:02.593 "name": "raid_bdev1", 00:13:02.593 "uuid": "f838604d-143f-41dc-a204-8b95e2c8d556", 00:13:02.593 "strip_size_kb": 64, 00:13:02.593 "state": "online", 00:13:02.593 "raid_level": "raid5f", 00:13:02.593 "superblock": false, 00:13:02.593 "num_base_bdevs": 3, 00:13:02.593 "num_base_bdevs_discovered": 2, 00:13:02.593 "num_base_bdevs_operational": 2, 00:13:02.593 "base_bdevs_list": [ 00:13:02.593 { 00:13:02.593 "name": null, 00:13:02.593 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:02.593 "is_configured": false, 00:13:02.593 "data_offset": 0, 00:13:02.593 "data_size": 65536 00:13:02.593 }, 00:13:02.593 { 00:13:02.593 "name": "BaseBdev2", 00:13:02.593 "uuid": "4c93a001-9bcf-521b-91e3-47ab5cd76e5e", 00:13:02.593 "is_configured": true, 00:13:02.593 "data_offset": 0, 00:13:02.593 "data_size": 65536 00:13:02.593 }, 00:13:02.593 { 00:13:02.593 "name": "BaseBdev3", 00:13:02.593 "uuid": "32f26f6f-bc03-5aad-a3d0-6487083c18fa", 00:13:02.593 "is_configured": true, 00:13:02.593 "data_offset": 0, 00:13:02.593 "data_size": 65536 00:13:02.593 } 00:13:02.593 ] 00:13:02.593 }' 00:13:02.593 23:46:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:02.593 23:46:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:13:02.593 23:46:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:02.593 23:46:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:13:02.593 23:46:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:13:02.594 23:46:50 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:02.594 23:46:50 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:02.594 [2024-11-26 23:46:50.640807] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:13:02.594 [2024-11-26 23:46:50.645271] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000027da0 00:13:02.594 23:46:50 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:02.594 23:46:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@663 -- # sleep 1 00:13:02.594 [2024-11-26 23:46:50.647446] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:13:03.533 23:46:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:03.533 23:46:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:03.533 23:46:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:03.533 23:46:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:03.533 23:46:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:03.533 23:46:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:03.533 23:46:51 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:03.533 23:46:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:03.533 23:46:51 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:03.792 23:46:51 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:03.793 23:46:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:03.793 "name": "raid_bdev1", 00:13:03.793 "uuid": "f838604d-143f-41dc-a204-8b95e2c8d556", 00:13:03.793 "strip_size_kb": 64, 00:13:03.793 "state": "online", 00:13:03.793 "raid_level": "raid5f", 00:13:03.793 "superblock": false, 00:13:03.793 "num_base_bdevs": 3, 00:13:03.793 "num_base_bdevs_discovered": 3, 00:13:03.793 "num_base_bdevs_operational": 3, 00:13:03.793 "process": { 00:13:03.793 "type": "rebuild", 00:13:03.793 "target": "spare", 00:13:03.793 "progress": { 00:13:03.793 "blocks": 20480, 00:13:03.793 "percent": 15 00:13:03.793 } 00:13:03.793 }, 00:13:03.793 "base_bdevs_list": [ 00:13:03.793 { 00:13:03.793 "name": "spare", 00:13:03.793 "uuid": "023771d5-72b8-5e19-a7f6-df9a242e54b2", 00:13:03.793 "is_configured": true, 00:13:03.793 "data_offset": 0, 00:13:03.793 "data_size": 65536 00:13:03.793 }, 00:13:03.793 { 00:13:03.793 "name": "BaseBdev2", 00:13:03.793 "uuid": "4c93a001-9bcf-521b-91e3-47ab5cd76e5e", 00:13:03.793 "is_configured": true, 00:13:03.793 "data_offset": 0, 00:13:03.793 "data_size": 65536 00:13:03.793 }, 00:13:03.793 { 00:13:03.793 "name": "BaseBdev3", 00:13:03.793 "uuid": "32f26f6f-bc03-5aad-a3d0-6487083c18fa", 00:13:03.793 "is_configured": true, 00:13:03.793 "data_offset": 0, 00:13:03.793 "data_size": 65536 00:13:03.793 } 00:13:03.793 ] 00:13:03.793 }' 00:13:03.793 23:46:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:03.793 23:46:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:03.793 23:46:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:03.793 23:46:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:03.793 23:46:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@666 -- # '[' false = true ']' 00:13:03.793 23:46:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=3 00:13:03.793 23:46:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@693 -- # '[' raid5f = raid1 ']' 00:13:03.793 23:46:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@706 -- # local timeout=443 00:13:03.793 23:46:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:03.793 23:46:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:03.793 23:46:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:03.793 23:46:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:03.793 23:46:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:03.793 23:46:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:03.793 23:46:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:03.793 23:46:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:03.793 23:46:51 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:03.793 23:46:51 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:03.793 23:46:51 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:03.793 23:46:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:03.793 "name": "raid_bdev1", 00:13:03.793 "uuid": "f838604d-143f-41dc-a204-8b95e2c8d556", 00:13:03.793 "strip_size_kb": 64, 00:13:03.793 "state": "online", 00:13:03.793 "raid_level": "raid5f", 00:13:03.793 "superblock": false, 00:13:03.793 "num_base_bdevs": 3, 00:13:03.793 "num_base_bdevs_discovered": 3, 00:13:03.793 "num_base_bdevs_operational": 3, 00:13:03.793 "process": { 00:13:03.793 "type": "rebuild", 00:13:03.793 "target": "spare", 00:13:03.793 "progress": { 00:13:03.793 "blocks": 22528, 00:13:03.793 "percent": 17 00:13:03.793 } 00:13:03.793 }, 00:13:03.793 "base_bdevs_list": [ 00:13:03.793 { 00:13:03.793 "name": "spare", 00:13:03.793 "uuid": "023771d5-72b8-5e19-a7f6-df9a242e54b2", 00:13:03.793 "is_configured": true, 00:13:03.793 "data_offset": 0, 00:13:03.793 "data_size": 65536 00:13:03.793 }, 00:13:03.793 { 00:13:03.793 "name": "BaseBdev2", 00:13:03.793 "uuid": "4c93a001-9bcf-521b-91e3-47ab5cd76e5e", 00:13:03.793 "is_configured": true, 00:13:03.793 "data_offset": 0, 00:13:03.793 "data_size": 65536 00:13:03.793 }, 00:13:03.793 { 00:13:03.793 "name": "BaseBdev3", 00:13:03.793 "uuid": "32f26f6f-bc03-5aad-a3d0-6487083c18fa", 00:13:03.793 "is_configured": true, 00:13:03.793 "data_offset": 0, 00:13:03.793 "data_size": 65536 00:13:03.793 } 00:13:03.793 ] 00:13:03.793 }' 00:13:03.793 23:46:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:03.793 23:46:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:03.793 23:46:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:03.793 23:46:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:03.793 23:46:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:13:05.205 23:46:52 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:05.205 23:46:52 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:05.205 23:46:52 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:05.205 23:46:52 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:05.205 23:46:52 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:05.205 23:46:52 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:05.205 23:46:52 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:05.205 23:46:52 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:05.205 23:46:52 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:05.205 23:46:52 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:05.205 23:46:52 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:05.205 23:46:52 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:05.205 "name": "raid_bdev1", 00:13:05.205 "uuid": "f838604d-143f-41dc-a204-8b95e2c8d556", 00:13:05.205 "strip_size_kb": 64, 00:13:05.205 "state": "online", 00:13:05.205 "raid_level": "raid5f", 00:13:05.205 "superblock": false, 00:13:05.205 "num_base_bdevs": 3, 00:13:05.205 "num_base_bdevs_discovered": 3, 00:13:05.205 "num_base_bdevs_operational": 3, 00:13:05.205 "process": { 00:13:05.205 "type": "rebuild", 00:13:05.205 "target": "spare", 00:13:05.205 "progress": { 00:13:05.205 "blocks": 45056, 00:13:05.205 "percent": 34 00:13:05.205 } 00:13:05.205 }, 00:13:05.205 "base_bdevs_list": [ 00:13:05.205 { 00:13:05.205 "name": "spare", 00:13:05.205 "uuid": "023771d5-72b8-5e19-a7f6-df9a242e54b2", 00:13:05.205 "is_configured": true, 00:13:05.205 "data_offset": 0, 00:13:05.205 "data_size": 65536 00:13:05.205 }, 00:13:05.205 { 00:13:05.205 "name": "BaseBdev2", 00:13:05.205 "uuid": "4c93a001-9bcf-521b-91e3-47ab5cd76e5e", 00:13:05.205 "is_configured": true, 00:13:05.205 "data_offset": 0, 00:13:05.205 "data_size": 65536 00:13:05.205 }, 00:13:05.205 { 00:13:05.205 "name": "BaseBdev3", 00:13:05.205 "uuid": "32f26f6f-bc03-5aad-a3d0-6487083c18fa", 00:13:05.205 "is_configured": true, 00:13:05.206 "data_offset": 0, 00:13:05.206 "data_size": 65536 00:13:05.206 } 00:13:05.206 ] 00:13:05.206 }' 00:13:05.206 23:46:52 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:05.206 23:46:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:05.206 23:46:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:05.206 23:46:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:05.206 23:46:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:13:06.147 23:46:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:06.147 23:46:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:06.147 23:46:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:06.147 23:46:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:06.147 23:46:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:06.147 23:46:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:06.147 23:46:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:06.147 23:46:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:06.147 23:46:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:06.147 23:46:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:06.147 23:46:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:06.147 23:46:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:06.147 "name": "raid_bdev1", 00:13:06.147 "uuid": "f838604d-143f-41dc-a204-8b95e2c8d556", 00:13:06.147 "strip_size_kb": 64, 00:13:06.147 "state": "online", 00:13:06.147 "raid_level": "raid5f", 00:13:06.147 "superblock": false, 00:13:06.147 "num_base_bdevs": 3, 00:13:06.147 "num_base_bdevs_discovered": 3, 00:13:06.147 "num_base_bdevs_operational": 3, 00:13:06.147 "process": { 00:13:06.147 "type": "rebuild", 00:13:06.147 "target": "spare", 00:13:06.147 "progress": { 00:13:06.147 "blocks": 69632, 00:13:06.147 "percent": 53 00:13:06.147 } 00:13:06.147 }, 00:13:06.147 "base_bdevs_list": [ 00:13:06.147 { 00:13:06.147 "name": "spare", 00:13:06.147 "uuid": "023771d5-72b8-5e19-a7f6-df9a242e54b2", 00:13:06.147 "is_configured": true, 00:13:06.147 "data_offset": 0, 00:13:06.147 "data_size": 65536 00:13:06.147 }, 00:13:06.147 { 00:13:06.147 "name": "BaseBdev2", 00:13:06.147 "uuid": "4c93a001-9bcf-521b-91e3-47ab5cd76e5e", 00:13:06.147 "is_configured": true, 00:13:06.147 "data_offset": 0, 00:13:06.147 "data_size": 65536 00:13:06.147 }, 00:13:06.147 { 00:13:06.147 "name": "BaseBdev3", 00:13:06.147 "uuid": "32f26f6f-bc03-5aad-a3d0-6487083c18fa", 00:13:06.147 "is_configured": true, 00:13:06.147 "data_offset": 0, 00:13:06.148 "data_size": 65536 00:13:06.148 } 00:13:06.148 ] 00:13:06.148 }' 00:13:06.148 23:46:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:06.148 23:46:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:06.148 23:46:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:06.148 23:46:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:06.148 23:46:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:13:07.088 23:46:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:07.088 23:46:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:07.088 23:46:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:07.088 23:46:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:07.088 23:46:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:07.088 23:46:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:07.088 23:46:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:07.088 23:46:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:07.088 23:46:55 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:07.088 23:46:55 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:07.349 23:46:55 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:07.349 23:46:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:07.349 "name": "raid_bdev1", 00:13:07.349 "uuid": "f838604d-143f-41dc-a204-8b95e2c8d556", 00:13:07.349 "strip_size_kb": 64, 00:13:07.349 "state": "online", 00:13:07.349 "raid_level": "raid5f", 00:13:07.349 "superblock": false, 00:13:07.349 "num_base_bdevs": 3, 00:13:07.349 "num_base_bdevs_discovered": 3, 00:13:07.349 "num_base_bdevs_operational": 3, 00:13:07.349 "process": { 00:13:07.349 "type": "rebuild", 00:13:07.349 "target": "spare", 00:13:07.349 "progress": { 00:13:07.349 "blocks": 92160, 00:13:07.349 "percent": 70 00:13:07.349 } 00:13:07.349 }, 00:13:07.349 "base_bdevs_list": [ 00:13:07.349 { 00:13:07.349 "name": "spare", 00:13:07.349 "uuid": "023771d5-72b8-5e19-a7f6-df9a242e54b2", 00:13:07.349 "is_configured": true, 00:13:07.349 "data_offset": 0, 00:13:07.349 "data_size": 65536 00:13:07.349 }, 00:13:07.349 { 00:13:07.349 "name": "BaseBdev2", 00:13:07.349 "uuid": "4c93a001-9bcf-521b-91e3-47ab5cd76e5e", 00:13:07.349 "is_configured": true, 00:13:07.349 "data_offset": 0, 00:13:07.349 "data_size": 65536 00:13:07.349 }, 00:13:07.349 { 00:13:07.349 "name": "BaseBdev3", 00:13:07.349 "uuid": "32f26f6f-bc03-5aad-a3d0-6487083c18fa", 00:13:07.349 "is_configured": true, 00:13:07.349 "data_offset": 0, 00:13:07.349 "data_size": 65536 00:13:07.349 } 00:13:07.349 ] 00:13:07.349 }' 00:13:07.349 23:46:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:07.349 23:46:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:07.349 23:46:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:07.349 23:46:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:07.349 23:46:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:13:08.288 23:46:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:08.288 23:46:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:08.288 23:46:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:08.288 23:46:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:08.288 23:46:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:08.288 23:46:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:08.288 23:46:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:08.288 23:46:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:08.288 23:46:56 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:08.288 23:46:56 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:08.288 23:46:56 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:08.288 23:46:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:08.288 "name": "raid_bdev1", 00:13:08.288 "uuid": "f838604d-143f-41dc-a204-8b95e2c8d556", 00:13:08.288 "strip_size_kb": 64, 00:13:08.288 "state": "online", 00:13:08.288 "raid_level": "raid5f", 00:13:08.288 "superblock": false, 00:13:08.288 "num_base_bdevs": 3, 00:13:08.288 "num_base_bdevs_discovered": 3, 00:13:08.288 "num_base_bdevs_operational": 3, 00:13:08.288 "process": { 00:13:08.288 "type": "rebuild", 00:13:08.288 "target": "spare", 00:13:08.288 "progress": { 00:13:08.288 "blocks": 114688, 00:13:08.288 "percent": 87 00:13:08.288 } 00:13:08.288 }, 00:13:08.288 "base_bdevs_list": [ 00:13:08.288 { 00:13:08.288 "name": "spare", 00:13:08.288 "uuid": "023771d5-72b8-5e19-a7f6-df9a242e54b2", 00:13:08.288 "is_configured": true, 00:13:08.288 "data_offset": 0, 00:13:08.288 "data_size": 65536 00:13:08.288 }, 00:13:08.288 { 00:13:08.288 "name": "BaseBdev2", 00:13:08.288 "uuid": "4c93a001-9bcf-521b-91e3-47ab5cd76e5e", 00:13:08.288 "is_configured": true, 00:13:08.288 "data_offset": 0, 00:13:08.288 "data_size": 65536 00:13:08.288 }, 00:13:08.288 { 00:13:08.288 "name": "BaseBdev3", 00:13:08.288 "uuid": "32f26f6f-bc03-5aad-a3d0-6487083c18fa", 00:13:08.288 "is_configured": true, 00:13:08.288 "data_offset": 0, 00:13:08.288 "data_size": 65536 00:13:08.288 } 00:13:08.288 ] 00:13:08.288 }' 00:13:08.288 23:46:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:08.548 23:46:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:08.548 23:46:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:08.548 23:46:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:08.548 23:46:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:13:09.118 [2024-11-26 23:46:57.081738] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:13:09.118 [2024-11-26 23:46:57.081846] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:13:09.118 [2024-11-26 23:46:57.081926] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:09.689 23:46:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:09.689 23:46:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:09.689 23:46:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:09.689 23:46:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:09.689 23:46:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:09.689 23:46:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:09.689 23:46:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:09.689 23:46:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:09.689 23:46:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:09.689 23:46:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:09.689 23:46:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:09.689 23:46:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:09.689 "name": "raid_bdev1", 00:13:09.689 "uuid": "f838604d-143f-41dc-a204-8b95e2c8d556", 00:13:09.689 "strip_size_kb": 64, 00:13:09.689 "state": "online", 00:13:09.689 "raid_level": "raid5f", 00:13:09.689 "superblock": false, 00:13:09.689 "num_base_bdevs": 3, 00:13:09.689 "num_base_bdevs_discovered": 3, 00:13:09.689 "num_base_bdevs_operational": 3, 00:13:09.689 "base_bdevs_list": [ 00:13:09.689 { 00:13:09.689 "name": "spare", 00:13:09.689 "uuid": "023771d5-72b8-5e19-a7f6-df9a242e54b2", 00:13:09.689 "is_configured": true, 00:13:09.689 "data_offset": 0, 00:13:09.689 "data_size": 65536 00:13:09.689 }, 00:13:09.689 { 00:13:09.689 "name": "BaseBdev2", 00:13:09.689 "uuid": "4c93a001-9bcf-521b-91e3-47ab5cd76e5e", 00:13:09.689 "is_configured": true, 00:13:09.689 "data_offset": 0, 00:13:09.689 "data_size": 65536 00:13:09.689 }, 00:13:09.689 { 00:13:09.689 "name": "BaseBdev3", 00:13:09.689 "uuid": "32f26f6f-bc03-5aad-a3d0-6487083c18fa", 00:13:09.689 "is_configured": true, 00:13:09.689 "data_offset": 0, 00:13:09.689 "data_size": 65536 00:13:09.689 } 00:13:09.689 ] 00:13:09.689 }' 00:13:09.689 23:46:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:09.689 23:46:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:13:09.689 23:46:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:09.689 23:46:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:13:09.689 23:46:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@709 -- # break 00:13:09.689 23:46:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:13:09.689 23:46:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:09.689 23:46:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:13:09.689 23:46:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=none 00:13:09.689 23:46:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:09.689 23:46:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:09.689 23:46:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:09.689 23:46:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:09.689 23:46:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:09.689 23:46:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:09.689 23:46:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:09.689 "name": "raid_bdev1", 00:13:09.689 "uuid": "f838604d-143f-41dc-a204-8b95e2c8d556", 00:13:09.689 "strip_size_kb": 64, 00:13:09.689 "state": "online", 00:13:09.689 "raid_level": "raid5f", 00:13:09.689 "superblock": false, 00:13:09.689 "num_base_bdevs": 3, 00:13:09.689 "num_base_bdevs_discovered": 3, 00:13:09.689 "num_base_bdevs_operational": 3, 00:13:09.689 "base_bdevs_list": [ 00:13:09.689 { 00:13:09.689 "name": "spare", 00:13:09.689 "uuid": "023771d5-72b8-5e19-a7f6-df9a242e54b2", 00:13:09.689 "is_configured": true, 00:13:09.689 "data_offset": 0, 00:13:09.689 "data_size": 65536 00:13:09.689 }, 00:13:09.689 { 00:13:09.689 "name": "BaseBdev2", 00:13:09.689 "uuid": "4c93a001-9bcf-521b-91e3-47ab5cd76e5e", 00:13:09.689 "is_configured": true, 00:13:09.689 "data_offset": 0, 00:13:09.689 "data_size": 65536 00:13:09.689 }, 00:13:09.689 { 00:13:09.689 "name": "BaseBdev3", 00:13:09.689 "uuid": "32f26f6f-bc03-5aad-a3d0-6487083c18fa", 00:13:09.689 "is_configured": true, 00:13:09.689 "data_offset": 0, 00:13:09.689 "data_size": 65536 00:13:09.689 } 00:13:09.689 ] 00:13:09.689 }' 00:13:09.689 23:46:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:09.689 23:46:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:13:09.689 23:46:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:09.689 23:46:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:13:09.689 23:46:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:13:09.689 23:46:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:09.689 23:46:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:09.689 23:46:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:09.689 23:46:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:09.689 23:46:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:09.689 23:46:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:09.689 23:46:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:09.689 23:46:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:09.689 23:46:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:09.689 23:46:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:09.689 23:46:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:09.689 23:46:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:09.689 23:46:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:09.689 23:46:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:09.950 23:46:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:09.950 "name": "raid_bdev1", 00:13:09.950 "uuid": "f838604d-143f-41dc-a204-8b95e2c8d556", 00:13:09.950 "strip_size_kb": 64, 00:13:09.950 "state": "online", 00:13:09.950 "raid_level": "raid5f", 00:13:09.950 "superblock": false, 00:13:09.950 "num_base_bdevs": 3, 00:13:09.950 "num_base_bdevs_discovered": 3, 00:13:09.950 "num_base_bdevs_operational": 3, 00:13:09.950 "base_bdevs_list": [ 00:13:09.950 { 00:13:09.950 "name": "spare", 00:13:09.950 "uuid": "023771d5-72b8-5e19-a7f6-df9a242e54b2", 00:13:09.950 "is_configured": true, 00:13:09.950 "data_offset": 0, 00:13:09.950 "data_size": 65536 00:13:09.950 }, 00:13:09.950 { 00:13:09.950 "name": "BaseBdev2", 00:13:09.950 "uuid": "4c93a001-9bcf-521b-91e3-47ab5cd76e5e", 00:13:09.950 "is_configured": true, 00:13:09.950 "data_offset": 0, 00:13:09.950 "data_size": 65536 00:13:09.950 }, 00:13:09.950 { 00:13:09.950 "name": "BaseBdev3", 00:13:09.950 "uuid": "32f26f6f-bc03-5aad-a3d0-6487083c18fa", 00:13:09.950 "is_configured": true, 00:13:09.950 "data_offset": 0, 00:13:09.950 "data_size": 65536 00:13:09.950 } 00:13:09.950 ] 00:13:09.950 }' 00:13:09.950 23:46:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:09.950 23:46:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:10.210 23:46:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:13:10.210 23:46:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:10.210 23:46:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:10.210 [2024-11-26 23:46:58.189875] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:13:10.210 [2024-11-26 23:46:58.189906] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:13:10.210 [2024-11-26 23:46:58.190014] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:13:10.210 [2024-11-26 23:46:58.190091] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:13:10.210 [2024-11-26 23:46:58.190101] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:13:10.210 23:46:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:10.210 23:46:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@720 -- # jq length 00:13:10.210 23:46:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:10.210 23:46:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:10.210 23:46:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:10.210 23:46:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:10.210 23:46:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:13:10.210 23:46:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:13:10.210 23:46:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:13:10.210 23:46:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:13:10.210 23:46:58 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:13:10.210 23:46:58 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:13:10.210 23:46:58 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@10 -- # local bdev_list 00:13:10.210 23:46:58 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:13:10.210 23:46:58 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@11 -- # local nbd_list 00:13:10.210 23:46:58 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@12 -- # local i 00:13:10.210 23:46:58 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:13:10.210 23:46:58 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:13:10.210 23:46:58 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:13:10.470 /dev/nbd0 00:13:10.470 23:46:58 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:13:10.470 23:46:58 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:13:10.470 23:46:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:13:10.470 23:46:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@873 -- # local i 00:13:10.470 23:46:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:13:10.470 23:46:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:13:10.470 23:46:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:13:10.470 23:46:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@877 -- # break 00:13:10.470 23:46:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:13:10.470 23:46:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:13:10.470 23:46:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:13:10.470 1+0 records in 00:13:10.470 1+0 records out 00:13:10.470 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000322183 s, 12.7 MB/s 00:13:10.470 23:46:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:10.470 23:46:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@890 -- # size=4096 00:13:10.470 23:46:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:10.470 23:46:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:13:10.470 23:46:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@893 -- # return 0 00:13:10.470 23:46:58 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:13:10.470 23:46:58 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:13:10.470 23:46:58 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:13:10.731 /dev/nbd1 00:13:10.731 23:46:58 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:13:10.731 23:46:58 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:13:10.731 23:46:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:13:10.731 23:46:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@873 -- # local i 00:13:10.731 23:46:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:13:10.731 23:46:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:13:10.731 23:46:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:13:10.731 23:46:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@877 -- # break 00:13:10.731 23:46:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:13:10.731 23:46:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:13:10.731 23:46:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:13:10.731 1+0 records in 00:13:10.731 1+0 records out 00:13:10.731 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000403389 s, 10.2 MB/s 00:13:10.731 23:46:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:10.731 23:46:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@890 -- # size=4096 00:13:10.731 23:46:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:10.731 23:46:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:13:10.731 23:46:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@893 -- # return 0 00:13:10.731 23:46:58 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:13:10.731 23:46:58 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:13:10.731 23:46:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@738 -- # cmp -i 0 /dev/nbd0 /dev/nbd1 00:13:10.731 23:46:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:13:10.731 23:46:58 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:13:10.731 23:46:58 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:13:10.731 23:46:58 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@50 -- # local nbd_list 00:13:10.731 23:46:58 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@51 -- # local i 00:13:10.731 23:46:58 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:13:10.731 23:46:58 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:13:10.992 23:46:58 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:13:10.992 23:46:58 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:13:10.992 23:46:58 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:13:10.992 23:46:58 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:13:10.992 23:46:58 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:13:10.992 23:46:58 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:13:10.992 23:46:58 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:13:10.992 23:46:58 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:13:10.992 23:46:58 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:13:10.992 23:46:58 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:13:10.992 23:46:59 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:13:10.992 23:46:59 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:13:10.992 23:46:59 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:13:10.992 23:46:59 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:13:10.992 23:46:59 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:13:10.992 23:46:59 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:13:10.992 23:46:59 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:13:10.992 23:46:59 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:13:11.252 23:46:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@743 -- # '[' false = true ']' 00:13:11.252 23:46:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@784 -- # killprocess 91772 00:13:11.252 23:46:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@954 -- # '[' -z 91772 ']' 00:13:11.252 23:46:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@958 -- # kill -0 91772 00:13:11.252 23:46:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@959 -- # uname 00:13:11.252 23:46:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:13:11.252 23:46:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 91772 00:13:11.252 killing process with pid 91772 00:13:11.252 Received shutdown signal, test time was about 60.000000 seconds 00:13:11.252 00:13:11.252 Latency(us) 00:13:11.252 [2024-11-26T23:46:59.384Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:13:11.252 [2024-11-26T23:46:59.384Z] =================================================================================================================== 00:13:11.252 [2024-11-26T23:46:59.384Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:13:11.252 23:46:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:13:11.252 23:46:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:13:11.252 23:46:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 91772' 00:13:11.252 23:46:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@973 -- # kill 91772 00:13:11.252 [2024-11-26 23:46:59.150770] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:13:11.252 23:46:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@978 -- # wait 91772 00:13:11.252 [2024-11-26 23:46:59.190708] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:13:11.511 23:46:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@786 -- # return 0 00:13:11.511 00:13:11.511 real 0m13.346s 00:13:11.511 user 0m16.732s 00:13:11.511 sys 0m1.767s 00:13:11.511 ************************************ 00:13:11.511 END TEST raid5f_rebuild_test 00:13:11.511 ************************************ 00:13:11.511 23:46:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:13:11.511 23:46:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:11.511 23:46:59 bdev_raid -- bdev/bdev_raid.sh@991 -- # run_test raid5f_rebuild_test_sb raid_rebuild_test raid5f 3 true false true 00:13:11.511 23:46:59 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:13:11.511 23:46:59 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:13:11.511 23:46:59 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:13:11.512 ************************************ 00:13:11.512 START TEST raid5f_rebuild_test_sb 00:13:11.512 ************************************ 00:13:11.512 23:46:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@1129 -- # raid_rebuild_test raid5f 3 true false true 00:13:11.512 23:46:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@569 -- # local raid_level=raid5f 00:13:11.512 23:46:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=3 00:13:11.512 23:46:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@571 -- # local superblock=true 00:13:11.512 23:46:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:13:11.512 23:46:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@573 -- # local verify=true 00:13:11.512 23:46:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:13:11.512 23:46:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:13:11.512 23:46:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:13:11.512 23:46:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:13:11.512 23:46:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:13:11.512 23:46:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:13:11.512 23:46:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:13:11.512 23:46:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:13:11.512 23:46:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev3 00:13:11.512 23:46:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:13:11.512 23:46:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:13:11.512 23:46:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:13:11.512 23:46:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:13:11.512 23:46:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:13:11.512 23:46:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # local strip_size 00:13:11.512 23:46:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@577 -- # local create_arg 00:13:11.512 23:46:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:13:11.512 23:46:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@579 -- # local data_offset 00:13:11.512 23:46:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@581 -- # '[' raid5f '!=' raid1 ']' 00:13:11.512 23:46:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@582 -- # '[' false = true ']' 00:13:11.512 23:46:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@586 -- # strip_size=64 00:13:11.512 23:46:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@587 -- # create_arg+=' -z 64' 00:13:11.512 23:46:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@592 -- # '[' true = true ']' 00:13:11.512 23:46:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@593 -- # create_arg+=' -s' 00:13:11.512 23:46:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:13:11.512 23:46:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@597 -- # raid_pid=92190 00:13:11.512 23:46:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@598 -- # waitforlisten 92190 00:13:11.512 23:46:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@835 -- # '[' -z 92190 ']' 00:13:11.512 23:46:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:13:11.512 23:46:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@840 -- # local max_retries=100 00:13:11.512 23:46:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:13:11.512 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:13:11.512 23:46:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@844 -- # xtrace_disable 00:13:11.512 23:46:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:11.512 [2024-11-26 23:46:59.519203] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:13:11.512 [2024-11-26 23:46:59.519413] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid92190 ] 00:13:11.512 I/O size of 3145728 is greater than zero copy threshold (65536). 00:13:11.512 Zero copy mechanism will not be used. 00:13:11.772 [2024-11-26 23:46:59.674652] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:13:11.772 [2024-11-26 23:46:59.698772] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:13:11.772 [2024-11-26 23:46:59.740063] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:13:11.772 [2024-11-26 23:46:59.740170] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:13:12.343 23:47:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:13:12.343 23:47:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@868 -- # return 0 00:13:12.343 23:47:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:13:12.343 23:47:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:13:12.343 23:47:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:12.343 23:47:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:12.343 BaseBdev1_malloc 00:13:12.343 23:47:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:12.343 23:47:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:13:12.343 23:47:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:12.343 23:47:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:12.343 [2024-11-26 23:47:00.382890] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:13:12.343 [2024-11-26 23:47:00.383012] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:12.343 [2024-11-26 23:47:00.383058] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:13:12.343 [2024-11-26 23:47:00.383106] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:12.343 [2024-11-26 23:47:00.385119] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:12.343 [2024-11-26 23:47:00.385186] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:13:12.343 BaseBdev1 00:13:12.343 23:47:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:12.343 23:47:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:13:12.343 23:47:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:13:12.343 23:47:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:12.343 23:47:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:12.343 BaseBdev2_malloc 00:13:12.343 23:47:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:12.343 23:47:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:13:12.343 23:47:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:12.343 23:47:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:12.343 [2024-11-26 23:47:00.411002] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:13:12.343 [2024-11-26 23:47:00.411055] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:12.343 [2024-11-26 23:47:00.411093] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:13:12.343 [2024-11-26 23:47:00.411102] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:12.343 [2024-11-26 23:47:00.413078] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:12.343 [2024-11-26 23:47:00.413129] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:13:12.343 BaseBdev2 00:13:12.343 23:47:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:12.343 23:47:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:13:12.343 23:47:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:13:12.343 23:47:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:12.343 23:47:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:12.343 BaseBdev3_malloc 00:13:12.343 23:47:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:12.343 23:47:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev3_malloc -p BaseBdev3 00:13:12.343 23:47:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:12.343 23:47:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:12.343 [2024-11-26 23:47:00.439077] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev3_malloc 00:13:12.343 [2024-11-26 23:47:00.439131] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:12.343 [2024-11-26 23:47:00.439171] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:13:12.343 [2024-11-26 23:47:00.439179] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:12.343 [2024-11-26 23:47:00.441189] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:12.343 [2024-11-26 23:47:00.441270] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:13:12.343 BaseBdev3 00:13:12.343 23:47:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:12.343 23:47:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:13:12.343 23:47:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:12.343 23:47:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:12.604 spare_malloc 00:13:12.604 23:47:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:12.604 23:47:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:13:12.604 23:47:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:12.604 23:47:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:12.604 spare_delay 00:13:12.604 23:47:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:12.604 23:47:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:13:12.604 23:47:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:12.604 23:47:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:12.604 [2024-11-26 23:47:00.495822] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:13:12.604 [2024-11-26 23:47:00.495878] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:12.604 [2024-11-26 23:47:00.495902] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:13:12.604 [2024-11-26 23:47:00.495911] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:12.604 [2024-11-26 23:47:00.497936] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:12.604 [2024-11-26 23:47:00.498009] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:13:12.604 spare 00:13:12.604 23:47:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:12.604 23:47:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n raid_bdev1 00:13:12.604 23:47:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:12.604 23:47:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:12.604 [2024-11-26 23:47:00.507857] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:13:12.604 [2024-11-26 23:47:00.509666] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:13:12.604 [2024-11-26 23:47:00.509723] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:13:12.604 [2024-11-26 23:47:00.509872] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:13:12.604 [2024-11-26 23:47:00.509889] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:13:12.604 [2024-11-26 23:47:00.510113] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:13:12.604 [2024-11-26 23:47:00.510532] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:13:12.604 [2024-11-26 23:47:00.510549] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:13:12.604 [2024-11-26 23:47:00.510677] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:12.604 23:47:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:12.604 23:47:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:13:12.604 23:47:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:12.604 23:47:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:12.604 23:47:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:12.604 23:47:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:12.604 23:47:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:12.604 23:47:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:12.604 23:47:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:12.604 23:47:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:12.604 23:47:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:12.604 23:47:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:12.604 23:47:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:12.604 23:47:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:12.604 23:47:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:12.604 23:47:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:12.604 23:47:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:12.604 "name": "raid_bdev1", 00:13:12.604 "uuid": "aeb0da41-a687-43e0-974e-937ec7519a0e", 00:13:12.604 "strip_size_kb": 64, 00:13:12.604 "state": "online", 00:13:12.604 "raid_level": "raid5f", 00:13:12.604 "superblock": true, 00:13:12.605 "num_base_bdevs": 3, 00:13:12.605 "num_base_bdevs_discovered": 3, 00:13:12.605 "num_base_bdevs_operational": 3, 00:13:12.605 "base_bdevs_list": [ 00:13:12.605 { 00:13:12.605 "name": "BaseBdev1", 00:13:12.605 "uuid": "750b6f65-5218-5851-bddf-b9ac0188a1d8", 00:13:12.605 "is_configured": true, 00:13:12.605 "data_offset": 2048, 00:13:12.605 "data_size": 63488 00:13:12.605 }, 00:13:12.605 { 00:13:12.605 "name": "BaseBdev2", 00:13:12.605 "uuid": "6f25ce10-ae25-567e-bda6-5434d31c5259", 00:13:12.605 "is_configured": true, 00:13:12.605 "data_offset": 2048, 00:13:12.605 "data_size": 63488 00:13:12.605 }, 00:13:12.605 { 00:13:12.605 "name": "BaseBdev3", 00:13:12.605 "uuid": "5538ab62-c9e7-5d6c-86e3-cd8f39be431c", 00:13:12.605 "is_configured": true, 00:13:12.605 "data_offset": 2048, 00:13:12.605 "data_size": 63488 00:13:12.605 } 00:13:12.605 ] 00:13:12.605 }' 00:13:12.605 23:47:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:12.605 23:47:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:12.868 23:47:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:13:12.868 23:47:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:12.868 23:47:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:13:12.868 23:47:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:12.868 [2024-11-26 23:47:00.931469] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:13:12.868 23:47:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:12.868 23:47:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=126976 00:13:12.868 23:47:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:12.868 23:47:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:12.868 23:47:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:12.868 23:47:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:13:12.868 23:47:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:13.132 23:47:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # data_offset=2048 00:13:13.132 23:47:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:13:13.132 23:47:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:13:13.132 23:47:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:13:13.132 23:47:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:13:13.132 23:47:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:13:13.132 23:47:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:13:13.132 23:47:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # local bdev_list 00:13:13.132 23:47:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:13:13.132 23:47:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # local nbd_list 00:13:13.132 23:47:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@12 -- # local i 00:13:13.132 23:47:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:13:13.132 23:47:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:13:13.132 23:47:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:13:13.132 [2024-11-26 23:47:01.178918] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000026d0 00:13:13.132 /dev/nbd0 00:13:13.132 23:47:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:13:13.132 23:47:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:13:13.132 23:47:01 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:13:13.132 23:47:01 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@873 -- # local i 00:13:13.132 23:47:01 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:13:13.132 23:47:01 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:13:13.132 23:47:01 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:13:13.132 23:47:01 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@877 -- # break 00:13:13.132 23:47:01 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:13:13.132 23:47:01 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:13:13.132 23:47:01 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:13:13.132 1+0 records in 00:13:13.132 1+0 records out 00:13:13.132 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000359665 s, 11.4 MB/s 00:13:13.132 23:47:01 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:13.132 23:47:01 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@890 -- # size=4096 00:13:13.132 23:47:01 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:13.132 23:47:01 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:13:13.132 23:47:01 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@893 -- # return 0 00:13:13.132 23:47:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:13:13.132 23:47:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:13:13.132 23:47:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@629 -- # '[' raid5f = raid5f ']' 00:13:13.132 23:47:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@630 -- # write_unit_size=256 00:13:13.132 23:47:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@631 -- # echo 128 00:13:13.132 23:47:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=131072 count=496 oflag=direct 00:13:13.701 496+0 records in 00:13:13.701 496+0 records out 00:13:13.701 65011712 bytes (65 MB, 62 MiB) copied, 0.277579 s, 234 MB/s 00:13:13.701 23:47:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:13:13.701 23:47:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:13:13.701 23:47:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:13:13.701 23:47:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # local nbd_list 00:13:13.701 23:47:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@51 -- # local i 00:13:13.701 23:47:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:13:13.701 23:47:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:13:13.701 23:47:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:13:13.701 23:47:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:13:13.701 23:47:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:13:13.701 [2024-11-26 23:47:01.745313] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:13.701 23:47:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:13:13.701 23:47:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:13:13.701 23:47:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:13:13.701 23:47:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:13:13.701 23:47:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:13:13.701 23:47:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:13:13.701 23:47:01 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:13.701 23:47:01 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:13.702 [2024-11-26 23:47:01.761425] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:13:13.702 23:47:01 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:13.702 23:47:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:13:13.702 23:47:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:13.702 23:47:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:13.702 23:47:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:13.702 23:47:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:13.702 23:47:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:13.702 23:47:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:13.702 23:47:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:13.702 23:47:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:13.702 23:47:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:13.702 23:47:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:13.702 23:47:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:13.702 23:47:01 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:13.702 23:47:01 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:13.702 23:47:01 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:13.702 23:47:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:13.702 "name": "raid_bdev1", 00:13:13.702 "uuid": "aeb0da41-a687-43e0-974e-937ec7519a0e", 00:13:13.702 "strip_size_kb": 64, 00:13:13.702 "state": "online", 00:13:13.702 "raid_level": "raid5f", 00:13:13.702 "superblock": true, 00:13:13.702 "num_base_bdevs": 3, 00:13:13.702 "num_base_bdevs_discovered": 2, 00:13:13.702 "num_base_bdevs_operational": 2, 00:13:13.702 "base_bdevs_list": [ 00:13:13.702 { 00:13:13.702 "name": null, 00:13:13.702 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:13.702 "is_configured": false, 00:13:13.702 "data_offset": 0, 00:13:13.702 "data_size": 63488 00:13:13.702 }, 00:13:13.702 { 00:13:13.702 "name": "BaseBdev2", 00:13:13.702 "uuid": "6f25ce10-ae25-567e-bda6-5434d31c5259", 00:13:13.702 "is_configured": true, 00:13:13.702 "data_offset": 2048, 00:13:13.702 "data_size": 63488 00:13:13.702 }, 00:13:13.702 { 00:13:13.702 "name": "BaseBdev3", 00:13:13.702 "uuid": "5538ab62-c9e7-5d6c-86e3-cd8f39be431c", 00:13:13.702 "is_configured": true, 00:13:13.702 "data_offset": 2048, 00:13:13.702 "data_size": 63488 00:13:13.702 } 00:13:13.702 ] 00:13:13.702 }' 00:13:13.702 23:47:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:13.702 23:47:01 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:14.271 23:47:02 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:13:14.271 23:47:02 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:14.271 23:47:02 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:14.271 [2024-11-26 23:47:02.228595] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:13:14.271 [2024-11-26 23:47:02.233213] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000255d0 00:13:14.271 23:47:02 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:14.271 23:47:02 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@647 -- # sleep 1 00:13:14.271 [2024-11-26 23:47:02.235329] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:13:15.216 23:47:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:15.216 23:47:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:15.216 23:47:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:15.216 23:47:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:15.216 23:47:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:15.216 23:47:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:15.216 23:47:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:15.216 23:47:03 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:15.216 23:47:03 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:15.216 23:47:03 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:15.216 23:47:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:15.216 "name": "raid_bdev1", 00:13:15.216 "uuid": "aeb0da41-a687-43e0-974e-937ec7519a0e", 00:13:15.216 "strip_size_kb": 64, 00:13:15.216 "state": "online", 00:13:15.216 "raid_level": "raid5f", 00:13:15.216 "superblock": true, 00:13:15.216 "num_base_bdevs": 3, 00:13:15.216 "num_base_bdevs_discovered": 3, 00:13:15.216 "num_base_bdevs_operational": 3, 00:13:15.216 "process": { 00:13:15.216 "type": "rebuild", 00:13:15.216 "target": "spare", 00:13:15.216 "progress": { 00:13:15.216 "blocks": 20480, 00:13:15.216 "percent": 16 00:13:15.216 } 00:13:15.216 }, 00:13:15.216 "base_bdevs_list": [ 00:13:15.216 { 00:13:15.216 "name": "spare", 00:13:15.216 "uuid": "a57a253c-5b63-5c17-a10d-19cb39fa9847", 00:13:15.216 "is_configured": true, 00:13:15.216 "data_offset": 2048, 00:13:15.216 "data_size": 63488 00:13:15.216 }, 00:13:15.216 { 00:13:15.216 "name": "BaseBdev2", 00:13:15.216 "uuid": "6f25ce10-ae25-567e-bda6-5434d31c5259", 00:13:15.216 "is_configured": true, 00:13:15.216 "data_offset": 2048, 00:13:15.216 "data_size": 63488 00:13:15.216 }, 00:13:15.216 { 00:13:15.216 "name": "BaseBdev3", 00:13:15.216 "uuid": "5538ab62-c9e7-5d6c-86e3-cd8f39be431c", 00:13:15.216 "is_configured": true, 00:13:15.216 "data_offset": 2048, 00:13:15.216 "data_size": 63488 00:13:15.216 } 00:13:15.216 ] 00:13:15.216 }' 00:13:15.216 23:47:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:15.216 23:47:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:15.216 23:47:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:15.476 23:47:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:15.476 23:47:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:13:15.476 23:47:03 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:15.476 23:47:03 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:15.476 [2024-11-26 23:47:03.391405] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:15.476 [2024-11-26 23:47:03.442176] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:13:15.476 [2024-11-26 23:47:03.442306] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:15.476 [2024-11-26 23:47:03.442325] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:15.476 [2024-11-26 23:47:03.442342] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:13:15.476 23:47:03 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:15.476 23:47:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:13:15.476 23:47:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:15.476 23:47:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:15.476 23:47:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:15.476 23:47:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:15.476 23:47:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:15.476 23:47:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:15.476 23:47:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:15.476 23:47:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:15.476 23:47:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:15.476 23:47:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:15.476 23:47:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:15.476 23:47:03 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:15.476 23:47:03 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:15.476 23:47:03 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:15.476 23:47:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:15.476 "name": "raid_bdev1", 00:13:15.476 "uuid": "aeb0da41-a687-43e0-974e-937ec7519a0e", 00:13:15.476 "strip_size_kb": 64, 00:13:15.476 "state": "online", 00:13:15.476 "raid_level": "raid5f", 00:13:15.476 "superblock": true, 00:13:15.476 "num_base_bdevs": 3, 00:13:15.476 "num_base_bdevs_discovered": 2, 00:13:15.476 "num_base_bdevs_operational": 2, 00:13:15.476 "base_bdevs_list": [ 00:13:15.476 { 00:13:15.476 "name": null, 00:13:15.476 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:15.476 "is_configured": false, 00:13:15.476 "data_offset": 0, 00:13:15.476 "data_size": 63488 00:13:15.476 }, 00:13:15.476 { 00:13:15.476 "name": "BaseBdev2", 00:13:15.476 "uuid": "6f25ce10-ae25-567e-bda6-5434d31c5259", 00:13:15.476 "is_configured": true, 00:13:15.476 "data_offset": 2048, 00:13:15.476 "data_size": 63488 00:13:15.476 }, 00:13:15.476 { 00:13:15.476 "name": "BaseBdev3", 00:13:15.476 "uuid": "5538ab62-c9e7-5d6c-86e3-cd8f39be431c", 00:13:15.476 "is_configured": true, 00:13:15.476 "data_offset": 2048, 00:13:15.476 "data_size": 63488 00:13:15.476 } 00:13:15.476 ] 00:13:15.476 }' 00:13:15.476 23:47:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:15.476 23:47:03 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:15.737 23:47:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:13:15.737 23:47:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:15.737 23:47:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:13:15.737 23:47:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:13:15.737 23:47:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:15.737 23:47:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:15.737 23:47:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:15.737 23:47:03 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:15.737 23:47:03 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:15.998 23:47:03 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:15.998 23:47:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:15.998 "name": "raid_bdev1", 00:13:15.998 "uuid": "aeb0da41-a687-43e0-974e-937ec7519a0e", 00:13:15.998 "strip_size_kb": 64, 00:13:15.998 "state": "online", 00:13:15.998 "raid_level": "raid5f", 00:13:15.998 "superblock": true, 00:13:15.998 "num_base_bdevs": 3, 00:13:15.998 "num_base_bdevs_discovered": 2, 00:13:15.998 "num_base_bdevs_operational": 2, 00:13:15.998 "base_bdevs_list": [ 00:13:15.998 { 00:13:15.998 "name": null, 00:13:15.998 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:15.998 "is_configured": false, 00:13:15.998 "data_offset": 0, 00:13:15.998 "data_size": 63488 00:13:15.998 }, 00:13:15.998 { 00:13:15.998 "name": "BaseBdev2", 00:13:15.998 "uuid": "6f25ce10-ae25-567e-bda6-5434d31c5259", 00:13:15.998 "is_configured": true, 00:13:15.998 "data_offset": 2048, 00:13:15.998 "data_size": 63488 00:13:15.998 }, 00:13:15.998 { 00:13:15.998 "name": "BaseBdev3", 00:13:15.998 "uuid": "5538ab62-c9e7-5d6c-86e3-cd8f39be431c", 00:13:15.998 "is_configured": true, 00:13:15.998 "data_offset": 2048, 00:13:15.998 "data_size": 63488 00:13:15.998 } 00:13:15.998 ] 00:13:15.998 }' 00:13:15.998 23:47:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:15.998 23:47:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:13:15.998 23:47:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:15.998 23:47:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:13:15.998 23:47:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:13:15.998 23:47:04 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:15.998 23:47:04 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:15.998 [2024-11-26 23:47:04.011273] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:13:15.998 [2024-11-26 23:47:04.015799] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000256a0 00:13:15.998 23:47:04 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:15.998 23:47:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@663 -- # sleep 1 00:13:15.998 [2024-11-26 23:47:04.017875] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:13:16.942 23:47:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:16.942 23:47:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:16.942 23:47:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:16.942 23:47:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:16.942 23:47:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:16.942 23:47:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:16.942 23:47:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:16.942 23:47:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:16.942 23:47:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:16.942 23:47:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:17.262 23:47:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:17.262 "name": "raid_bdev1", 00:13:17.262 "uuid": "aeb0da41-a687-43e0-974e-937ec7519a0e", 00:13:17.262 "strip_size_kb": 64, 00:13:17.262 "state": "online", 00:13:17.262 "raid_level": "raid5f", 00:13:17.262 "superblock": true, 00:13:17.262 "num_base_bdevs": 3, 00:13:17.262 "num_base_bdevs_discovered": 3, 00:13:17.262 "num_base_bdevs_operational": 3, 00:13:17.262 "process": { 00:13:17.262 "type": "rebuild", 00:13:17.262 "target": "spare", 00:13:17.262 "progress": { 00:13:17.262 "blocks": 20480, 00:13:17.262 "percent": 16 00:13:17.262 } 00:13:17.262 }, 00:13:17.262 "base_bdevs_list": [ 00:13:17.262 { 00:13:17.262 "name": "spare", 00:13:17.262 "uuid": "a57a253c-5b63-5c17-a10d-19cb39fa9847", 00:13:17.262 "is_configured": true, 00:13:17.262 "data_offset": 2048, 00:13:17.262 "data_size": 63488 00:13:17.262 }, 00:13:17.262 { 00:13:17.262 "name": "BaseBdev2", 00:13:17.262 "uuid": "6f25ce10-ae25-567e-bda6-5434d31c5259", 00:13:17.262 "is_configured": true, 00:13:17.262 "data_offset": 2048, 00:13:17.262 "data_size": 63488 00:13:17.262 }, 00:13:17.262 { 00:13:17.262 "name": "BaseBdev3", 00:13:17.262 "uuid": "5538ab62-c9e7-5d6c-86e3-cd8f39be431c", 00:13:17.262 "is_configured": true, 00:13:17.262 "data_offset": 2048, 00:13:17.262 "data_size": 63488 00:13:17.262 } 00:13:17.262 ] 00:13:17.262 }' 00:13:17.262 23:47:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:17.262 23:47:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:17.262 23:47:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:17.262 23:47:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:17.262 23:47:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@666 -- # '[' true = true ']' 00:13:17.262 23:47:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@666 -- # '[' = false ']' 00:13:17.262 /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh: line 666: [: =: unary operator expected 00:13:17.262 23:47:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=3 00:13:17.262 23:47:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@693 -- # '[' raid5f = raid1 ']' 00:13:17.262 23:47:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@706 -- # local timeout=457 00:13:17.262 23:47:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:17.262 23:47:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:17.262 23:47:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:17.262 23:47:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:17.262 23:47:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:17.262 23:47:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:17.262 23:47:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:17.262 23:47:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:17.262 23:47:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:17.262 23:47:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:17.262 23:47:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:17.262 23:47:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:17.262 "name": "raid_bdev1", 00:13:17.262 "uuid": "aeb0da41-a687-43e0-974e-937ec7519a0e", 00:13:17.262 "strip_size_kb": 64, 00:13:17.262 "state": "online", 00:13:17.262 "raid_level": "raid5f", 00:13:17.262 "superblock": true, 00:13:17.262 "num_base_bdevs": 3, 00:13:17.262 "num_base_bdevs_discovered": 3, 00:13:17.262 "num_base_bdevs_operational": 3, 00:13:17.262 "process": { 00:13:17.262 "type": "rebuild", 00:13:17.262 "target": "spare", 00:13:17.262 "progress": { 00:13:17.262 "blocks": 22528, 00:13:17.262 "percent": 17 00:13:17.262 } 00:13:17.262 }, 00:13:17.262 "base_bdevs_list": [ 00:13:17.262 { 00:13:17.262 "name": "spare", 00:13:17.262 "uuid": "a57a253c-5b63-5c17-a10d-19cb39fa9847", 00:13:17.262 "is_configured": true, 00:13:17.262 "data_offset": 2048, 00:13:17.262 "data_size": 63488 00:13:17.262 }, 00:13:17.262 { 00:13:17.262 "name": "BaseBdev2", 00:13:17.262 "uuid": "6f25ce10-ae25-567e-bda6-5434d31c5259", 00:13:17.262 "is_configured": true, 00:13:17.262 "data_offset": 2048, 00:13:17.262 "data_size": 63488 00:13:17.262 }, 00:13:17.262 { 00:13:17.262 "name": "BaseBdev3", 00:13:17.262 "uuid": "5538ab62-c9e7-5d6c-86e3-cd8f39be431c", 00:13:17.262 "is_configured": true, 00:13:17.262 "data_offset": 2048, 00:13:17.262 "data_size": 63488 00:13:17.262 } 00:13:17.262 ] 00:13:17.262 }' 00:13:17.262 23:47:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:17.262 23:47:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:17.262 23:47:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:17.262 23:47:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:17.262 23:47:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:13:18.201 23:47:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:18.201 23:47:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:18.201 23:47:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:18.201 23:47:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:18.201 23:47:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:18.201 23:47:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:18.201 23:47:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:18.201 23:47:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:18.201 23:47:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:18.201 23:47:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:18.201 23:47:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:18.460 23:47:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:18.460 "name": "raid_bdev1", 00:13:18.460 "uuid": "aeb0da41-a687-43e0-974e-937ec7519a0e", 00:13:18.460 "strip_size_kb": 64, 00:13:18.460 "state": "online", 00:13:18.460 "raid_level": "raid5f", 00:13:18.460 "superblock": true, 00:13:18.460 "num_base_bdevs": 3, 00:13:18.460 "num_base_bdevs_discovered": 3, 00:13:18.460 "num_base_bdevs_operational": 3, 00:13:18.460 "process": { 00:13:18.460 "type": "rebuild", 00:13:18.460 "target": "spare", 00:13:18.460 "progress": { 00:13:18.460 "blocks": 45056, 00:13:18.460 "percent": 35 00:13:18.460 } 00:13:18.460 }, 00:13:18.460 "base_bdevs_list": [ 00:13:18.460 { 00:13:18.460 "name": "spare", 00:13:18.460 "uuid": "a57a253c-5b63-5c17-a10d-19cb39fa9847", 00:13:18.460 "is_configured": true, 00:13:18.460 "data_offset": 2048, 00:13:18.460 "data_size": 63488 00:13:18.460 }, 00:13:18.460 { 00:13:18.460 "name": "BaseBdev2", 00:13:18.460 "uuid": "6f25ce10-ae25-567e-bda6-5434d31c5259", 00:13:18.460 "is_configured": true, 00:13:18.460 "data_offset": 2048, 00:13:18.460 "data_size": 63488 00:13:18.460 }, 00:13:18.460 { 00:13:18.460 "name": "BaseBdev3", 00:13:18.460 "uuid": "5538ab62-c9e7-5d6c-86e3-cd8f39be431c", 00:13:18.460 "is_configured": true, 00:13:18.460 "data_offset": 2048, 00:13:18.460 "data_size": 63488 00:13:18.460 } 00:13:18.460 ] 00:13:18.460 }' 00:13:18.460 23:47:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:18.460 23:47:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:18.460 23:47:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:18.460 23:47:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:18.460 23:47:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:13:19.397 23:47:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:19.397 23:47:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:19.397 23:47:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:19.397 23:47:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:19.397 23:47:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:19.397 23:47:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:19.397 23:47:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:19.397 23:47:07 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:19.397 23:47:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:19.397 23:47:07 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:19.397 23:47:07 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:19.397 23:47:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:19.397 "name": "raid_bdev1", 00:13:19.397 "uuid": "aeb0da41-a687-43e0-974e-937ec7519a0e", 00:13:19.397 "strip_size_kb": 64, 00:13:19.397 "state": "online", 00:13:19.397 "raid_level": "raid5f", 00:13:19.397 "superblock": true, 00:13:19.397 "num_base_bdevs": 3, 00:13:19.397 "num_base_bdevs_discovered": 3, 00:13:19.397 "num_base_bdevs_operational": 3, 00:13:19.397 "process": { 00:13:19.397 "type": "rebuild", 00:13:19.397 "target": "spare", 00:13:19.397 "progress": { 00:13:19.397 "blocks": 67584, 00:13:19.397 "percent": 53 00:13:19.397 } 00:13:19.397 }, 00:13:19.397 "base_bdevs_list": [ 00:13:19.397 { 00:13:19.397 "name": "spare", 00:13:19.397 "uuid": "a57a253c-5b63-5c17-a10d-19cb39fa9847", 00:13:19.397 "is_configured": true, 00:13:19.397 "data_offset": 2048, 00:13:19.397 "data_size": 63488 00:13:19.397 }, 00:13:19.397 { 00:13:19.397 "name": "BaseBdev2", 00:13:19.397 "uuid": "6f25ce10-ae25-567e-bda6-5434d31c5259", 00:13:19.397 "is_configured": true, 00:13:19.397 "data_offset": 2048, 00:13:19.398 "data_size": 63488 00:13:19.398 }, 00:13:19.398 { 00:13:19.398 "name": "BaseBdev3", 00:13:19.398 "uuid": "5538ab62-c9e7-5d6c-86e3-cd8f39be431c", 00:13:19.398 "is_configured": true, 00:13:19.398 "data_offset": 2048, 00:13:19.398 "data_size": 63488 00:13:19.398 } 00:13:19.398 ] 00:13:19.398 }' 00:13:19.398 23:47:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:19.398 23:47:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:19.398 23:47:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:19.657 23:47:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:19.657 23:47:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:13:20.596 23:47:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:20.596 23:47:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:20.596 23:47:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:20.596 23:47:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:20.596 23:47:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:20.596 23:47:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:20.596 23:47:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:20.596 23:47:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:20.597 23:47:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:20.597 23:47:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:20.597 23:47:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:20.597 23:47:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:20.597 "name": "raid_bdev1", 00:13:20.597 "uuid": "aeb0da41-a687-43e0-974e-937ec7519a0e", 00:13:20.597 "strip_size_kb": 64, 00:13:20.597 "state": "online", 00:13:20.597 "raid_level": "raid5f", 00:13:20.597 "superblock": true, 00:13:20.597 "num_base_bdevs": 3, 00:13:20.597 "num_base_bdevs_discovered": 3, 00:13:20.597 "num_base_bdevs_operational": 3, 00:13:20.597 "process": { 00:13:20.597 "type": "rebuild", 00:13:20.597 "target": "spare", 00:13:20.597 "progress": { 00:13:20.597 "blocks": 92160, 00:13:20.597 "percent": 72 00:13:20.597 } 00:13:20.597 }, 00:13:20.597 "base_bdevs_list": [ 00:13:20.597 { 00:13:20.597 "name": "spare", 00:13:20.597 "uuid": "a57a253c-5b63-5c17-a10d-19cb39fa9847", 00:13:20.597 "is_configured": true, 00:13:20.597 "data_offset": 2048, 00:13:20.597 "data_size": 63488 00:13:20.597 }, 00:13:20.597 { 00:13:20.597 "name": "BaseBdev2", 00:13:20.597 "uuid": "6f25ce10-ae25-567e-bda6-5434d31c5259", 00:13:20.597 "is_configured": true, 00:13:20.597 "data_offset": 2048, 00:13:20.597 "data_size": 63488 00:13:20.597 }, 00:13:20.597 { 00:13:20.597 "name": "BaseBdev3", 00:13:20.597 "uuid": "5538ab62-c9e7-5d6c-86e3-cd8f39be431c", 00:13:20.597 "is_configured": true, 00:13:20.597 "data_offset": 2048, 00:13:20.597 "data_size": 63488 00:13:20.597 } 00:13:20.597 ] 00:13:20.597 }' 00:13:20.597 23:47:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:20.597 23:47:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:20.597 23:47:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:20.597 23:47:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:20.597 23:47:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:13:21.977 23:47:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:21.977 23:47:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:21.977 23:47:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:21.977 23:47:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:21.977 23:47:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:21.977 23:47:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:21.977 23:47:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:21.977 23:47:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:21.977 23:47:09 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:21.977 23:47:09 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:21.977 23:47:09 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:21.977 23:47:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:21.977 "name": "raid_bdev1", 00:13:21.977 "uuid": "aeb0da41-a687-43e0-974e-937ec7519a0e", 00:13:21.977 "strip_size_kb": 64, 00:13:21.977 "state": "online", 00:13:21.977 "raid_level": "raid5f", 00:13:21.977 "superblock": true, 00:13:21.977 "num_base_bdevs": 3, 00:13:21.977 "num_base_bdevs_discovered": 3, 00:13:21.977 "num_base_bdevs_operational": 3, 00:13:21.977 "process": { 00:13:21.977 "type": "rebuild", 00:13:21.977 "target": "spare", 00:13:21.977 "progress": { 00:13:21.977 "blocks": 114688, 00:13:21.977 "percent": 90 00:13:21.977 } 00:13:21.977 }, 00:13:21.977 "base_bdevs_list": [ 00:13:21.977 { 00:13:21.977 "name": "spare", 00:13:21.977 "uuid": "a57a253c-5b63-5c17-a10d-19cb39fa9847", 00:13:21.977 "is_configured": true, 00:13:21.977 "data_offset": 2048, 00:13:21.977 "data_size": 63488 00:13:21.977 }, 00:13:21.977 { 00:13:21.977 "name": "BaseBdev2", 00:13:21.977 "uuid": "6f25ce10-ae25-567e-bda6-5434d31c5259", 00:13:21.977 "is_configured": true, 00:13:21.977 "data_offset": 2048, 00:13:21.977 "data_size": 63488 00:13:21.977 }, 00:13:21.977 { 00:13:21.977 "name": "BaseBdev3", 00:13:21.977 "uuid": "5538ab62-c9e7-5d6c-86e3-cd8f39be431c", 00:13:21.977 "is_configured": true, 00:13:21.977 "data_offset": 2048, 00:13:21.977 "data_size": 63488 00:13:21.977 } 00:13:21.977 ] 00:13:21.977 }' 00:13:21.977 23:47:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:21.977 23:47:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:21.977 23:47:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:21.977 23:47:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:21.977 23:47:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:13:22.236 [2024-11-26 23:47:10.251283] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:13:22.236 [2024-11-26 23:47:10.251348] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:13:22.236 [2024-11-26 23:47:10.251483] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:22.805 23:47:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:22.805 23:47:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:22.805 23:47:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:22.805 23:47:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:22.805 23:47:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:22.805 23:47:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:22.805 23:47:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:22.805 23:47:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:22.805 23:47:10 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:22.805 23:47:10 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:22.805 23:47:10 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:22.805 23:47:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:22.805 "name": "raid_bdev1", 00:13:22.805 "uuid": "aeb0da41-a687-43e0-974e-937ec7519a0e", 00:13:22.805 "strip_size_kb": 64, 00:13:22.805 "state": "online", 00:13:22.805 "raid_level": "raid5f", 00:13:22.805 "superblock": true, 00:13:22.805 "num_base_bdevs": 3, 00:13:22.805 "num_base_bdevs_discovered": 3, 00:13:22.805 "num_base_bdevs_operational": 3, 00:13:22.805 "base_bdevs_list": [ 00:13:22.805 { 00:13:22.805 "name": "spare", 00:13:22.805 "uuid": "a57a253c-5b63-5c17-a10d-19cb39fa9847", 00:13:22.805 "is_configured": true, 00:13:22.805 "data_offset": 2048, 00:13:22.805 "data_size": 63488 00:13:22.805 }, 00:13:22.805 { 00:13:22.805 "name": "BaseBdev2", 00:13:22.805 "uuid": "6f25ce10-ae25-567e-bda6-5434d31c5259", 00:13:22.805 "is_configured": true, 00:13:22.805 "data_offset": 2048, 00:13:22.805 "data_size": 63488 00:13:22.805 }, 00:13:22.805 { 00:13:22.805 "name": "BaseBdev3", 00:13:22.805 "uuid": "5538ab62-c9e7-5d6c-86e3-cd8f39be431c", 00:13:22.805 "is_configured": true, 00:13:22.805 "data_offset": 2048, 00:13:22.805 "data_size": 63488 00:13:22.805 } 00:13:22.805 ] 00:13:22.805 }' 00:13:22.805 23:47:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:23.065 23:47:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:13:23.065 23:47:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:23.065 23:47:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:13:23.066 23:47:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@709 -- # break 00:13:23.066 23:47:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:13:23.066 23:47:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:23.066 23:47:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:13:23.066 23:47:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:13:23.066 23:47:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:23.066 23:47:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:23.066 23:47:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:23.066 23:47:11 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:23.066 23:47:11 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:23.066 23:47:11 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:23.066 23:47:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:23.066 "name": "raid_bdev1", 00:13:23.066 "uuid": "aeb0da41-a687-43e0-974e-937ec7519a0e", 00:13:23.066 "strip_size_kb": 64, 00:13:23.066 "state": "online", 00:13:23.066 "raid_level": "raid5f", 00:13:23.066 "superblock": true, 00:13:23.066 "num_base_bdevs": 3, 00:13:23.066 "num_base_bdevs_discovered": 3, 00:13:23.066 "num_base_bdevs_operational": 3, 00:13:23.066 "base_bdevs_list": [ 00:13:23.066 { 00:13:23.066 "name": "spare", 00:13:23.066 "uuid": "a57a253c-5b63-5c17-a10d-19cb39fa9847", 00:13:23.066 "is_configured": true, 00:13:23.066 "data_offset": 2048, 00:13:23.066 "data_size": 63488 00:13:23.066 }, 00:13:23.066 { 00:13:23.066 "name": "BaseBdev2", 00:13:23.066 "uuid": "6f25ce10-ae25-567e-bda6-5434d31c5259", 00:13:23.066 "is_configured": true, 00:13:23.066 "data_offset": 2048, 00:13:23.066 "data_size": 63488 00:13:23.066 }, 00:13:23.066 { 00:13:23.066 "name": "BaseBdev3", 00:13:23.066 "uuid": "5538ab62-c9e7-5d6c-86e3-cd8f39be431c", 00:13:23.066 "is_configured": true, 00:13:23.066 "data_offset": 2048, 00:13:23.066 "data_size": 63488 00:13:23.066 } 00:13:23.066 ] 00:13:23.066 }' 00:13:23.066 23:47:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:23.066 23:47:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:13:23.066 23:47:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:23.066 23:47:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:13:23.066 23:47:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:13:23.066 23:47:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:23.066 23:47:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:23.066 23:47:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:23.066 23:47:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:23.066 23:47:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:23.066 23:47:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:23.066 23:47:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:23.066 23:47:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:23.066 23:47:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:23.066 23:47:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:23.066 23:47:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:23.066 23:47:11 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:23.066 23:47:11 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:23.066 23:47:11 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:23.066 23:47:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:23.066 "name": "raid_bdev1", 00:13:23.066 "uuid": "aeb0da41-a687-43e0-974e-937ec7519a0e", 00:13:23.066 "strip_size_kb": 64, 00:13:23.066 "state": "online", 00:13:23.066 "raid_level": "raid5f", 00:13:23.066 "superblock": true, 00:13:23.066 "num_base_bdevs": 3, 00:13:23.066 "num_base_bdevs_discovered": 3, 00:13:23.066 "num_base_bdevs_operational": 3, 00:13:23.066 "base_bdevs_list": [ 00:13:23.066 { 00:13:23.066 "name": "spare", 00:13:23.066 "uuid": "a57a253c-5b63-5c17-a10d-19cb39fa9847", 00:13:23.066 "is_configured": true, 00:13:23.066 "data_offset": 2048, 00:13:23.066 "data_size": 63488 00:13:23.066 }, 00:13:23.066 { 00:13:23.066 "name": "BaseBdev2", 00:13:23.066 "uuid": "6f25ce10-ae25-567e-bda6-5434d31c5259", 00:13:23.066 "is_configured": true, 00:13:23.066 "data_offset": 2048, 00:13:23.066 "data_size": 63488 00:13:23.066 }, 00:13:23.066 { 00:13:23.066 "name": "BaseBdev3", 00:13:23.066 "uuid": "5538ab62-c9e7-5d6c-86e3-cd8f39be431c", 00:13:23.066 "is_configured": true, 00:13:23.066 "data_offset": 2048, 00:13:23.066 "data_size": 63488 00:13:23.066 } 00:13:23.066 ] 00:13:23.066 }' 00:13:23.066 23:47:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:23.066 23:47:11 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:23.636 23:47:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:13:23.636 23:47:11 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:23.636 23:47:11 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:23.636 [2024-11-26 23:47:11.582675] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:13:23.636 [2024-11-26 23:47:11.582754] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:13:23.636 [2024-11-26 23:47:11.582862] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:13:23.636 [2024-11-26 23:47:11.582964] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:13:23.636 [2024-11-26 23:47:11.583024] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:13:23.636 23:47:11 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:23.636 23:47:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:23.636 23:47:11 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:23.636 23:47:11 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:23.636 23:47:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # jq length 00:13:23.636 23:47:11 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:23.636 23:47:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:13:23.636 23:47:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:13:23.636 23:47:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:13:23.636 23:47:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:13:23.636 23:47:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:13:23.636 23:47:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:13:23.636 23:47:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # local bdev_list 00:13:23.636 23:47:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:13:23.636 23:47:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # local nbd_list 00:13:23.636 23:47:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@12 -- # local i 00:13:23.636 23:47:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:13:23.636 23:47:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:13:23.636 23:47:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:13:23.896 /dev/nbd0 00:13:23.896 23:47:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:13:23.896 23:47:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:13:23.896 23:47:11 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:13:23.896 23:47:11 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@873 -- # local i 00:13:23.896 23:47:11 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:13:23.896 23:47:11 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:13:23.896 23:47:11 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:13:23.896 23:47:11 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@877 -- # break 00:13:23.896 23:47:11 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:13:23.896 23:47:11 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:13:23.896 23:47:11 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:13:23.896 1+0 records in 00:13:23.896 1+0 records out 00:13:23.896 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000202191 s, 20.3 MB/s 00:13:23.896 23:47:11 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:23.896 23:47:11 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@890 -- # size=4096 00:13:23.896 23:47:11 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:23.896 23:47:11 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:13:23.896 23:47:11 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@893 -- # return 0 00:13:23.896 23:47:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:13:23.896 23:47:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:13:23.896 23:47:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:13:24.156 /dev/nbd1 00:13:24.156 23:47:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:13:24.156 23:47:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:13:24.156 23:47:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:13:24.156 23:47:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@873 -- # local i 00:13:24.156 23:47:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:13:24.156 23:47:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:13:24.156 23:47:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:13:24.156 23:47:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@877 -- # break 00:13:24.156 23:47:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:13:24.156 23:47:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:13:24.156 23:47:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:13:24.156 1+0 records in 00:13:24.156 1+0 records out 00:13:24.156 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000405034 s, 10.1 MB/s 00:13:24.156 23:47:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:24.156 23:47:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@890 -- # size=4096 00:13:24.156 23:47:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:24.156 23:47:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:13:24.156 23:47:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@893 -- # return 0 00:13:24.156 23:47:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:13:24.156 23:47:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:13:24.156 23:47:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@738 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:13:24.156 23:47:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:13:24.156 23:47:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:13:24.156 23:47:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:13:24.156 23:47:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # local nbd_list 00:13:24.156 23:47:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@51 -- # local i 00:13:24.156 23:47:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:13:24.156 23:47:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:13:24.416 23:47:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:13:24.416 23:47:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:13:24.416 23:47:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:13:24.416 23:47:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:13:24.416 23:47:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:13:24.416 23:47:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:13:24.416 23:47:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:13:24.416 23:47:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:13:24.416 23:47:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:13:24.416 23:47:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:13:24.675 23:47:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:13:24.675 23:47:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:13:24.675 23:47:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:13:24.675 23:47:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:13:24.675 23:47:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:13:24.675 23:47:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:13:24.676 23:47:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:13:24.676 23:47:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:13:24.676 23:47:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@743 -- # '[' true = true ']' 00:13:24.676 23:47:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@745 -- # rpc_cmd bdev_passthru_delete spare 00:13:24.676 23:47:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:24.676 23:47:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:24.676 23:47:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:24.676 23:47:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@746 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:13:24.676 23:47:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:24.676 23:47:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:24.676 [2024-11-26 23:47:12.619436] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:13:24.676 [2024-11-26 23:47:12.619528] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:24.676 [2024-11-26 23:47:12.619565] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:13:24.676 [2024-11-26 23:47:12.619592] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:24.676 [2024-11-26 23:47:12.621687] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:24.676 [2024-11-26 23:47:12.621770] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:13:24.676 [2024-11-26 23:47:12.621873] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:13:24.676 [2024-11-26 23:47:12.621912] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:13:24.676 [2024-11-26 23:47:12.622051] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:13:24.676 [2024-11-26 23:47:12.622138] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:13:24.676 spare 00:13:24.676 23:47:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:24.676 23:47:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@747 -- # rpc_cmd bdev_wait_for_examine 00:13:24.676 23:47:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:24.676 23:47:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:24.676 [2024-11-26 23:47:12.722022] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001580 00:13:24.676 [2024-11-26 23:47:12.722046] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:13:24.676 [2024-11-26 23:47:12.722298] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000043d50 00:13:24.676 [2024-11-26 23:47:12.722749] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001580 00:13:24.676 [2024-11-26 23:47:12.722770] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001580 00:13:24.676 [2024-11-26 23:47:12.722897] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:24.676 23:47:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:24.676 23:47:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@749 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:13:24.676 23:47:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:24.676 23:47:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:24.676 23:47:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:24.676 23:47:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:24.676 23:47:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:24.676 23:47:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:24.676 23:47:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:24.676 23:47:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:24.676 23:47:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:24.676 23:47:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:24.676 23:47:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:24.676 23:47:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:24.676 23:47:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:24.676 23:47:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:24.676 23:47:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:24.676 "name": "raid_bdev1", 00:13:24.676 "uuid": "aeb0da41-a687-43e0-974e-937ec7519a0e", 00:13:24.676 "strip_size_kb": 64, 00:13:24.676 "state": "online", 00:13:24.676 "raid_level": "raid5f", 00:13:24.676 "superblock": true, 00:13:24.676 "num_base_bdevs": 3, 00:13:24.676 "num_base_bdevs_discovered": 3, 00:13:24.676 "num_base_bdevs_operational": 3, 00:13:24.676 "base_bdevs_list": [ 00:13:24.676 { 00:13:24.676 "name": "spare", 00:13:24.676 "uuid": "a57a253c-5b63-5c17-a10d-19cb39fa9847", 00:13:24.676 "is_configured": true, 00:13:24.676 "data_offset": 2048, 00:13:24.676 "data_size": 63488 00:13:24.676 }, 00:13:24.676 { 00:13:24.676 "name": "BaseBdev2", 00:13:24.676 "uuid": "6f25ce10-ae25-567e-bda6-5434d31c5259", 00:13:24.676 "is_configured": true, 00:13:24.676 "data_offset": 2048, 00:13:24.676 "data_size": 63488 00:13:24.676 }, 00:13:24.676 { 00:13:24.676 "name": "BaseBdev3", 00:13:24.676 "uuid": "5538ab62-c9e7-5d6c-86e3-cd8f39be431c", 00:13:24.676 "is_configured": true, 00:13:24.676 "data_offset": 2048, 00:13:24.676 "data_size": 63488 00:13:24.676 } 00:13:24.676 ] 00:13:24.676 }' 00:13:24.676 23:47:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:24.676 23:47:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:25.244 23:47:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@750 -- # verify_raid_bdev_process raid_bdev1 none none 00:13:25.244 23:47:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:25.244 23:47:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:13:25.244 23:47:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:13:25.244 23:47:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:25.244 23:47:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:25.244 23:47:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:25.244 23:47:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:25.244 23:47:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:25.244 23:47:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:25.244 23:47:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:25.244 "name": "raid_bdev1", 00:13:25.244 "uuid": "aeb0da41-a687-43e0-974e-937ec7519a0e", 00:13:25.244 "strip_size_kb": 64, 00:13:25.244 "state": "online", 00:13:25.244 "raid_level": "raid5f", 00:13:25.244 "superblock": true, 00:13:25.244 "num_base_bdevs": 3, 00:13:25.244 "num_base_bdevs_discovered": 3, 00:13:25.244 "num_base_bdevs_operational": 3, 00:13:25.244 "base_bdevs_list": [ 00:13:25.244 { 00:13:25.244 "name": "spare", 00:13:25.244 "uuid": "a57a253c-5b63-5c17-a10d-19cb39fa9847", 00:13:25.244 "is_configured": true, 00:13:25.244 "data_offset": 2048, 00:13:25.244 "data_size": 63488 00:13:25.244 }, 00:13:25.244 { 00:13:25.244 "name": "BaseBdev2", 00:13:25.244 "uuid": "6f25ce10-ae25-567e-bda6-5434d31c5259", 00:13:25.244 "is_configured": true, 00:13:25.244 "data_offset": 2048, 00:13:25.244 "data_size": 63488 00:13:25.244 }, 00:13:25.244 { 00:13:25.244 "name": "BaseBdev3", 00:13:25.244 "uuid": "5538ab62-c9e7-5d6c-86e3-cd8f39be431c", 00:13:25.244 "is_configured": true, 00:13:25.244 "data_offset": 2048, 00:13:25.244 "data_size": 63488 00:13:25.244 } 00:13:25.244 ] 00:13:25.244 }' 00:13:25.244 23:47:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:25.244 23:47:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:13:25.244 23:47:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:25.244 23:47:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:13:25.244 23:47:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:25.244 23:47:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # jq -r '.[].base_bdevs_list[0].name' 00:13:25.244 23:47:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:25.244 23:47:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:25.244 23:47:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:25.244 23:47:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # [[ spare == \s\p\a\r\e ]] 00:13:25.244 23:47:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@754 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:13:25.244 23:47:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:25.244 23:47:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:25.244 [2024-11-26 23:47:13.363011] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:25.244 23:47:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:25.244 23:47:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@755 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:13:25.244 23:47:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:25.244 23:47:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:25.244 23:47:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:25.244 23:47:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:25.244 23:47:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:25.244 23:47:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:25.244 23:47:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:25.244 23:47:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:25.244 23:47:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:25.503 23:47:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:25.503 23:47:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:25.503 23:47:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:25.503 23:47:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:25.503 23:47:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:25.503 23:47:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:25.503 "name": "raid_bdev1", 00:13:25.503 "uuid": "aeb0da41-a687-43e0-974e-937ec7519a0e", 00:13:25.503 "strip_size_kb": 64, 00:13:25.503 "state": "online", 00:13:25.503 "raid_level": "raid5f", 00:13:25.503 "superblock": true, 00:13:25.503 "num_base_bdevs": 3, 00:13:25.503 "num_base_bdevs_discovered": 2, 00:13:25.503 "num_base_bdevs_operational": 2, 00:13:25.503 "base_bdevs_list": [ 00:13:25.503 { 00:13:25.503 "name": null, 00:13:25.503 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:25.503 "is_configured": false, 00:13:25.503 "data_offset": 0, 00:13:25.503 "data_size": 63488 00:13:25.503 }, 00:13:25.503 { 00:13:25.503 "name": "BaseBdev2", 00:13:25.503 "uuid": "6f25ce10-ae25-567e-bda6-5434d31c5259", 00:13:25.503 "is_configured": true, 00:13:25.503 "data_offset": 2048, 00:13:25.504 "data_size": 63488 00:13:25.504 }, 00:13:25.504 { 00:13:25.504 "name": "BaseBdev3", 00:13:25.504 "uuid": "5538ab62-c9e7-5d6c-86e3-cd8f39be431c", 00:13:25.504 "is_configured": true, 00:13:25.504 "data_offset": 2048, 00:13:25.504 "data_size": 63488 00:13:25.504 } 00:13:25.504 ] 00:13:25.504 }' 00:13:25.504 23:47:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:25.504 23:47:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:25.763 23:47:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@756 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:13:25.763 23:47:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:25.763 23:47:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:25.763 [2024-11-26 23:47:13.746506] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:13:25.763 [2024-11-26 23:47:13.746725] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:13:25.763 [2024-11-26 23:47:13.746786] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:13:25.763 [2024-11-26 23:47:13.746858] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:13:25.763 [2024-11-26 23:47:13.751117] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000043e20 00:13:25.763 23:47:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:25.763 23:47:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@757 -- # sleep 1 00:13:25.763 [2024-11-26 23:47:13.753182] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:13:26.699 23:47:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@758 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:26.699 23:47:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:26.699 23:47:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:26.699 23:47:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:26.699 23:47:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:26.699 23:47:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:26.699 23:47:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:26.699 23:47:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:26.699 23:47:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:26.699 23:47:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:26.699 23:47:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:26.699 "name": "raid_bdev1", 00:13:26.699 "uuid": "aeb0da41-a687-43e0-974e-937ec7519a0e", 00:13:26.699 "strip_size_kb": 64, 00:13:26.699 "state": "online", 00:13:26.699 "raid_level": "raid5f", 00:13:26.699 "superblock": true, 00:13:26.699 "num_base_bdevs": 3, 00:13:26.699 "num_base_bdevs_discovered": 3, 00:13:26.699 "num_base_bdevs_operational": 3, 00:13:26.699 "process": { 00:13:26.699 "type": "rebuild", 00:13:26.699 "target": "spare", 00:13:26.699 "progress": { 00:13:26.699 "blocks": 20480, 00:13:26.699 "percent": 16 00:13:26.699 } 00:13:26.699 }, 00:13:26.699 "base_bdevs_list": [ 00:13:26.699 { 00:13:26.699 "name": "spare", 00:13:26.699 "uuid": "a57a253c-5b63-5c17-a10d-19cb39fa9847", 00:13:26.699 "is_configured": true, 00:13:26.699 "data_offset": 2048, 00:13:26.699 "data_size": 63488 00:13:26.699 }, 00:13:26.699 { 00:13:26.699 "name": "BaseBdev2", 00:13:26.699 "uuid": "6f25ce10-ae25-567e-bda6-5434d31c5259", 00:13:26.699 "is_configured": true, 00:13:26.699 "data_offset": 2048, 00:13:26.699 "data_size": 63488 00:13:26.699 }, 00:13:26.699 { 00:13:26.699 "name": "BaseBdev3", 00:13:26.699 "uuid": "5538ab62-c9e7-5d6c-86e3-cd8f39be431c", 00:13:26.699 "is_configured": true, 00:13:26.699 "data_offset": 2048, 00:13:26.699 "data_size": 63488 00:13:26.699 } 00:13:26.699 ] 00:13:26.699 }' 00:13:26.699 23:47:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:26.957 23:47:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:26.957 23:47:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:26.957 23:47:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:26.958 23:47:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@761 -- # rpc_cmd bdev_passthru_delete spare 00:13:26.958 23:47:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:26.958 23:47:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:26.958 [2024-11-26 23:47:14.909767] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:26.958 [2024-11-26 23:47:14.960003] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:13:26.958 [2024-11-26 23:47:14.960054] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:26.958 [2024-11-26 23:47:14.960070] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:26.958 [2024-11-26 23:47:14.960077] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:13:26.958 23:47:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:26.958 23:47:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@762 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:13:26.958 23:47:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:26.958 23:47:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:26.958 23:47:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:26.958 23:47:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:26.958 23:47:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:26.958 23:47:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:26.958 23:47:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:26.958 23:47:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:26.958 23:47:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:26.958 23:47:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:26.958 23:47:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:26.958 23:47:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:26.958 23:47:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:26.958 23:47:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:26.958 23:47:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:26.958 "name": "raid_bdev1", 00:13:26.958 "uuid": "aeb0da41-a687-43e0-974e-937ec7519a0e", 00:13:26.958 "strip_size_kb": 64, 00:13:26.958 "state": "online", 00:13:26.958 "raid_level": "raid5f", 00:13:26.958 "superblock": true, 00:13:26.958 "num_base_bdevs": 3, 00:13:26.958 "num_base_bdevs_discovered": 2, 00:13:26.958 "num_base_bdevs_operational": 2, 00:13:26.958 "base_bdevs_list": [ 00:13:26.958 { 00:13:26.958 "name": null, 00:13:26.958 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:26.958 "is_configured": false, 00:13:26.958 "data_offset": 0, 00:13:26.958 "data_size": 63488 00:13:26.958 }, 00:13:26.958 { 00:13:26.958 "name": "BaseBdev2", 00:13:26.958 "uuid": "6f25ce10-ae25-567e-bda6-5434d31c5259", 00:13:26.958 "is_configured": true, 00:13:26.958 "data_offset": 2048, 00:13:26.958 "data_size": 63488 00:13:26.958 }, 00:13:26.958 { 00:13:26.958 "name": "BaseBdev3", 00:13:26.958 "uuid": "5538ab62-c9e7-5d6c-86e3-cd8f39be431c", 00:13:26.958 "is_configured": true, 00:13:26.958 "data_offset": 2048, 00:13:26.958 "data_size": 63488 00:13:26.958 } 00:13:26.958 ] 00:13:26.958 }' 00:13:26.958 23:47:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:26.958 23:47:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:27.525 23:47:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@763 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:13:27.525 23:47:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:27.525 23:47:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:27.525 [2024-11-26 23:47:15.404587] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:13:27.525 [2024-11-26 23:47:15.404682] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:27.525 [2024-11-26 23:47:15.404720] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ab80 00:13:27.525 [2024-11-26 23:47:15.404746] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:27.525 [2024-11-26 23:47:15.405186] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:27.525 [2024-11-26 23:47:15.405241] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:13:27.525 [2024-11-26 23:47:15.405352] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:13:27.525 [2024-11-26 23:47:15.405391] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:13:27.525 [2024-11-26 23:47:15.405451] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:13:27.525 [2024-11-26 23:47:15.405496] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:13:27.525 [2024-11-26 23:47:15.409588] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000043ef0 00:13:27.525 spare 00:13:27.525 23:47:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:27.525 23:47:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@764 -- # sleep 1 00:13:27.525 [2024-11-26 23:47:15.411677] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:13:28.461 23:47:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@765 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:28.461 23:47:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:28.461 23:47:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:28.461 23:47:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:28.461 23:47:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:28.461 23:47:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:28.461 23:47:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:28.461 23:47:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:28.461 23:47:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:28.461 23:47:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:28.461 23:47:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:28.461 "name": "raid_bdev1", 00:13:28.461 "uuid": "aeb0da41-a687-43e0-974e-937ec7519a0e", 00:13:28.461 "strip_size_kb": 64, 00:13:28.461 "state": "online", 00:13:28.461 "raid_level": "raid5f", 00:13:28.461 "superblock": true, 00:13:28.461 "num_base_bdevs": 3, 00:13:28.461 "num_base_bdevs_discovered": 3, 00:13:28.461 "num_base_bdevs_operational": 3, 00:13:28.461 "process": { 00:13:28.461 "type": "rebuild", 00:13:28.461 "target": "spare", 00:13:28.461 "progress": { 00:13:28.461 "blocks": 20480, 00:13:28.461 "percent": 16 00:13:28.461 } 00:13:28.461 }, 00:13:28.461 "base_bdevs_list": [ 00:13:28.461 { 00:13:28.461 "name": "spare", 00:13:28.461 "uuid": "a57a253c-5b63-5c17-a10d-19cb39fa9847", 00:13:28.461 "is_configured": true, 00:13:28.461 "data_offset": 2048, 00:13:28.461 "data_size": 63488 00:13:28.461 }, 00:13:28.461 { 00:13:28.461 "name": "BaseBdev2", 00:13:28.461 "uuid": "6f25ce10-ae25-567e-bda6-5434d31c5259", 00:13:28.461 "is_configured": true, 00:13:28.461 "data_offset": 2048, 00:13:28.461 "data_size": 63488 00:13:28.461 }, 00:13:28.461 { 00:13:28.461 "name": "BaseBdev3", 00:13:28.461 "uuid": "5538ab62-c9e7-5d6c-86e3-cd8f39be431c", 00:13:28.461 "is_configured": true, 00:13:28.461 "data_offset": 2048, 00:13:28.461 "data_size": 63488 00:13:28.461 } 00:13:28.461 ] 00:13:28.461 }' 00:13:28.461 23:47:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:28.461 23:47:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:28.461 23:47:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:28.461 23:47:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:28.461 23:47:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@768 -- # rpc_cmd bdev_passthru_delete spare 00:13:28.461 23:47:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:28.461 23:47:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:28.461 [2024-11-26 23:47:16.571583] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:28.721 [2024-11-26 23:47:16.618231] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:13:28.721 [2024-11-26 23:47:16.618283] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:28.721 [2024-11-26 23:47:16.618297] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:28.721 [2024-11-26 23:47:16.618307] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:13:28.721 23:47:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:28.721 23:47:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@769 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:13:28.721 23:47:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:28.721 23:47:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:28.721 23:47:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:28.721 23:47:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:28.721 23:47:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:28.721 23:47:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:28.721 23:47:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:28.721 23:47:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:28.721 23:47:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:28.721 23:47:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:28.721 23:47:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:28.721 23:47:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:28.721 23:47:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:28.721 23:47:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:28.721 23:47:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:28.721 "name": "raid_bdev1", 00:13:28.721 "uuid": "aeb0da41-a687-43e0-974e-937ec7519a0e", 00:13:28.721 "strip_size_kb": 64, 00:13:28.721 "state": "online", 00:13:28.721 "raid_level": "raid5f", 00:13:28.721 "superblock": true, 00:13:28.721 "num_base_bdevs": 3, 00:13:28.721 "num_base_bdevs_discovered": 2, 00:13:28.721 "num_base_bdevs_operational": 2, 00:13:28.721 "base_bdevs_list": [ 00:13:28.721 { 00:13:28.721 "name": null, 00:13:28.721 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:28.721 "is_configured": false, 00:13:28.721 "data_offset": 0, 00:13:28.721 "data_size": 63488 00:13:28.721 }, 00:13:28.721 { 00:13:28.721 "name": "BaseBdev2", 00:13:28.721 "uuid": "6f25ce10-ae25-567e-bda6-5434d31c5259", 00:13:28.721 "is_configured": true, 00:13:28.721 "data_offset": 2048, 00:13:28.721 "data_size": 63488 00:13:28.721 }, 00:13:28.721 { 00:13:28.721 "name": "BaseBdev3", 00:13:28.721 "uuid": "5538ab62-c9e7-5d6c-86e3-cd8f39be431c", 00:13:28.721 "is_configured": true, 00:13:28.721 "data_offset": 2048, 00:13:28.721 "data_size": 63488 00:13:28.721 } 00:13:28.721 ] 00:13:28.721 }' 00:13:28.721 23:47:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:28.721 23:47:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:28.981 23:47:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@770 -- # verify_raid_bdev_process raid_bdev1 none none 00:13:28.981 23:47:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:28.981 23:47:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:13:28.981 23:47:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:13:28.981 23:47:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:28.981 23:47:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:28.981 23:47:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:28.981 23:47:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:28.981 23:47:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:28.981 23:47:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:28.981 23:47:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:28.981 "name": "raid_bdev1", 00:13:28.981 "uuid": "aeb0da41-a687-43e0-974e-937ec7519a0e", 00:13:28.981 "strip_size_kb": 64, 00:13:28.981 "state": "online", 00:13:28.981 "raid_level": "raid5f", 00:13:28.981 "superblock": true, 00:13:28.981 "num_base_bdevs": 3, 00:13:28.981 "num_base_bdevs_discovered": 2, 00:13:28.981 "num_base_bdevs_operational": 2, 00:13:28.981 "base_bdevs_list": [ 00:13:28.981 { 00:13:28.981 "name": null, 00:13:28.981 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:28.981 "is_configured": false, 00:13:28.981 "data_offset": 0, 00:13:28.981 "data_size": 63488 00:13:28.981 }, 00:13:28.981 { 00:13:28.981 "name": "BaseBdev2", 00:13:28.981 "uuid": "6f25ce10-ae25-567e-bda6-5434d31c5259", 00:13:28.981 "is_configured": true, 00:13:28.981 "data_offset": 2048, 00:13:28.981 "data_size": 63488 00:13:28.981 }, 00:13:28.981 { 00:13:28.981 "name": "BaseBdev3", 00:13:28.981 "uuid": "5538ab62-c9e7-5d6c-86e3-cd8f39be431c", 00:13:28.981 "is_configured": true, 00:13:28.981 "data_offset": 2048, 00:13:28.981 "data_size": 63488 00:13:28.981 } 00:13:28.981 ] 00:13:28.981 }' 00:13:28.981 23:47:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:28.981 23:47:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:13:28.981 23:47:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:29.242 23:47:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:13:29.242 23:47:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@773 -- # rpc_cmd bdev_passthru_delete BaseBdev1 00:13:29.242 23:47:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:29.242 23:47:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:29.242 23:47:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:29.242 23:47:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@774 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:13:29.242 23:47:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:29.242 23:47:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:29.242 [2024-11-26 23:47:17.146687] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:13:29.242 [2024-11-26 23:47:17.146739] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:29.242 [2024-11-26 23:47:17.146757] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b180 00:13:29.242 [2024-11-26 23:47:17.146768] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:29.242 [2024-11-26 23:47:17.147144] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:29.242 [2024-11-26 23:47:17.147163] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:13:29.242 [2024-11-26 23:47:17.147225] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:13:29.242 [2024-11-26 23:47:17.147239] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:13:29.242 [2024-11-26 23:47:17.147247] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:13:29.242 [2024-11-26 23:47:17.147258] bdev_raid.c:3894:raid_bdev_examine_done: *ERROR*: Failed to examine bdev BaseBdev1: Invalid argument 00:13:29.242 BaseBdev1 00:13:29.242 23:47:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:29.242 23:47:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@775 -- # sleep 1 00:13:30.180 23:47:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@776 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:13:30.180 23:47:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:30.180 23:47:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:30.180 23:47:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:30.180 23:47:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:30.180 23:47:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:30.180 23:47:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:30.180 23:47:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:30.180 23:47:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:30.180 23:47:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:30.180 23:47:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:30.180 23:47:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:30.180 23:47:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:30.180 23:47:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:30.180 23:47:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:30.180 23:47:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:30.180 "name": "raid_bdev1", 00:13:30.180 "uuid": "aeb0da41-a687-43e0-974e-937ec7519a0e", 00:13:30.180 "strip_size_kb": 64, 00:13:30.180 "state": "online", 00:13:30.180 "raid_level": "raid5f", 00:13:30.180 "superblock": true, 00:13:30.180 "num_base_bdevs": 3, 00:13:30.180 "num_base_bdevs_discovered": 2, 00:13:30.180 "num_base_bdevs_operational": 2, 00:13:30.180 "base_bdevs_list": [ 00:13:30.180 { 00:13:30.180 "name": null, 00:13:30.180 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:30.180 "is_configured": false, 00:13:30.180 "data_offset": 0, 00:13:30.180 "data_size": 63488 00:13:30.180 }, 00:13:30.180 { 00:13:30.180 "name": "BaseBdev2", 00:13:30.180 "uuid": "6f25ce10-ae25-567e-bda6-5434d31c5259", 00:13:30.180 "is_configured": true, 00:13:30.180 "data_offset": 2048, 00:13:30.180 "data_size": 63488 00:13:30.180 }, 00:13:30.180 { 00:13:30.180 "name": "BaseBdev3", 00:13:30.180 "uuid": "5538ab62-c9e7-5d6c-86e3-cd8f39be431c", 00:13:30.180 "is_configured": true, 00:13:30.180 "data_offset": 2048, 00:13:30.180 "data_size": 63488 00:13:30.180 } 00:13:30.180 ] 00:13:30.180 }' 00:13:30.180 23:47:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:30.180 23:47:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:30.749 23:47:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@777 -- # verify_raid_bdev_process raid_bdev1 none none 00:13:30.749 23:47:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:30.749 23:47:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:13:30.749 23:47:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:13:30.749 23:47:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:30.749 23:47:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:30.749 23:47:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:30.749 23:47:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:30.749 23:47:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:30.749 23:47:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:30.749 23:47:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:30.749 "name": "raid_bdev1", 00:13:30.749 "uuid": "aeb0da41-a687-43e0-974e-937ec7519a0e", 00:13:30.749 "strip_size_kb": 64, 00:13:30.749 "state": "online", 00:13:30.749 "raid_level": "raid5f", 00:13:30.749 "superblock": true, 00:13:30.749 "num_base_bdevs": 3, 00:13:30.749 "num_base_bdevs_discovered": 2, 00:13:30.749 "num_base_bdevs_operational": 2, 00:13:30.749 "base_bdevs_list": [ 00:13:30.749 { 00:13:30.749 "name": null, 00:13:30.749 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:30.749 "is_configured": false, 00:13:30.749 "data_offset": 0, 00:13:30.749 "data_size": 63488 00:13:30.749 }, 00:13:30.749 { 00:13:30.749 "name": "BaseBdev2", 00:13:30.749 "uuid": "6f25ce10-ae25-567e-bda6-5434d31c5259", 00:13:30.749 "is_configured": true, 00:13:30.749 "data_offset": 2048, 00:13:30.749 "data_size": 63488 00:13:30.749 }, 00:13:30.749 { 00:13:30.749 "name": "BaseBdev3", 00:13:30.749 "uuid": "5538ab62-c9e7-5d6c-86e3-cd8f39be431c", 00:13:30.749 "is_configured": true, 00:13:30.749 "data_offset": 2048, 00:13:30.749 "data_size": 63488 00:13:30.749 } 00:13:30.749 ] 00:13:30.749 }' 00:13:30.749 23:47:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:30.749 23:47:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:13:30.749 23:47:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:30.749 23:47:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:13:30.749 23:47:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@778 -- # NOT rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:13:30.749 23:47:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@652 -- # local es=0 00:13:30.749 23:47:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:13:30.749 23:47:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:13:30.749 23:47:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:13:30.749 23:47:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:13:30.749 23:47:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:13:30.749 23:47:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:13:30.749 23:47:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:30.749 23:47:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:30.749 [2024-11-26 23:47:18.732238] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:13:30.749 [2024-11-26 23:47:18.732398] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:13:30.749 [2024-11-26 23:47:18.732414] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:13:30.749 request: 00:13:30.749 { 00:13:30.749 "base_bdev": "BaseBdev1", 00:13:30.749 "raid_bdev": "raid_bdev1", 00:13:30.749 "method": "bdev_raid_add_base_bdev", 00:13:30.749 "req_id": 1 00:13:30.749 } 00:13:30.749 Got JSON-RPC error response 00:13:30.749 response: 00:13:30.749 { 00:13:30.749 "code": -22, 00:13:30.749 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:13:30.749 } 00:13:30.749 23:47:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:13:30.749 23:47:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@655 -- # es=1 00:13:30.749 23:47:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:13:30.749 23:47:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:13:30.749 23:47:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:13:30.749 23:47:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@779 -- # sleep 1 00:13:31.701 23:47:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@780 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:13:31.701 23:47:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:31.701 23:47:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:31.701 23:47:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:31.701 23:47:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:31.701 23:47:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:31.701 23:47:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:31.701 23:47:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:31.701 23:47:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:31.701 23:47:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:31.701 23:47:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:31.701 23:47:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:31.701 23:47:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:31.701 23:47:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:31.701 23:47:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:31.701 23:47:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:31.701 "name": "raid_bdev1", 00:13:31.701 "uuid": "aeb0da41-a687-43e0-974e-937ec7519a0e", 00:13:31.701 "strip_size_kb": 64, 00:13:31.701 "state": "online", 00:13:31.701 "raid_level": "raid5f", 00:13:31.701 "superblock": true, 00:13:31.701 "num_base_bdevs": 3, 00:13:31.701 "num_base_bdevs_discovered": 2, 00:13:31.701 "num_base_bdevs_operational": 2, 00:13:31.701 "base_bdevs_list": [ 00:13:31.701 { 00:13:31.701 "name": null, 00:13:31.701 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:31.701 "is_configured": false, 00:13:31.701 "data_offset": 0, 00:13:31.701 "data_size": 63488 00:13:31.701 }, 00:13:31.701 { 00:13:31.701 "name": "BaseBdev2", 00:13:31.701 "uuid": "6f25ce10-ae25-567e-bda6-5434d31c5259", 00:13:31.701 "is_configured": true, 00:13:31.701 "data_offset": 2048, 00:13:31.701 "data_size": 63488 00:13:31.701 }, 00:13:31.701 { 00:13:31.701 "name": "BaseBdev3", 00:13:31.701 "uuid": "5538ab62-c9e7-5d6c-86e3-cd8f39be431c", 00:13:31.701 "is_configured": true, 00:13:31.701 "data_offset": 2048, 00:13:31.701 "data_size": 63488 00:13:31.701 } 00:13:31.701 ] 00:13:31.701 }' 00:13:31.701 23:47:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:31.701 23:47:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:32.271 23:47:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@781 -- # verify_raid_bdev_process raid_bdev1 none none 00:13:32.271 23:47:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:32.271 23:47:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:13:32.271 23:47:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:13:32.271 23:47:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:32.271 23:47:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:32.271 23:47:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:32.271 23:47:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:32.271 23:47:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:32.271 23:47:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:32.271 23:47:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:32.271 "name": "raid_bdev1", 00:13:32.271 "uuid": "aeb0da41-a687-43e0-974e-937ec7519a0e", 00:13:32.271 "strip_size_kb": 64, 00:13:32.271 "state": "online", 00:13:32.271 "raid_level": "raid5f", 00:13:32.271 "superblock": true, 00:13:32.271 "num_base_bdevs": 3, 00:13:32.271 "num_base_bdevs_discovered": 2, 00:13:32.271 "num_base_bdevs_operational": 2, 00:13:32.271 "base_bdevs_list": [ 00:13:32.271 { 00:13:32.271 "name": null, 00:13:32.271 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:32.271 "is_configured": false, 00:13:32.271 "data_offset": 0, 00:13:32.271 "data_size": 63488 00:13:32.271 }, 00:13:32.271 { 00:13:32.271 "name": "BaseBdev2", 00:13:32.271 "uuid": "6f25ce10-ae25-567e-bda6-5434d31c5259", 00:13:32.271 "is_configured": true, 00:13:32.271 "data_offset": 2048, 00:13:32.271 "data_size": 63488 00:13:32.271 }, 00:13:32.271 { 00:13:32.271 "name": "BaseBdev3", 00:13:32.271 "uuid": "5538ab62-c9e7-5d6c-86e3-cd8f39be431c", 00:13:32.271 "is_configured": true, 00:13:32.271 "data_offset": 2048, 00:13:32.271 "data_size": 63488 00:13:32.271 } 00:13:32.271 ] 00:13:32.271 }' 00:13:32.271 23:47:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:32.271 23:47:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:13:32.271 23:47:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:32.271 23:47:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:13:32.271 23:47:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@784 -- # killprocess 92190 00:13:32.271 23:47:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@954 -- # '[' -z 92190 ']' 00:13:32.271 23:47:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@958 -- # kill -0 92190 00:13:32.271 23:47:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@959 -- # uname 00:13:32.271 23:47:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:13:32.271 23:47:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 92190 00:13:32.271 23:47:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:13:32.271 23:47:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:13:32.271 23:47:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@972 -- # echo 'killing process with pid 92190' 00:13:32.271 killing process with pid 92190 00:13:32.271 23:47:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@973 -- # kill 92190 00:13:32.271 Received shutdown signal, test time was about 60.000000 seconds 00:13:32.271 00:13:32.271 Latency(us) 00:13:32.271 [2024-11-26T23:47:20.403Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:13:32.271 [2024-11-26T23:47:20.403Z] =================================================================================================================== 00:13:32.271 [2024-11-26T23:47:20.403Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:13:32.271 [2024-11-26 23:47:20.299798] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:13:32.271 [2024-11-26 23:47:20.299906] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:13:32.271 23:47:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@978 -- # wait 92190 00:13:32.271 [2024-11-26 23:47:20.299967] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:13:32.271 [2024-11-26 23:47:20.299976] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state offline 00:13:32.271 [2024-11-26 23:47:20.340109] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:13:32.531 23:47:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@786 -- # return 0 00:13:32.531 00:13:32.531 real 0m21.088s 00:13:32.531 user 0m27.399s 00:13:32.531 sys 0m2.484s 00:13:32.531 23:47:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@1130 -- # xtrace_disable 00:13:32.531 ************************************ 00:13:32.531 END TEST raid5f_rebuild_test_sb 00:13:32.531 ************************************ 00:13:32.531 23:47:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:32.531 23:47:20 bdev_raid -- bdev/bdev_raid.sh@985 -- # for n in {3..4} 00:13:32.531 23:47:20 bdev_raid -- bdev/bdev_raid.sh@986 -- # run_test raid5f_state_function_test raid_state_function_test raid5f 4 false 00:13:32.531 23:47:20 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:13:32.531 23:47:20 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:13:32.531 23:47:20 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:13:32.531 ************************************ 00:13:32.531 START TEST raid5f_state_function_test 00:13:32.531 ************************************ 00:13:32.531 23:47:20 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@1129 -- # raid_state_function_test raid5f 4 false 00:13:32.531 23:47:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=raid5f 00:13:32.531 23:47:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=4 00:13:32.531 23:47:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:13:32.531 23:47:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:13:32.531 23:47:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:13:32.531 23:47:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:13:32.531 23:47:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:13:32.531 23:47:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:13:32.531 23:47:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:13:32.531 23:47:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:13:32.531 23:47:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:13:32.532 23:47:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:13:32.532 23:47:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:13:32.532 23:47:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:13:32.532 23:47:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:13:32.532 23:47:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev4 00:13:32.532 23:47:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:13:32.532 23:47:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:13:32.532 23:47:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:13:32.532 23:47:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:13:32.532 23:47:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:13:32.532 23:47:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:13:32.532 23:47:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:13:32.532 23:47:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:13:32.532 23:47:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' raid5f '!=' raid1 ']' 00:13:32.532 23:47:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:13:32.532 23:47:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:13:32.532 Process raid pid: 92920 00:13:32.532 23:47:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:13:32.532 23:47:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:13:32.532 23:47:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=92920 00:13:32.532 23:47:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:13:32.532 23:47:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 92920' 00:13:32.532 23:47:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 92920 00:13:32.532 23:47:20 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@835 -- # '[' -z 92920 ']' 00:13:32.532 23:47:20 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:13:32.532 23:47:20 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:13:32.532 23:47:20 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:13:32.532 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:13:32.532 23:47:20 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:13:32.532 23:47:20 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:32.792 [2024-11-26 23:47:20.704707] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:13:32.792 [2024-11-26 23:47:20.704910] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:13:32.792 [2024-11-26 23:47:20.855059] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:13:32.792 [2024-11-26 23:47:20.879547] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:13:33.052 [2024-11-26 23:47:20.922016] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:13:33.052 [2024-11-26 23:47:20.922121] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:13:33.621 23:47:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:13:33.622 23:47:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@868 -- # return 0 00:13:33.622 23:47:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:13:33.622 23:47:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:33.622 23:47:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:33.622 [2024-11-26 23:47:21.520811] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:13:33.622 [2024-11-26 23:47:21.520919] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:13:33.622 [2024-11-26 23:47:21.520964] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:13:33.622 [2024-11-26 23:47:21.520987] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:13:33.622 [2024-11-26 23:47:21.521005] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:13:33.622 [2024-11-26 23:47:21.521028] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:13:33.622 [2024-11-26 23:47:21.521045] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:13:33.622 [2024-11-26 23:47:21.521065] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:13:33.622 23:47:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:33.622 23:47:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:13:33.622 23:47:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:33.622 23:47:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:33.622 23:47:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:33.622 23:47:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:33.622 23:47:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:13:33.622 23:47:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:33.622 23:47:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:33.622 23:47:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:33.622 23:47:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:33.622 23:47:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:33.622 23:47:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:33.622 23:47:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:33.622 23:47:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:33.622 23:47:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:33.622 23:47:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:33.622 "name": "Existed_Raid", 00:13:33.622 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:33.622 "strip_size_kb": 64, 00:13:33.622 "state": "configuring", 00:13:33.622 "raid_level": "raid5f", 00:13:33.622 "superblock": false, 00:13:33.622 "num_base_bdevs": 4, 00:13:33.622 "num_base_bdevs_discovered": 0, 00:13:33.622 "num_base_bdevs_operational": 4, 00:13:33.622 "base_bdevs_list": [ 00:13:33.622 { 00:13:33.622 "name": "BaseBdev1", 00:13:33.622 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:33.622 "is_configured": false, 00:13:33.622 "data_offset": 0, 00:13:33.622 "data_size": 0 00:13:33.622 }, 00:13:33.622 { 00:13:33.622 "name": "BaseBdev2", 00:13:33.622 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:33.622 "is_configured": false, 00:13:33.622 "data_offset": 0, 00:13:33.622 "data_size": 0 00:13:33.622 }, 00:13:33.622 { 00:13:33.622 "name": "BaseBdev3", 00:13:33.622 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:33.622 "is_configured": false, 00:13:33.622 "data_offset": 0, 00:13:33.622 "data_size": 0 00:13:33.622 }, 00:13:33.622 { 00:13:33.622 "name": "BaseBdev4", 00:13:33.622 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:33.622 "is_configured": false, 00:13:33.622 "data_offset": 0, 00:13:33.622 "data_size": 0 00:13:33.622 } 00:13:33.622 ] 00:13:33.622 }' 00:13:33.622 23:47:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:33.622 23:47:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:33.882 23:47:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:13:33.882 23:47:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:33.882 23:47:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:33.882 [2024-11-26 23:47:21.912059] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:13:33.882 [2024-11-26 23:47:21.912134] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:13:33.882 23:47:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:33.882 23:47:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:13:33.882 23:47:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:33.882 23:47:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:33.882 [2024-11-26 23:47:21.920073] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:13:33.882 [2024-11-26 23:47:21.920114] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:13:33.883 [2024-11-26 23:47:21.920123] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:13:33.883 [2024-11-26 23:47:21.920132] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:13:33.883 [2024-11-26 23:47:21.920138] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:13:33.883 [2024-11-26 23:47:21.920145] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:13:33.883 [2024-11-26 23:47:21.920151] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:13:33.883 [2024-11-26 23:47:21.920159] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:13:33.883 23:47:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:33.883 23:47:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:13:33.883 23:47:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:33.883 23:47:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:33.883 [2024-11-26 23:47:21.936982] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:13:33.883 BaseBdev1 00:13:33.883 23:47:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:33.883 23:47:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:13:33.883 23:47:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:13:33.883 23:47:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:13:33.883 23:47:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:13:33.883 23:47:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:13:33.883 23:47:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:13:33.883 23:47:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:13:33.883 23:47:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:33.883 23:47:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:33.883 23:47:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:33.883 23:47:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:13:33.883 23:47:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:33.883 23:47:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:33.883 [ 00:13:33.883 { 00:13:33.883 "name": "BaseBdev1", 00:13:33.883 "aliases": [ 00:13:33.883 "8ee86c6d-adff-4712-921c-1b9a589b639d" 00:13:33.883 ], 00:13:33.883 "product_name": "Malloc disk", 00:13:33.883 "block_size": 512, 00:13:33.883 "num_blocks": 65536, 00:13:33.883 "uuid": "8ee86c6d-adff-4712-921c-1b9a589b639d", 00:13:33.883 "assigned_rate_limits": { 00:13:33.883 "rw_ios_per_sec": 0, 00:13:33.883 "rw_mbytes_per_sec": 0, 00:13:33.883 "r_mbytes_per_sec": 0, 00:13:33.883 "w_mbytes_per_sec": 0 00:13:33.883 }, 00:13:33.883 "claimed": true, 00:13:33.883 "claim_type": "exclusive_write", 00:13:33.883 "zoned": false, 00:13:33.883 "supported_io_types": { 00:13:33.883 "read": true, 00:13:33.883 "write": true, 00:13:33.883 "unmap": true, 00:13:33.883 "flush": true, 00:13:33.883 "reset": true, 00:13:33.883 "nvme_admin": false, 00:13:33.883 "nvme_io": false, 00:13:33.883 "nvme_io_md": false, 00:13:33.883 "write_zeroes": true, 00:13:33.883 "zcopy": true, 00:13:33.883 "get_zone_info": false, 00:13:33.883 "zone_management": false, 00:13:33.883 "zone_append": false, 00:13:33.883 "compare": false, 00:13:33.883 "compare_and_write": false, 00:13:33.883 "abort": true, 00:13:33.883 "seek_hole": false, 00:13:33.883 "seek_data": false, 00:13:33.883 "copy": true, 00:13:33.883 "nvme_iov_md": false 00:13:33.883 }, 00:13:33.883 "memory_domains": [ 00:13:33.883 { 00:13:33.883 "dma_device_id": "system", 00:13:33.883 "dma_device_type": 1 00:13:33.883 }, 00:13:33.883 { 00:13:33.883 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:33.883 "dma_device_type": 2 00:13:33.883 } 00:13:33.883 ], 00:13:33.883 "driver_specific": {} 00:13:33.883 } 00:13:33.883 ] 00:13:33.883 23:47:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:33.883 23:47:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:13:33.883 23:47:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:13:33.883 23:47:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:33.883 23:47:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:33.883 23:47:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:33.883 23:47:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:33.883 23:47:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:13:33.883 23:47:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:33.883 23:47:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:33.883 23:47:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:33.883 23:47:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:33.883 23:47:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:33.883 23:47:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:33.883 23:47:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:33.883 23:47:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:33.883 23:47:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:34.143 23:47:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:34.143 "name": "Existed_Raid", 00:13:34.143 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:34.143 "strip_size_kb": 64, 00:13:34.143 "state": "configuring", 00:13:34.144 "raid_level": "raid5f", 00:13:34.144 "superblock": false, 00:13:34.144 "num_base_bdevs": 4, 00:13:34.144 "num_base_bdevs_discovered": 1, 00:13:34.144 "num_base_bdevs_operational": 4, 00:13:34.144 "base_bdevs_list": [ 00:13:34.144 { 00:13:34.144 "name": "BaseBdev1", 00:13:34.144 "uuid": "8ee86c6d-adff-4712-921c-1b9a589b639d", 00:13:34.144 "is_configured": true, 00:13:34.144 "data_offset": 0, 00:13:34.144 "data_size": 65536 00:13:34.144 }, 00:13:34.144 { 00:13:34.144 "name": "BaseBdev2", 00:13:34.144 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:34.144 "is_configured": false, 00:13:34.144 "data_offset": 0, 00:13:34.144 "data_size": 0 00:13:34.144 }, 00:13:34.144 { 00:13:34.144 "name": "BaseBdev3", 00:13:34.144 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:34.144 "is_configured": false, 00:13:34.144 "data_offset": 0, 00:13:34.144 "data_size": 0 00:13:34.144 }, 00:13:34.144 { 00:13:34.144 "name": "BaseBdev4", 00:13:34.144 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:34.144 "is_configured": false, 00:13:34.144 "data_offset": 0, 00:13:34.144 "data_size": 0 00:13:34.144 } 00:13:34.144 ] 00:13:34.144 }' 00:13:34.144 23:47:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:34.144 23:47:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:34.402 23:47:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:13:34.402 23:47:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:34.402 23:47:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:34.402 [2024-11-26 23:47:22.420228] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:13:34.402 [2024-11-26 23:47:22.420272] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:13:34.402 23:47:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:34.402 23:47:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:13:34.402 23:47:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:34.402 23:47:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:34.402 [2024-11-26 23:47:22.432250] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:13:34.402 [2024-11-26 23:47:22.434046] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:13:34.402 [2024-11-26 23:47:22.434091] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:13:34.402 [2024-11-26 23:47:22.434100] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:13:34.402 [2024-11-26 23:47:22.434109] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:13:34.402 [2024-11-26 23:47:22.434115] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:13:34.402 [2024-11-26 23:47:22.434123] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:13:34.402 23:47:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:34.402 23:47:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:13:34.402 23:47:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:13:34.402 23:47:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:13:34.402 23:47:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:34.403 23:47:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:34.403 23:47:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:34.403 23:47:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:34.403 23:47:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:13:34.403 23:47:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:34.403 23:47:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:34.403 23:47:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:34.403 23:47:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:34.403 23:47:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:34.403 23:47:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:34.403 23:47:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:34.403 23:47:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:34.403 23:47:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:34.403 23:47:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:34.403 "name": "Existed_Raid", 00:13:34.403 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:34.403 "strip_size_kb": 64, 00:13:34.403 "state": "configuring", 00:13:34.403 "raid_level": "raid5f", 00:13:34.403 "superblock": false, 00:13:34.403 "num_base_bdevs": 4, 00:13:34.403 "num_base_bdevs_discovered": 1, 00:13:34.403 "num_base_bdevs_operational": 4, 00:13:34.403 "base_bdevs_list": [ 00:13:34.403 { 00:13:34.403 "name": "BaseBdev1", 00:13:34.403 "uuid": "8ee86c6d-adff-4712-921c-1b9a589b639d", 00:13:34.403 "is_configured": true, 00:13:34.403 "data_offset": 0, 00:13:34.403 "data_size": 65536 00:13:34.403 }, 00:13:34.403 { 00:13:34.403 "name": "BaseBdev2", 00:13:34.403 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:34.403 "is_configured": false, 00:13:34.403 "data_offset": 0, 00:13:34.403 "data_size": 0 00:13:34.403 }, 00:13:34.403 { 00:13:34.403 "name": "BaseBdev3", 00:13:34.403 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:34.403 "is_configured": false, 00:13:34.403 "data_offset": 0, 00:13:34.403 "data_size": 0 00:13:34.403 }, 00:13:34.403 { 00:13:34.403 "name": "BaseBdev4", 00:13:34.403 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:34.403 "is_configured": false, 00:13:34.403 "data_offset": 0, 00:13:34.403 "data_size": 0 00:13:34.403 } 00:13:34.403 ] 00:13:34.403 }' 00:13:34.403 23:47:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:34.403 23:47:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:34.972 23:47:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:13:34.972 23:47:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:34.972 23:47:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:34.972 [2024-11-26 23:47:22.890168] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:13:34.972 BaseBdev2 00:13:34.972 23:47:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:34.972 23:47:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:13:34.972 23:47:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:13:34.972 23:47:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:13:34.972 23:47:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:13:34.972 23:47:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:13:34.972 23:47:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:13:34.972 23:47:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:13:34.972 23:47:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:34.972 23:47:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:34.972 23:47:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:34.972 23:47:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:13:34.972 23:47:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:34.972 23:47:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:34.972 [ 00:13:34.972 { 00:13:34.972 "name": "BaseBdev2", 00:13:34.972 "aliases": [ 00:13:34.972 "ce3736e2-ae89-4868-b996-d519e5cb185e" 00:13:34.972 ], 00:13:34.972 "product_name": "Malloc disk", 00:13:34.972 "block_size": 512, 00:13:34.972 "num_blocks": 65536, 00:13:34.972 "uuid": "ce3736e2-ae89-4868-b996-d519e5cb185e", 00:13:34.972 "assigned_rate_limits": { 00:13:34.972 "rw_ios_per_sec": 0, 00:13:34.972 "rw_mbytes_per_sec": 0, 00:13:34.972 "r_mbytes_per_sec": 0, 00:13:34.972 "w_mbytes_per_sec": 0 00:13:34.972 }, 00:13:34.972 "claimed": true, 00:13:34.972 "claim_type": "exclusive_write", 00:13:34.972 "zoned": false, 00:13:34.972 "supported_io_types": { 00:13:34.972 "read": true, 00:13:34.972 "write": true, 00:13:34.972 "unmap": true, 00:13:34.972 "flush": true, 00:13:34.972 "reset": true, 00:13:34.972 "nvme_admin": false, 00:13:34.972 "nvme_io": false, 00:13:34.972 "nvme_io_md": false, 00:13:34.972 "write_zeroes": true, 00:13:34.972 "zcopy": true, 00:13:34.972 "get_zone_info": false, 00:13:34.972 "zone_management": false, 00:13:34.972 "zone_append": false, 00:13:34.972 "compare": false, 00:13:34.972 "compare_and_write": false, 00:13:34.972 "abort": true, 00:13:34.972 "seek_hole": false, 00:13:34.972 "seek_data": false, 00:13:34.972 "copy": true, 00:13:34.972 "nvme_iov_md": false 00:13:34.972 }, 00:13:34.972 "memory_domains": [ 00:13:34.972 { 00:13:34.972 "dma_device_id": "system", 00:13:34.972 "dma_device_type": 1 00:13:34.972 }, 00:13:34.972 { 00:13:34.972 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:34.972 "dma_device_type": 2 00:13:34.972 } 00:13:34.972 ], 00:13:34.972 "driver_specific": {} 00:13:34.972 } 00:13:34.972 ] 00:13:34.972 23:47:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:34.972 23:47:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:13:34.972 23:47:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:13:34.972 23:47:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:13:34.972 23:47:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:13:34.972 23:47:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:34.972 23:47:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:34.972 23:47:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:34.972 23:47:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:34.972 23:47:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:13:34.972 23:47:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:34.972 23:47:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:34.972 23:47:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:34.972 23:47:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:34.972 23:47:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:34.972 23:47:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:34.972 23:47:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:34.972 23:47:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:34.972 23:47:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:34.972 23:47:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:34.972 "name": "Existed_Raid", 00:13:34.972 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:34.972 "strip_size_kb": 64, 00:13:34.972 "state": "configuring", 00:13:34.972 "raid_level": "raid5f", 00:13:34.972 "superblock": false, 00:13:34.972 "num_base_bdevs": 4, 00:13:34.972 "num_base_bdevs_discovered": 2, 00:13:34.972 "num_base_bdevs_operational": 4, 00:13:34.972 "base_bdevs_list": [ 00:13:34.972 { 00:13:34.972 "name": "BaseBdev1", 00:13:34.972 "uuid": "8ee86c6d-adff-4712-921c-1b9a589b639d", 00:13:34.972 "is_configured": true, 00:13:34.972 "data_offset": 0, 00:13:34.972 "data_size": 65536 00:13:34.972 }, 00:13:34.972 { 00:13:34.972 "name": "BaseBdev2", 00:13:34.972 "uuid": "ce3736e2-ae89-4868-b996-d519e5cb185e", 00:13:34.972 "is_configured": true, 00:13:34.972 "data_offset": 0, 00:13:34.972 "data_size": 65536 00:13:34.972 }, 00:13:34.972 { 00:13:34.972 "name": "BaseBdev3", 00:13:34.972 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:34.972 "is_configured": false, 00:13:34.972 "data_offset": 0, 00:13:34.972 "data_size": 0 00:13:34.972 }, 00:13:34.972 { 00:13:34.972 "name": "BaseBdev4", 00:13:34.972 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:34.972 "is_configured": false, 00:13:34.972 "data_offset": 0, 00:13:34.972 "data_size": 0 00:13:34.972 } 00:13:34.972 ] 00:13:34.972 }' 00:13:34.972 23:47:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:34.972 23:47:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:35.232 23:47:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:13:35.232 23:47:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:35.232 23:47:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:35.492 [2024-11-26 23:47:23.375249] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:13:35.492 BaseBdev3 00:13:35.492 23:47:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:35.492 23:47:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:13:35.492 23:47:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:13:35.492 23:47:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:13:35.492 23:47:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:13:35.492 23:47:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:13:35.492 23:47:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:13:35.492 23:47:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:13:35.492 23:47:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:35.492 23:47:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:35.492 23:47:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:35.492 23:47:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:13:35.492 23:47:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:35.492 23:47:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:35.492 [ 00:13:35.492 { 00:13:35.492 "name": "BaseBdev3", 00:13:35.492 "aliases": [ 00:13:35.492 "756f3b10-35c2-4f69-816d-b4b19be2fff3" 00:13:35.492 ], 00:13:35.492 "product_name": "Malloc disk", 00:13:35.492 "block_size": 512, 00:13:35.492 "num_blocks": 65536, 00:13:35.492 "uuid": "756f3b10-35c2-4f69-816d-b4b19be2fff3", 00:13:35.492 "assigned_rate_limits": { 00:13:35.492 "rw_ios_per_sec": 0, 00:13:35.492 "rw_mbytes_per_sec": 0, 00:13:35.492 "r_mbytes_per_sec": 0, 00:13:35.492 "w_mbytes_per_sec": 0 00:13:35.492 }, 00:13:35.492 "claimed": true, 00:13:35.492 "claim_type": "exclusive_write", 00:13:35.492 "zoned": false, 00:13:35.492 "supported_io_types": { 00:13:35.492 "read": true, 00:13:35.492 "write": true, 00:13:35.492 "unmap": true, 00:13:35.492 "flush": true, 00:13:35.492 "reset": true, 00:13:35.492 "nvme_admin": false, 00:13:35.492 "nvme_io": false, 00:13:35.492 "nvme_io_md": false, 00:13:35.492 "write_zeroes": true, 00:13:35.492 "zcopy": true, 00:13:35.492 "get_zone_info": false, 00:13:35.492 "zone_management": false, 00:13:35.492 "zone_append": false, 00:13:35.492 "compare": false, 00:13:35.492 "compare_and_write": false, 00:13:35.492 "abort": true, 00:13:35.492 "seek_hole": false, 00:13:35.492 "seek_data": false, 00:13:35.492 "copy": true, 00:13:35.492 "nvme_iov_md": false 00:13:35.492 }, 00:13:35.492 "memory_domains": [ 00:13:35.492 { 00:13:35.492 "dma_device_id": "system", 00:13:35.492 "dma_device_type": 1 00:13:35.492 }, 00:13:35.492 { 00:13:35.492 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:35.492 "dma_device_type": 2 00:13:35.492 } 00:13:35.492 ], 00:13:35.492 "driver_specific": {} 00:13:35.492 } 00:13:35.492 ] 00:13:35.492 23:47:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:35.492 23:47:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:13:35.492 23:47:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:13:35.492 23:47:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:13:35.492 23:47:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:13:35.492 23:47:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:35.492 23:47:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:35.492 23:47:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:35.492 23:47:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:35.492 23:47:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:13:35.492 23:47:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:35.492 23:47:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:35.492 23:47:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:35.492 23:47:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:35.492 23:47:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:35.492 23:47:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:35.492 23:47:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:35.492 23:47:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:35.492 23:47:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:35.492 23:47:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:35.492 "name": "Existed_Raid", 00:13:35.492 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:35.492 "strip_size_kb": 64, 00:13:35.492 "state": "configuring", 00:13:35.492 "raid_level": "raid5f", 00:13:35.492 "superblock": false, 00:13:35.492 "num_base_bdevs": 4, 00:13:35.492 "num_base_bdevs_discovered": 3, 00:13:35.492 "num_base_bdevs_operational": 4, 00:13:35.492 "base_bdevs_list": [ 00:13:35.492 { 00:13:35.492 "name": "BaseBdev1", 00:13:35.492 "uuid": "8ee86c6d-adff-4712-921c-1b9a589b639d", 00:13:35.492 "is_configured": true, 00:13:35.492 "data_offset": 0, 00:13:35.492 "data_size": 65536 00:13:35.492 }, 00:13:35.492 { 00:13:35.492 "name": "BaseBdev2", 00:13:35.492 "uuid": "ce3736e2-ae89-4868-b996-d519e5cb185e", 00:13:35.492 "is_configured": true, 00:13:35.492 "data_offset": 0, 00:13:35.492 "data_size": 65536 00:13:35.492 }, 00:13:35.492 { 00:13:35.492 "name": "BaseBdev3", 00:13:35.492 "uuid": "756f3b10-35c2-4f69-816d-b4b19be2fff3", 00:13:35.492 "is_configured": true, 00:13:35.492 "data_offset": 0, 00:13:35.492 "data_size": 65536 00:13:35.492 }, 00:13:35.492 { 00:13:35.492 "name": "BaseBdev4", 00:13:35.492 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:35.492 "is_configured": false, 00:13:35.492 "data_offset": 0, 00:13:35.492 "data_size": 0 00:13:35.492 } 00:13:35.492 ] 00:13:35.492 }' 00:13:35.492 23:47:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:35.492 23:47:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:35.752 23:47:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:13:35.752 23:47:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:35.752 23:47:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:35.752 [2024-11-26 23:47:23.805147] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:13:35.752 [2024-11-26 23:47:23.805203] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:13:35.752 [2024-11-26 23:47:23.805211] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 196608, blocklen 512 00:13:35.752 [2024-11-26 23:47:23.805504] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:13:35.752 [2024-11-26 23:47:23.805978] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:13:35.752 [2024-11-26 23:47:23.805997] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:13:35.752 BaseBdev4 00:13:35.752 [2024-11-26 23:47:23.806196] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:35.752 23:47:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:35.752 23:47:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev4 00:13:35.752 23:47:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:13:35.752 23:47:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:13:35.752 23:47:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:13:35.752 23:47:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:13:35.752 23:47:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:13:35.752 23:47:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:13:35.752 23:47:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:35.752 23:47:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:35.752 23:47:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:35.752 23:47:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:13:35.752 23:47:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:35.752 23:47:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:35.752 [ 00:13:35.752 { 00:13:35.752 "name": "BaseBdev4", 00:13:35.752 "aliases": [ 00:13:35.752 "75351cef-4289-4786-8487-7b601f67f8eb" 00:13:35.752 ], 00:13:35.752 "product_name": "Malloc disk", 00:13:35.752 "block_size": 512, 00:13:35.752 "num_blocks": 65536, 00:13:35.752 "uuid": "75351cef-4289-4786-8487-7b601f67f8eb", 00:13:35.752 "assigned_rate_limits": { 00:13:35.752 "rw_ios_per_sec": 0, 00:13:35.752 "rw_mbytes_per_sec": 0, 00:13:35.752 "r_mbytes_per_sec": 0, 00:13:35.752 "w_mbytes_per_sec": 0 00:13:35.752 }, 00:13:35.752 "claimed": true, 00:13:35.752 "claim_type": "exclusive_write", 00:13:35.752 "zoned": false, 00:13:35.752 "supported_io_types": { 00:13:35.752 "read": true, 00:13:35.752 "write": true, 00:13:35.752 "unmap": true, 00:13:35.752 "flush": true, 00:13:35.752 "reset": true, 00:13:35.752 "nvme_admin": false, 00:13:35.752 "nvme_io": false, 00:13:35.752 "nvme_io_md": false, 00:13:35.752 "write_zeroes": true, 00:13:35.752 "zcopy": true, 00:13:35.752 "get_zone_info": false, 00:13:35.752 "zone_management": false, 00:13:35.752 "zone_append": false, 00:13:35.752 "compare": false, 00:13:35.752 "compare_and_write": false, 00:13:35.752 "abort": true, 00:13:35.752 "seek_hole": false, 00:13:35.752 "seek_data": false, 00:13:35.752 "copy": true, 00:13:35.752 "nvme_iov_md": false 00:13:35.752 }, 00:13:35.752 "memory_domains": [ 00:13:35.752 { 00:13:35.752 "dma_device_id": "system", 00:13:35.752 "dma_device_type": 1 00:13:35.752 }, 00:13:35.752 { 00:13:35.752 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:35.752 "dma_device_type": 2 00:13:35.752 } 00:13:35.752 ], 00:13:35.752 "driver_specific": {} 00:13:35.752 } 00:13:35.752 ] 00:13:35.752 23:47:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:35.752 23:47:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:13:35.752 23:47:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:13:35.752 23:47:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:13:35.752 23:47:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 4 00:13:35.752 23:47:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:35.752 23:47:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:35.752 23:47:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:35.752 23:47:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:35.752 23:47:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:13:35.752 23:47:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:35.752 23:47:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:35.752 23:47:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:35.752 23:47:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:35.752 23:47:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:35.752 23:47:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:35.752 23:47:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:35.752 23:47:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:35.752 23:47:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:35.752 23:47:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:35.752 "name": "Existed_Raid", 00:13:35.752 "uuid": "d4169941-3969-4706-a84a-3b0f9c26e3f3", 00:13:35.752 "strip_size_kb": 64, 00:13:35.752 "state": "online", 00:13:35.752 "raid_level": "raid5f", 00:13:35.752 "superblock": false, 00:13:35.752 "num_base_bdevs": 4, 00:13:35.752 "num_base_bdevs_discovered": 4, 00:13:35.752 "num_base_bdevs_operational": 4, 00:13:35.752 "base_bdevs_list": [ 00:13:35.752 { 00:13:35.752 "name": "BaseBdev1", 00:13:35.752 "uuid": "8ee86c6d-adff-4712-921c-1b9a589b639d", 00:13:35.752 "is_configured": true, 00:13:35.752 "data_offset": 0, 00:13:35.752 "data_size": 65536 00:13:35.752 }, 00:13:35.752 { 00:13:35.752 "name": "BaseBdev2", 00:13:35.752 "uuid": "ce3736e2-ae89-4868-b996-d519e5cb185e", 00:13:35.752 "is_configured": true, 00:13:35.752 "data_offset": 0, 00:13:35.752 "data_size": 65536 00:13:35.752 }, 00:13:35.752 { 00:13:35.752 "name": "BaseBdev3", 00:13:35.752 "uuid": "756f3b10-35c2-4f69-816d-b4b19be2fff3", 00:13:35.752 "is_configured": true, 00:13:35.752 "data_offset": 0, 00:13:35.752 "data_size": 65536 00:13:35.752 }, 00:13:35.752 { 00:13:35.752 "name": "BaseBdev4", 00:13:35.752 "uuid": "75351cef-4289-4786-8487-7b601f67f8eb", 00:13:35.752 "is_configured": true, 00:13:35.752 "data_offset": 0, 00:13:35.752 "data_size": 65536 00:13:35.752 } 00:13:35.752 ] 00:13:35.752 }' 00:13:35.752 23:47:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:35.752 23:47:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:36.321 23:47:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:13:36.321 23:47:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:13:36.321 23:47:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:13:36.321 23:47:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:13:36.321 23:47:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:13:36.321 23:47:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:13:36.321 23:47:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:13:36.321 23:47:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:36.321 23:47:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:36.321 23:47:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:13:36.321 [2024-11-26 23:47:24.212667] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:13:36.321 23:47:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:36.321 23:47:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:13:36.321 "name": "Existed_Raid", 00:13:36.321 "aliases": [ 00:13:36.321 "d4169941-3969-4706-a84a-3b0f9c26e3f3" 00:13:36.321 ], 00:13:36.321 "product_name": "Raid Volume", 00:13:36.321 "block_size": 512, 00:13:36.321 "num_blocks": 196608, 00:13:36.321 "uuid": "d4169941-3969-4706-a84a-3b0f9c26e3f3", 00:13:36.321 "assigned_rate_limits": { 00:13:36.321 "rw_ios_per_sec": 0, 00:13:36.321 "rw_mbytes_per_sec": 0, 00:13:36.321 "r_mbytes_per_sec": 0, 00:13:36.321 "w_mbytes_per_sec": 0 00:13:36.321 }, 00:13:36.321 "claimed": false, 00:13:36.321 "zoned": false, 00:13:36.321 "supported_io_types": { 00:13:36.321 "read": true, 00:13:36.321 "write": true, 00:13:36.321 "unmap": false, 00:13:36.321 "flush": false, 00:13:36.321 "reset": true, 00:13:36.321 "nvme_admin": false, 00:13:36.321 "nvme_io": false, 00:13:36.321 "nvme_io_md": false, 00:13:36.321 "write_zeroes": true, 00:13:36.321 "zcopy": false, 00:13:36.321 "get_zone_info": false, 00:13:36.321 "zone_management": false, 00:13:36.321 "zone_append": false, 00:13:36.321 "compare": false, 00:13:36.321 "compare_and_write": false, 00:13:36.321 "abort": false, 00:13:36.321 "seek_hole": false, 00:13:36.321 "seek_data": false, 00:13:36.321 "copy": false, 00:13:36.321 "nvme_iov_md": false 00:13:36.321 }, 00:13:36.321 "driver_specific": { 00:13:36.321 "raid": { 00:13:36.321 "uuid": "d4169941-3969-4706-a84a-3b0f9c26e3f3", 00:13:36.321 "strip_size_kb": 64, 00:13:36.321 "state": "online", 00:13:36.321 "raid_level": "raid5f", 00:13:36.321 "superblock": false, 00:13:36.321 "num_base_bdevs": 4, 00:13:36.321 "num_base_bdevs_discovered": 4, 00:13:36.321 "num_base_bdevs_operational": 4, 00:13:36.321 "base_bdevs_list": [ 00:13:36.321 { 00:13:36.321 "name": "BaseBdev1", 00:13:36.321 "uuid": "8ee86c6d-adff-4712-921c-1b9a589b639d", 00:13:36.321 "is_configured": true, 00:13:36.321 "data_offset": 0, 00:13:36.321 "data_size": 65536 00:13:36.321 }, 00:13:36.321 { 00:13:36.321 "name": "BaseBdev2", 00:13:36.321 "uuid": "ce3736e2-ae89-4868-b996-d519e5cb185e", 00:13:36.321 "is_configured": true, 00:13:36.321 "data_offset": 0, 00:13:36.321 "data_size": 65536 00:13:36.321 }, 00:13:36.321 { 00:13:36.321 "name": "BaseBdev3", 00:13:36.321 "uuid": "756f3b10-35c2-4f69-816d-b4b19be2fff3", 00:13:36.321 "is_configured": true, 00:13:36.321 "data_offset": 0, 00:13:36.321 "data_size": 65536 00:13:36.321 }, 00:13:36.321 { 00:13:36.321 "name": "BaseBdev4", 00:13:36.321 "uuid": "75351cef-4289-4786-8487-7b601f67f8eb", 00:13:36.321 "is_configured": true, 00:13:36.321 "data_offset": 0, 00:13:36.321 "data_size": 65536 00:13:36.321 } 00:13:36.321 ] 00:13:36.321 } 00:13:36.321 } 00:13:36.321 }' 00:13:36.321 23:47:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:13:36.321 23:47:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:13:36.321 BaseBdev2 00:13:36.321 BaseBdev3 00:13:36.321 BaseBdev4' 00:13:36.321 23:47:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:36.321 23:47:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:13:36.321 23:47:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:36.321 23:47:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:13:36.321 23:47:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:36.321 23:47:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:36.321 23:47:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:36.321 23:47:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:36.321 23:47:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:36.321 23:47:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:36.321 23:47:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:36.321 23:47:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:36.321 23:47:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:13:36.321 23:47:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:36.321 23:47:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:36.321 23:47:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:36.322 23:47:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:36.322 23:47:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:36.322 23:47:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:36.322 23:47:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:13:36.322 23:47:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:36.322 23:47:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:36.322 23:47:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:36.322 23:47:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:36.582 23:47:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:36.582 23:47:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:36.582 23:47:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:36.582 23:47:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:36.582 23:47:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:13:36.582 23:47:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:36.582 23:47:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:36.582 23:47:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:36.582 23:47:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:36.582 23:47:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:36.582 23:47:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:13:36.582 23:47:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:36.582 23:47:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:36.582 [2024-11-26 23:47:24.504013] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:13:36.582 23:47:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:36.582 23:47:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:13:36.582 23:47:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy raid5f 00:13:36.582 23:47:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:13:36.582 23:47:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@199 -- # return 0 00:13:36.582 23:47:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:13:36.582 23:47:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 3 00:13:36.582 23:47:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:36.582 23:47:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:36.582 23:47:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:36.582 23:47:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:36.582 23:47:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:36.582 23:47:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:36.582 23:47:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:36.582 23:47:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:36.582 23:47:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:36.582 23:47:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:36.582 23:47:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:36.582 23:47:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:36.582 23:47:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:36.582 23:47:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:36.582 23:47:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:36.582 "name": "Existed_Raid", 00:13:36.582 "uuid": "d4169941-3969-4706-a84a-3b0f9c26e3f3", 00:13:36.582 "strip_size_kb": 64, 00:13:36.582 "state": "online", 00:13:36.582 "raid_level": "raid5f", 00:13:36.582 "superblock": false, 00:13:36.582 "num_base_bdevs": 4, 00:13:36.582 "num_base_bdevs_discovered": 3, 00:13:36.582 "num_base_bdevs_operational": 3, 00:13:36.582 "base_bdevs_list": [ 00:13:36.582 { 00:13:36.582 "name": null, 00:13:36.582 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:36.582 "is_configured": false, 00:13:36.582 "data_offset": 0, 00:13:36.582 "data_size": 65536 00:13:36.582 }, 00:13:36.582 { 00:13:36.582 "name": "BaseBdev2", 00:13:36.582 "uuid": "ce3736e2-ae89-4868-b996-d519e5cb185e", 00:13:36.582 "is_configured": true, 00:13:36.582 "data_offset": 0, 00:13:36.582 "data_size": 65536 00:13:36.582 }, 00:13:36.582 { 00:13:36.582 "name": "BaseBdev3", 00:13:36.582 "uuid": "756f3b10-35c2-4f69-816d-b4b19be2fff3", 00:13:36.582 "is_configured": true, 00:13:36.582 "data_offset": 0, 00:13:36.582 "data_size": 65536 00:13:36.582 }, 00:13:36.582 { 00:13:36.582 "name": "BaseBdev4", 00:13:36.582 "uuid": "75351cef-4289-4786-8487-7b601f67f8eb", 00:13:36.582 "is_configured": true, 00:13:36.582 "data_offset": 0, 00:13:36.582 "data_size": 65536 00:13:36.582 } 00:13:36.582 ] 00:13:36.582 }' 00:13:36.582 23:47:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:36.582 23:47:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:36.848 23:47:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:13:36.848 23:47:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:13:36.848 23:47:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:36.848 23:47:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:13:36.848 23:47:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:36.848 23:47:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:36.848 23:47:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:36.848 23:47:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:13:37.121 23:47:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:13:37.121 23:47:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:13:37.121 23:47:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:37.121 23:47:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:37.121 [2024-11-26 23:47:24.978415] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:13:37.121 [2024-11-26 23:47:24.978562] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:13:37.121 [2024-11-26 23:47:24.989422] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:13:37.121 23:47:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:37.121 23:47:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:13:37.121 23:47:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:13:37.121 23:47:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:13:37.121 23:47:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:37.121 23:47:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:37.121 23:47:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:37.121 23:47:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:37.121 23:47:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:13:37.121 23:47:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:13:37.121 23:47:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:13:37.121 23:47:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:37.121 23:47:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:37.121 [2024-11-26 23:47:25.049326] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:13:37.121 23:47:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:37.121 23:47:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:13:37.121 23:47:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:13:37.121 23:47:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:37.121 23:47:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:37.121 23:47:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:37.121 23:47:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:13:37.121 23:47:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:37.121 23:47:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:13:37.121 23:47:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:13:37.121 23:47:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev4 00:13:37.121 23:47:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:37.121 23:47:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:37.121 [2024-11-26 23:47:25.120083] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev4 00:13:37.121 [2024-11-26 23:47:25.120125] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:13:37.121 23:47:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:37.121 23:47:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:13:37.121 23:47:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:13:37.121 23:47:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:37.121 23:47:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:13:37.121 23:47:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:37.121 23:47:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:37.121 23:47:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:37.121 23:47:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:13:37.121 23:47:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:13:37.121 23:47:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 4 -gt 2 ']' 00:13:37.121 23:47:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:13:37.121 23:47:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:13:37.121 23:47:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:13:37.121 23:47:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:37.121 23:47:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:37.121 BaseBdev2 00:13:37.121 23:47:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:37.121 23:47:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:13:37.121 23:47:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:13:37.121 23:47:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:13:37.121 23:47:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:13:37.121 23:47:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:13:37.122 23:47:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:13:37.122 23:47:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:13:37.122 23:47:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:37.122 23:47:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:37.122 23:47:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:37.122 23:47:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:13:37.122 23:47:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:37.122 23:47:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:37.122 [ 00:13:37.122 { 00:13:37.122 "name": "BaseBdev2", 00:13:37.122 "aliases": [ 00:13:37.122 "fdf0e830-e6b3-4a21-bf7a-f8a1de74ee57" 00:13:37.122 ], 00:13:37.122 "product_name": "Malloc disk", 00:13:37.122 "block_size": 512, 00:13:37.122 "num_blocks": 65536, 00:13:37.122 "uuid": "fdf0e830-e6b3-4a21-bf7a-f8a1de74ee57", 00:13:37.122 "assigned_rate_limits": { 00:13:37.122 "rw_ios_per_sec": 0, 00:13:37.122 "rw_mbytes_per_sec": 0, 00:13:37.122 "r_mbytes_per_sec": 0, 00:13:37.122 "w_mbytes_per_sec": 0 00:13:37.122 }, 00:13:37.122 "claimed": false, 00:13:37.122 "zoned": false, 00:13:37.122 "supported_io_types": { 00:13:37.122 "read": true, 00:13:37.122 "write": true, 00:13:37.122 "unmap": true, 00:13:37.122 "flush": true, 00:13:37.122 "reset": true, 00:13:37.122 "nvme_admin": false, 00:13:37.122 "nvme_io": false, 00:13:37.122 "nvme_io_md": false, 00:13:37.122 "write_zeroes": true, 00:13:37.122 "zcopy": true, 00:13:37.122 "get_zone_info": false, 00:13:37.122 "zone_management": false, 00:13:37.122 "zone_append": false, 00:13:37.122 "compare": false, 00:13:37.122 "compare_and_write": false, 00:13:37.122 "abort": true, 00:13:37.122 "seek_hole": false, 00:13:37.122 "seek_data": false, 00:13:37.122 "copy": true, 00:13:37.122 "nvme_iov_md": false 00:13:37.122 }, 00:13:37.122 "memory_domains": [ 00:13:37.122 { 00:13:37.122 "dma_device_id": "system", 00:13:37.122 "dma_device_type": 1 00:13:37.122 }, 00:13:37.122 { 00:13:37.122 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:37.122 "dma_device_type": 2 00:13:37.122 } 00:13:37.122 ], 00:13:37.122 "driver_specific": {} 00:13:37.122 } 00:13:37.122 ] 00:13:37.122 23:47:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:37.122 23:47:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:13:37.122 23:47:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:13:37.122 23:47:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:13:37.122 23:47:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:13:37.122 23:47:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:37.122 23:47:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:37.395 BaseBdev3 00:13:37.395 23:47:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:37.395 23:47:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:13:37.395 23:47:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:13:37.395 23:47:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:13:37.395 23:47:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:13:37.395 23:47:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:13:37.395 23:47:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:13:37.395 23:47:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:13:37.395 23:47:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:37.395 23:47:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:37.395 23:47:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:37.395 23:47:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:13:37.395 23:47:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:37.395 23:47:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:37.395 [ 00:13:37.395 { 00:13:37.395 "name": "BaseBdev3", 00:13:37.395 "aliases": [ 00:13:37.395 "e89fa23e-7fb8-4d81-9eed-6d5086f83907" 00:13:37.395 ], 00:13:37.395 "product_name": "Malloc disk", 00:13:37.395 "block_size": 512, 00:13:37.395 "num_blocks": 65536, 00:13:37.395 "uuid": "e89fa23e-7fb8-4d81-9eed-6d5086f83907", 00:13:37.395 "assigned_rate_limits": { 00:13:37.395 "rw_ios_per_sec": 0, 00:13:37.395 "rw_mbytes_per_sec": 0, 00:13:37.395 "r_mbytes_per_sec": 0, 00:13:37.395 "w_mbytes_per_sec": 0 00:13:37.395 }, 00:13:37.395 "claimed": false, 00:13:37.395 "zoned": false, 00:13:37.395 "supported_io_types": { 00:13:37.395 "read": true, 00:13:37.395 "write": true, 00:13:37.395 "unmap": true, 00:13:37.395 "flush": true, 00:13:37.395 "reset": true, 00:13:37.395 "nvme_admin": false, 00:13:37.395 "nvme_io": false, 00:13:37.395 "nvme_io_md": false, 00:13:37.395 "write_zeroes": true, 00:13:37.395 "zcopy": true, 00:13:37.395 "get_zone_info": false, 00:13:37.395 "zone_management": false, 00:13:37.395 "zone_append": false, 00:13:37.395 "compare": false, 00:13:37.395 "compare_and_write": false, 00:13:37.395 "abort": true, 00:13:37.395 "seek_hole": false, 00:13:37.395 "seek_data": false, 00:13:37.395 "copy": true, 00:13:37.395 "nvme_iov_md": false 00:13:37.395 }, 00:13:37.395 "memory_domains": [ 00:13:37.395 { 00:13:37.395 "dma_device_id": "system", 00:13:37.395 "dma_device_type": 1 00:13:37.395 }, 00:13:37.395 { 00:13:37.395 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:37.395 "dma_device_type": 2 00:13:37.395 } 00:13:37.395 ], 00:13:37.395 "driver_specific": {} 00:13:37.395 } 00:13:37.395 ] 00:13:37.395 23:47:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:37.395 23:47:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:13:37.395 23:47:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:13:37.395 23:47:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:13:37.395 23:47:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:13:37.395 23:47:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:37.395 23:47:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:37.395 BaseBdev4 00:13:37.395 23:47:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:37.395 23:47:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev4 00:13:37.395 23:47:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:13:37.395 23:47:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:13:37.395 23:47:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:13:37.395 23:47:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:13:37.395 23:47:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:13:37.395 23:47:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:13:37.395 23:47:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:37.395 23:47:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:37.395 23:47:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:37.395 23:47:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:13:37.395 23:47:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:37.395 23:47:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:37.395 [ 00:13:37.395 { 00:13:37.395 "name": "BaseBdev4", 00:13:37.395 "aliases": [ 00:13:37.395 "ef6ce586-35c4-4135-973b-d5b8a4083996" 00:13:37.395 ], 00:13:37.395 "product_name": "Malloc disk", 00:13:37.395 "block_size": 512, 00:13:37.396 "num_blocks": 65536, 00:13:37.396 "uuid": "ef6ce586-35c4-4135-973b-d5b8a4083996", 00:13:37.396 "assigned_rate_limits": { 00:13:37.396 "rw_ios_per_sec": 0, 00:13:37.396 "rw_mbytes_per_sec": 0, 00:13:37.396 "r_mbytes_per_sec": 0, 00:13:37.396 "w_mbytes_per_sec": 0 00:13:37.396 }, 00:13:37.396 "claimed": false, 00:13:37.396 "zoned": false, 00:13:37.396 "supported_io_types": { 00:13:37.396 "read": true, 00:13:37.396 "write": true, 00:13:37.396 "unmap": true, 00:13:37.396 "flush": true, 00:13:37.396 "reset": true, 00:13:37.396 "nvme_admin": false, 00:13:37.396 "nvme_io": false, 00:13:37.396 "nvme_io_md": false, 00:13:37.396 "write_zeroes": true, 00:13:37.396 "zcopy": true, 00:13:37.396 "get_zone_info": false, 00:13:37.396 "zone_management": false, 00:13:37.396 "zone_append": false, 00:13:37.396 "compare": false, 00:13:37.396 "compare_and_write": false, 00:13:37.396 "abort": true, 00:13:37.396 "seek_hole": false, 00:13:37.396 "seek_data": false, 00:13:37.396 "copy": true, 00:13:37.396 "nvme_iov_md": false 00:13:37.396 }, 00:13:37.396 "memory_domains": [ 00:13:37.396 { 00:13:37.396 "dma_device_id": "system", 00:13:37.396 "dma_device_type": 1 00:13:37.396 }, 00:13:37.396 { 00:13:37.396 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:37.396 "dma_device_type": 2 00:13:37.396 } 00:13:37.396 ], 00:13:37.396 "driver_specific": {} 00:13:37.396 } 00:13:37.396 ] 00:13:37.396 23:47:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:37.396 23:47:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:13:37.396 23:47:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:13:37.396 23:47:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:13:37.396 23:47:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:13:37.396 23:47:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:37.396 23:47:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:37.396 [2024-11-26 23:47:25.341478] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:13:37.396 [2024-11-26 23:47:25.341561] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:13:37.396 [2024-11-26 23:47:25.341622] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:13:37.396 [2024-11-26 23:47:25.343334] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:13:37.396 [2024-11-26 23:47:25.343428] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:13:37.396 23:47:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:37.396 23:47:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:13:37.396 23:47:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:37.396 23:47:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:37.396 23:47:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:37.396 23:47:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:37.396 23:47:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:13:37.396 23:47:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:37.396 23:47:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:37.396 23:47:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:37.396 23:47:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:37.396 23:47:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:37.396 23:47:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:37.396 23:47:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:37.396 23:47:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:37.396 23:47:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:37.396 23:47:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:37.396 "name": "Existed_Raid", 00:13:37.396 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:37.396 "strip_size_kb": 64, 00:13:37.396 "state": "configuring", 00:13:37.396 "raid_level": "raid5f", 00:13:37.396 "superblock": false, 00:13:37.396 "num_base_bdevs": 4, 00:13:37.396 "num_base_bdevs_discovered": 3, 00:13:37.396 "num_base_bdevs_operational": 4, 00:13:37.396 "base_bdevs_list": [ 00:13:37.396 { 00:13:37.396 "name": "BaseBdev1", 00:13:37.396 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:37.396 "is_configured": false, 00:13:37.396 "data_offset": 0, 00:13:37.396 "data_size": 0 00:13:37.396 }, 00:13:37.396 { 00:13:37.396 "name": "BaseBdev2", 00:13:37.396 "uuid": "fdf0e830-e6b3-4a21-bf7a-f8a1de74ee57", 00:13:37.396 "is_configured": true, 00:13:37.396 "data_offset": 0, 00:13:37.396 "data_size": 65536 00:13:37.396 }, 00:13:37.396 { 00:13:37.396 "name": "BaseBdev3", 00:13:37.396 "uuid": "e89fa23e-7fb8-4d81-9eed-6d5086f83907", 00:13:37.396 "is_configured": true, 00:13:37.396 "data_offset": 0, 00:13:37.396 "data_size": 65536 00:13:37.396 }, 00:13:37.396 { 00:13:37.396 "name": "BaseBdev4", 00:13:37.396 "uuid": "ef6ce586-35c4-4135-973b-d5b8a4083996", 00:13:37.396 "is_configured": true, 00:13:37.396 "data_offset": 0, 00:13:37.396 "data_size": 65536 00:13:37.396 } 00:13:37.396 ] 00:13:37.396 }' 00:13:37.396 23:47:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:37.396 23:47:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:37.656 23:47:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:13:37.656 23:47:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:37.656 23:47:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:37.656 [2024-11-26 23:47:25.752810] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:13:37.656 23:47:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:37.656 23:47:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:13:37.656 23:47:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:37.656 23:47:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:37.656 23:47:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:37.656 23:47:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:37.656 23:47:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:13:37.656 23:47:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:37.656 23:47:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:37.656 23:47:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:37.656 23:47:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:37.656 23:47:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:37.656 23:47:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:37.656 23:47:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:37.656 23:47:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:37.916 23:47:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:37.916 23:47:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:37.916 "name": "Existed_Raid", 00:13:37.916 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:37.916 "strip_size_kb": 64, 00:13:37.916 "state": "configuring", 00:13:37.916 "raid_level": "raid5f", 00:13:37.916 "superblock": false, 00:13:37.916 "num_base_bdevs": 4, 00:13:37.916 "num_base_bdevs_discovered": 2, 00:13:37.916 "num_base_bdevs_operational": 4, 00:13:37.916 "base_bdevs_list": [ 00:13:37.916 { 00:13:37.916 "name": "BaseBdev1", 00:13:37.916 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:37.916 "is_configured": false, 00:13:37.916 "data_offset": 0, 00:13:37.916 "data_size": 0 00:13:37.916 }, 00:13:37.916 { 00:13:37.916 "name": null, 00:13:37.916 "uuid": "fdf0e830-e6b3-4a21-bf7a-f8a1de74ee57", 00:13:37.916 "is_configured": false, 00:13:37.916 "data_offset": 0, 00:13:37.916 "data_size": 65536 00:13:37.916 }, 00:13:37.916 { 00:13:37.916 "name": "BaseBdev3", 00:13:37.916 "uuid": "e89fa23e-7fb8-4d81-9eed-6d5086f83907", 00:13:37.916 "is_configured": true, 00:13:37.916 "data_offset": 0, 00:13:37.916 "data_size": 65536 00:13:37.916 }, 00:13:37.916 { 00:13:37.916 "name": "BaseBdev4", 00:13:37.916 "uuid": "ef6ce586-35c4-4135-973b-d5b8a4083996", 00:13:37.916 "is_configured": true, 00:13:37.916 "data_offset": 0, 00:13:37.916 "data_size": 65536 00:13:37.916 } 00:13:37.916 ] 00:13:37.916 }' 00:13:37.916 23:47:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:37.916 23:47:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:38.177 23:47:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:38.177 23:47:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:38.177 23:47:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:13:38.177 23:47:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:38.177 23:47:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:38.177 23:47:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:13:38.177 23:47:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:13:38.177 23:47:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:38.177 23:47:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:38.177 [2024-11-26 23:47:26.174800] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:13:38.177 BaseBdev1 00:13:38.177 23:47:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:38.177 23:47:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:13:38.177 23:47:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:13:38.177 23:47:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:13:38.177 23:47:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:13:38.177 23:47:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:13:38.177 23:47:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:13:38.177 23:47:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:13:38.177 23:47:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:38.177 23:47:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:38.177 23:47:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:38.177 23:47:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:13:38.177 23:47:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:38.177 23:47:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:38.177 [ 00:13:38.177 { 00:13:38.177 "name": "BaseBdev1", 00:13:38.177 "aliases": [ 00:13:38.177 "809ff151-2b02-4576-a310-e4d655fababd" 00:13:38.177 ], 00:13:38.177 "product_name": "Malloc disk", 00:13:38.177 "block_size": 512, 00:13:38.177 "num_blocks": 65536, 00:13:38.177 "uuid": "809ff151-2b02-4576-a310-e4d655fababd", 00:13:38.177 "assigned_rate_limits": { 00:13:38.177 "rw_ios_per_sec": 0, 00:13:38.177 "rw_mbytes_per_sec": 0, 00:13:38.177 "r_mbytes_per_sec": 0, 00:13:38.177 "w_mbytes_per_sec": 0 00:13:38.177 }, 00:13:38.177 "claimed": true, 00:13:38.177 "claim_type": "exclusive_write", 00:13:38.177 "zoned": false, 00:13:38.177 "supported_io_types": { 00:13:38.177 "read": true, 00:13:38.177 "write": true, 00:13:38.177 "unmap": true, 00:13:38.177 "flush": true, 00:13:38.177 "reset": true, 00:13:38.177 "nvme_admin": false, 00:13:38.177 "nvme_io": false, 00:13:38.177 "nvme_io_md": false, 00:13:38.177 "write_zeroes": true, 00:13:38.177 "zcopy": true, 00:13:38.177 "get_zone_info": false, 00:13:38.177 "zone_management": false, 00:13:38.177 "zone_append": false, 00:13:38.177 "compare": false, 00:13:38.177 "compare_and_write": false, 00:13:38.177 "abort": true, 00:13:38.177 "seek_hole": false, 00:13:38.177 "seek_data": false, 00:13:38.177 "copy": true, 00:13:38.177 "nvme_iov_md": false 00:13:38.177 }, 00:13:38.177 "memory_domains": [ 00:13:38.177 { 00:13:38.177 "dma_device_id": "system", 00:13:38.177 "dma_device_type": 1 00:13:38.177 }, 00:13:38.177 { 00:13:38.177 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:38.177 "dma_device_type": 2 00:13:38.177 } 00:13:38.177 ], 00:13:38.177 "driver_specific": {} 00:13:38.177 } 00:13:38.177 ] 00:13:38.177 23:47:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:38.177 23:47:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:13:38.177 23:47:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:13:38.177 23:47:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:38.177 23:47:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:38.177 23:47:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:38.177 23:47:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:38.177 23:47:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:13:38.177 23:47:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:38.177 23:47:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:38.177 23:47:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:38.177 23:47:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:38.177 23:47:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:38.177 23:47:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:38.177 23:47:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:38.177 23:47:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:38.177 23:47:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:38.177 23:47:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:38.177 "name": "Existed_Raid", 00:13:38.178 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:38.178 "strip_size_kb": 64, 00:13:38.178 "state": "configuring", 00:13:38.178 "raid_level": "raid5f", 00:13:38.178 "superblock": false, 00:13:38.178 "num_base_bdevs": 4, 00:13:38.178 "num_base_bdevs_discovered": 3, 00:13:38.178 "num_base_bdevs_operational": 4, 00:13:38.178 "base_bdevs_list": [ 00:13:38.178 { 00:13:38.178 "name": "BaseBdev1", 00:13:38.178 "uuid": "809ff151-2b02-4576-a310-e4d655fababd", 00:13:38.178 "is_configured": true, 00:13:38.178 "data_offset": 0, 00:13:38.178 "data_size": 65536 00:13:38.178 }, 00:13:38.178 { 00:13:38.178 "name": null, 00:13:38.178 "uuid": "fdf0e830-e6b3-4a21-bf7a-f8a1de74ee57", 00:13:38.178 "is_configured": false, 00:13:38.178 "data_offset": 0, 00:13:38.178 "data_size": 65536 00:13:38.178 }, 00:13:38.178 { 00:13:38.178 "name": "BaseBdev3", 00:13:38.178 "uuid": "e89fa23e-7fb8-4d81-9eed-6d5086f83907", 00:13:38.178 "is_configured": true, 00:13:38.178 "data_offset": 0, 00:13:38.178 "data_size": 65536 00:13:38.178 }, 00:13:38.178 { 00:13:38.178 "name": "BaseBdev4", 00:13:38.178 "uuid": "ef6ce586-35c4-4135-973b-d5b8a4083996", 00:13:38.178 "is_configured": true, 00:13:38.178 "data_offset": 0, 00:13:38.178 "data_size": 65536 00:13:38.178 } 00:13:38.178 ] 00:13:38.178 }' 00:13:38.178 23:47:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:38.178 23:47:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:38.746 23:47:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:13:38.746 23:47:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:38.746 23:47:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:38.746 23:47:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:38.746 23:47:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:38.746 23:47:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:13:38.746 23:47:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:13:38.746 23:47:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:38.746 23:47:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:38.746 [2024-11-26 23:47:26.666044] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:13:38.746 23:47:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:38.746 23:47:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:13:38.746 23:47:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:38.746 23:47:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:38.746 23:47:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:38.746 23:47:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:38.746 23:47:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:13:38.746 23:47:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:38.746 23:47:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:38.746 23:47:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:38.746 23:47:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:38.746 23:47:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:38.746 23:47:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:38.746 23:47:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:38.746 23:47:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:38.746 23:47:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:38.746 23:47:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:38.746 "name": "Existed_Raid", 00:13:38.746 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:38.746 "strip_size_kb": 64, 00:13:38.746 "state": "configuring", 00:13:38.746 "raid_level": "raid5f", 00:13:38.746 "superblock": false, 00:13:38.746 "num_base_bdevs": 4, 00:13:38.746 "num_base_bdevs_discovered": 2, 00:13:38.746 "num_base_bdevs_operational": 4, 00:13:38.746 "base_bdevs_list": [ 00:13:38.746 { 00:13:38.746 "name": "BaseBdev1", 00:13:38.746 "uuid": "809ff151-2b02-4576-a310-e4d655fababd", 00:13:38.746 "is_configured": true, 00:13:38.746 "data_offset": 0, 00:13:38.746 "data_size": 65536 00:13:38.746 }, 00:13:38.746 { 00:13:38.746 "name": null, 00:13:38.746 "uuid": "fdf0e830-e6b3-4a21-bf7a-f8a1de74ee57", 00:13:38.746 "is_configured": false, 00:13:38.746 "data_offset": 0, 00:13:38.746 "data_size": 65536 00:13:38.746 }, 00:13:38.746 { 00:13:38.746 "name": null, 00:13:38.746 "uuid": "e89fa23e-7fb8-4d81-9eed-6d5086f83907", 00:13:38.746 "is_configured": false, 00:13:38.746 "data_offset": 0, 00:13:38.746 "data_size": 65536 00:13:38.746 }, 00:13:38.746 { 00:13:38.746 "name": "BaseBdev4", 00:13:38.746 "uuid": "ef6ce586-35c4-4135-973b-d5b8a4083996", 00:13:38.746 "is_configured": true, 00:13:38.746 "data_offset": 0, 00:13:38.746 "data_size": 65536 00:13:38.746 } 00:13:38.746 ] 00:13:38.746 }' 00:13:38.746 23:47:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:38.746 23:47:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:39.005 23:47:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:39.005 23:47:27 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:39.005 23:47:27 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:39.005 23:47:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:13:39.265 23:47:27 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:39.265 23:47:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:13:39.265 23:47:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:13:39.265 23:47:27 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:39.265 23:47:27 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:39.265 [2024-11-26 23:47:27.177215] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:13:39.265 23:47:27 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:39.265 23:47:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:13:39.265 23:47:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:39.265 23:47:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:39.265 23:47:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:39.265 23:47:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:39.265 23:47:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:13:39.265 23:47:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:39.265 23:47:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:39.265 23:47:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:39.265 23:47:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:39.265 23:47:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:39.265 23:47:27 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:39.265 23:47:27 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:39.265 23:47:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:39.265 23:47:27 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:39.265 23:47:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:39.265 "name": "Existed_Raid", 00:13:39.265 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:39.265 "strip_size_kb": 64, 00:13:39.265 "state": "configuring", 00:13:39.265 "raid_level": "raid5f", 00:13:39.265 "superblock": false, 00:13:39.265 "num_base_bdevs": 4, 00:13:39.265 "num_base_bdevs_discovered": 3, 00:13:39.265 "num_base_bdevs_operational": 4, 00:13:39.265 "base_bdevs_list": [ 00:13:39.265 { 00:13:39.265 "name": "BaseBdev1", 00:13:39.265 "uuid": "809ff151-2b02-4576-a310-e4d655fababd", 00:13:39.265 "is_configured": true, 00:13:39.265 "data_offset": 0, 00:13:39.265 "data_size": 65536 00:13:39.265 }, 00:13:39.265 { 00:13:39.265 "name": null, 00:13:39.265 "uuid": "fdf0e830-e6b3-4a21-bf7a-f8a1de74ee57", 00:13:39.265 "is_configured": false, 00:13:39.265 "data_offset": 0, 00:13:39.265 "data_size": 65536 00:13:39.265 }, 00:13:39.265 { 00:13:39.265 "name": "BaseBdev3", 00:13:39.265 "uuid": "e89fa23e-7fb8-4d81-9eed-6d5086f83907", 00:13:39.265 "is_configured": true, 00:13:39.265 "data_offset": 0, 00:13:39.265 "data_size": 65536 00:13:39.265 }, 00:13:39.265 { 00:13:39.265 "name": "BaseBdev4", 00:13:39.265 "uuid": "ef6ce586-35c4-4135-973b-d5b8a4083996", 00:13:39.265 "is_configured": true, 00:13:39.265 "data_offset": 0, 00:13:39.265 "data_size": 65536 00:13:39.265 } 00:13:39.265 ] 00:13:39.265 }' 00:13:39.265 23:47:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:39.265 23:47:27 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:39.524 23:47:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:13:39.524 23:47:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:39.524 23:47:27 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:39.524 23:47:27 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:39.524 23:47:27 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:39.783 23:47:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:13:39.783 23:47:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:13:39.783 23:47:27 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:39.783 23:47:27 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:39.783 [2024-11-26 23:47:27.680463] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:13:39.783 23:47:27 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:39.783 23:47:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:13:39.783 23:47:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:39.783 23:47:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:39.783 23:47:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:39.783 23:47:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:39.783 23:47:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:13:39.783 23:47:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:39.783 23:47:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:39.783 23:47:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:39.783 23:47:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:39.783 23:47:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:39.783 23:47:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:39.783 23:47:27 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:39.783 23:47:27 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:39.783 23:47:27 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:39.783 23:47:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:39.783 "name": "Existed_Raid", 00:13:39.783 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:39.783 "strip_size_kb": 64, 00:13:39.783 "state": "configuring", 00:13:39.783 "raid_level": "raid5f", 00:13:39.783 "superblock": false, 00:13:39.783 "num_base_bdevs": 4, 00:13:39.783 "num_base_bdevs_discovered": 2, 00:13:39.783 "num_base_bdevs_operational": 4, 00:13:39.783 "base_bdevs_list": [ 00:13:39.783 { 00:13:39.783 "name": null, 00:13:39.783 "uuid": "809ff151-2b02-4576-a310-e4d655fababd", 00:13:39.783 "is_configured": false, 00:13:39.783 "data_offset": 0, 00:13:39.783 "data_size": 65536 00:13:39.783 }, 00:13:39.783 { 00:13:39.783 "name": null, 00:13:39.783 "uuid": "fdf0e830-e6b3-4a21-bf7a-f8a1de74ee57", 00:13:39.783 "is_configured": false, 00:13:39.783 "data_offset": 0, 00:13:39.783 "data_size": 65536 00:13:39.783 }, 00:13:39.783 { 00:13:39.783 "name": "BaseBdev3", 00:13:39.783 "uuid": "e89fa23e-7fb8-4d81-9eed-6d5086f83907", 00:13:39.783 "is_configured": true, 00:13:39.783 "data_offset": 0, 00:13:39.783 "data_size": 65536 00:13:39.783 }, 00:13:39.783 { 00:13:39.783 "name": "BaseBdev4", 00:13:39.783 "uuid": "ef6ce586-35c4-4135-973b-d5b8a4083996", 00:13:39.783 "is_configured": true, 00:13:39.783 "data_offset": 0, 00:13:39.783 "data_size": 65536 00:13:39.783 } 00:13:39.783 ] 00:13:39.783 }' 00:13:39.783 23:47:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:39.783 23:47:27 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:40.042 23:47:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:40.042 23:47:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:40.042 23:47:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:40.042 23:47:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:13:40.042 23:47:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:40.302 23:47:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:13:40.302 23:47:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:13:40.302 23:47:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:40.302 23:47:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:40.302 [2024-11-26 23:47:28.194232] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:13:40.302 23:47:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:40.302 23:47:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:13:40.302 23:47:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:40.302 23:47:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:40.302 23:47:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:40.302 23:47:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:40.302 23:47:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:13:40.302 23:47:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:40.302 23:47:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:40.302 23:47:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:40.302 23:47:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:40.302 23:47:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:40.302 23:47:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:40.302 23:47:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:40.302 23:47:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:40.302 23:47:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:40.302 23:47:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:40.302 "name": "Existed_Raid", 00:13:40.302 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:40.302 "strip_size_kb": 64, 00:13:40.302 "state": "configuring", 00:13:40.302 "raid_level": "raid5f", 00:13:40.302 "superblock": false, 00:13:40.302 "num_base_bdevs": 4, 00:13:40.302 "num_base_bdevs_discovered": 3, 00:13:40.302 "num_base_bdevs_operational": 4, 00:13:40.302 "base_bdevs_list": [ 00:13:40.302 { 00:13:40.302 "name": null, 00:13:40.302 "uuid": "809ff151-2b02-4576-a310-e4d655fababd", 00:13:40.302 "is_configured": false, 00:13:40.302 "data_offset": 0, 00:13:40.302 "data_size": 65536 00:13:40.302 }, 00:13:40.302 { 00:13:40.302 "name": "BaseBdev2", 00:13:40.302 "uuid": "fdf0e830-e6b3-4a21-bf7a-f8a1de74ee57", 00:13:40.302 "is_configured": true, 00:13:40.302 "data_offset": 0, 00:13:40.302 "data_size": 65536 00:13:40.302 }, 00:13:40.302 { 00:13:40.302 "name": "BaseBdev3", 00:13:40.302 "uuid": "e89fa23e-7fb8-4d81-9eed-6d5086f83907", 00:13:40.302 "is_configured": true, 00:13:40.302 "data_offset": 0, 00:13:40.302 "data_size": 65536 00:13:40.302 }, 00:13:40.302 { 00:13:40.302 "name": "BaseBdev4", 00:13:40.302 "uuid": "ef6ce586-35c4-4135-973b-d5b8a4083996", 00:13:40.302 "is_configured": true, 00:13:40.302 "data_offset": 0, 00:13:40.302 "data_size": 65536 00:13:40.302 } 00:13:40.302 ] 00:13:40.302 }' 00:13:40.302 23:47:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:40.302 23:47:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:40.561 23:47:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:13:40.561 23:47:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:40.561 23:47:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:40.561 23:47:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:40.561 23:47:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:40.561 23:47:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:13:40.821 23:47:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:40.821 23:47:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:40.821 23:47:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:40.821 23:47:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:13:40.821 23:47:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:40.821 23:47:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 809ff151-2b02-4576-a310-e4d655fababd 00:13:40.821 23:47:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:40.821 23:47:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:40.821 [2024-11-26 23:47:28.752028] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:13:40.821 [2024-11-26 23:47:28.752075] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:13:40.821 [2024-11-26 23:47:28.752082] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 196608, blocklen 512 00:13:40.821 [2024-11-26 23:47:28.752329] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002a10 00:13:40.821 [2024-11-26 23:47:28.752768] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:13:40.821 [2024-11-26 23:47:28.752782] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001c80 00:13:40.821 [2024-11-26 23:47:28.752946] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:40.821 NewBaseBdev 00:13:40.821 23:47:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:40.821 23:47:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:13:40.821 23:47:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:13:40.821 23:47:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:13:40.821 23:47:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:13:40.821 23:47:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:13:40.821 23:47:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:13:40.821 23:47:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:13:40.821 23:47:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:40.821 23:47:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:40.821 23:47:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:40.821 23:47:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:13:40.821 23:47:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:40.821 23:47:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:40.821 [ 00:13:40.821 { 00:13:40.821 "name": "NewBaseBdev", 00:13:40.821 "aliases": [ 00:13:40.821 "809ff151-2b02-4576-a310-e4d655fababd" 00:13:40.821 ], 00:13:40.821 "product_name": "Malloc disk", 00:13:40.821 "block_size": 512, 00:13:40.821 "num_blocks": 65536, 00:13:40.821 "uuid": "809ff151-2b02-4576-a310-e4d655fababd", 00:13:40.821 "assigned_rate_limits": { 00:13:40.821 "rw_ios_per_sec": 0, 00:13:40.821 "rw_mbytes_per_sec": 0, 00:13:40.821 "r_mbytes_per_sec": 0, 00:13:40.821 "w_mbytes_per_sec": 0 00:13:40.821 }, 00:13:40.821 "claimed": true, 00:13:40.821 "claim_type": "exclusive_write", 00:13:40.821 "zoned": false, 00:13:40.821 "supported_io_types": { 00:13:40.821 "read": true, 00:13:40.821 "write": true, 00:13:40.821 "unmap": true, 00:13:40.821 "flush": true, 00:13:40.821 "reset": true, 00:13:40.821 "nvme_admin": false, 00:13:40.821 "nvme_io": false, 00:13:40.821 "nvme_io_md": false, 00:13:40.821 "write_zeroes": true, 00:13:40.821 "zcopy": true, 00:13:40.821 "get_zone_info": false, 00:13:40.821 "zone_management": false, 00:13:40.821 "zone_append": false, 00:13:40.821 "compare": false, 00:13:40.821 "compare_and_write": false, 00:13:40.821 "abort": true, 00:13:40.821 "seek_hole": false, 00:13:40.821 "seek_data": false, 00:13:40.821 "copy": true, 00:13:40.821 "nvme_iov_md": false 00:13:40.821 }, 00:13:40.821 "memory_domains": [ 00:13:40.821 { 00:13:40.821 "dma_device_id": "system", 00:13:40.821 "dma_device_type": 1 00:13:40.821 }, 00:13:40.821 { 00:13:40.821 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:40.821 "dma_device_type": 2 00:13:40.821 } 00:13:40.821 ], 00:13:40.821 "driver_specific": {} 00:13:40.821 } 00:13:40.821 ] 00:13:40.821 23:47:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:40.821 23:47:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:13:40.821 23:47:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 4 00:13:40.821 23:47:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:40.821 23:47:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:40.821 23:47:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:40.821 23:47:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:40.821 23:47:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:13:40.821 23:47:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:40.821 23:47:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:40.821 23:47:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:40.821 23:47:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:40.821 23:47:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:40.821 23:47:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:40.821 23:47:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:40.821 23:47:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:40.821 23:47:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:40.821 23:47:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:40.821 "name": "Existed_Raid", 00:13:40.821 "uuid": "d1915686-b250-4a21-8825-2f851c12d703", 00:13:40.821 "strip_size_kb": 64, 00:13:40.821 "state": "online", 00:13:40.821 "raid_level": "raid5f", 00:13:40.821 "superblock": false, 00:13:40.821 "num_base_bdevs": 4, 00:13:40.821 "num_base_bdevs_discovered": 4, 00:13:40.821 "num_base_bdevs_operational": 4, 00:13:40.821 "base_bdevs_list": [ 00:13:40.821 { 00:13:40.821 "name": "NewBaseBdev", 00:13:40.821 "uuid": "809ff151-2b02-4576-a310-e4d655fababd", 00:13:40.821 "is_configured": true, 00:13:40.822 "data_offset": 0, 00:13:40.822 "data_size": 65536 00:13:40.822 }, 00:13:40.822 { 00:13:40.822 "name": "BaseBdev2", 00:13:40.822 "uuid": "fdf0e830-e6b3-4a21-bf7a-f8a1de74ee57", 00:13:40.822 "is_configured": true, 00:13:40.822 "data_offset": 0, 00:13:40.822 "data_size": 65536 00:13:40.822 }, 00:13:40.822 { 00:13:40.822 "name": "BaseBdev3", 00:13:40.822 "uuid": "e89fa23e-7fb8-4d81-9eed-6d5086f83907", 00:13:40.822 "is_configured": true, 00:13:40.822 "data_offset": 0, 00:13:40.822 "data_size": 65536 00:13:40.822 }, 00:13:40.822 { 00:13:40.822 "name": "BaseBdev4", 00:13:40.822 "uuid": "ef6ce586-35c4-4135-973b-d5b8a4083996", 00:13:40.822 "is_configured": true, 00:13:40.822 "data_offset": 0, 00:13:40.822 "data_size": 65536 00:13:40.822 } 00:13:40.822 ] 00:13:40.822 }' 00:13:40.822 23:47:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:40.822 23:47:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:41.081 23:47:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:13:41.081 23:47:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:13:41.081 23:47:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:13:41.081 23:47:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:13:41.081 23:47:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:13:41.081 23:47:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:13:41.081 23:47:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:13:41.081 23:47:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:13:41.081 23:47:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:41.081 23:47:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:41.081 [2024-11-26 23:47:29.179541] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:13:41.081 23:47:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:41.339 23:47:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:13:41.339 "name": "Existed_Raid", 00:13:41.339 "aliases": [ 00:13:41.339 "d1915686-b250-4a21-8825-2f851c12d703" 00:13:41.339 ], 00:13:41.339 "product_name": "Raid Volume", 00:13:41.340 "block_size": 512, 00:13:41.340 "num_blocks": 196608, 00:13:41.340 "uuid": "d1915686-b250-4a21-8825-2f851c12d703", 00:13:41.340 "assigned_rate_limits": { 00:13:41.340 "rw_ios_per_sec": 0, 00:13:41.340 "rw_mbytes_per_sec": 0, 00:13:41.340 "r_mbytes_per_sec": 0, 00:13:41.340 "w_mbytes_per_sec": 0 00:13:41.340 }, 00:13:41.340 "claimed": false, 00:13:41.340 "zoned": false, 00:13:41.340 "supported_io_types": { 00:13:41.340 "read": true, 00:13:41.340 "write": true, 00:13:41.340 "unmap": false, 00:13:41.340 "flush": false, 00:13:41.340 "reset": true, 00:13:41.340 "nvme_admin": false, 00:13:41.340 "nvme_io": false, 00:13:41.340 "nvme_io_md": false, 00:13:41.340 "write_zeroes": true, 00:13:41.340 "zcopy": false, 00:13:41.340 "get_zone_info": false, 00:13:41.340 "zone_management": false, 00:13:41.340 "zone_append": false, 00:13:41.340 "compare": false, 00:13:41.340 "compare_and_write": false, 00:13:41.340 "abort": false, 00:13:41.340 "seek_hole": false, 00:13:41.340 "seek_data": false, 00:13:41.340 "copy": false, 00:13:41.340 "nvme_iov_md": false 00:13:41.340 }, 00:13:41.340 "driver_specific": { 00:13:41.340 "raid": { 00:13:41.340 "uuid": "d1915686-b250-4a21-8825-2f851c12d703", 00:13:41.340 "strip_size_kb": 64, 00:13:41.340 "state": "online", 00:13:41.340 "raid_level": "raid5f", 00:13:41.340 "superblock": false, 00:13:41.340 "num_base_bdevs": 4, 00:13:41.340 "num_base_bdevs_discovered": 4, 00:13:41.340 "num_base_bdevs_operational": 4, 00:13:41.340 "base_bdevs_list": [ 00:13:41.340 { 00:13:41.340 "name": "NewBaseBdev", 00:13:41.340 "uuid": "809ff151-2b02-4576-a310-e4d655fababd", 00:13:41.340 "is_configured": true, 00:13:41.340 "data_offset": 0, 00:13:41.340 "data_size": 65536 00:13:41.340 }, 00:13:41.340 { 00:13:41.340 "name": "BaseBdev2", 00:13:41.340 "uuid": "fdf0e830-e6b3-4a21-bf7a-f8a1de74ee57", 00:13:41.340 "is_configured": true, 00:13:41.340 "data_offset": 0, 00:13:41.340 "data_size": 65536 00:13:41.340 }, 00:13:41.340 { 00:13:41.340 "name": "BaseBdev3", 00:13:41.340 "uuid": "e89fa23e-7fb8-4d81-9eed-6d5086f83907", 00:13:41.340 "is_configured": true, 00:13:41.340 "data_offset": 0, 00:13:41.340 "data_size": 65536 00:13:41.340 }, 00:13:41.340 { 00:13:41.340 "name": "BaseBdev4", 00:13:41.340 "uuid": "ef6ce586-35c4-4135-973b-d5b8a4083996", 00:13:41.340 "is_configured": true, 00:13:41.340 "data_offset": 0, 00:13:41.340 "data_size": 65536 00:13:41.340 } 00:13:41.340 ] 00:13:41.340 } 00:13:41.340 } 00:13:41.340 }' 00:13:41.340 23:47:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:13:41.340 23:47:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:13:41.340 BaseBdev2 00:13:41.340 BaseBdev3 00:13:41.340 BaseBdev4' 00:13:41.340 23:47:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:41.340 23:47:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:13:41.340 23:47:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:41.340 23:47:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:13:41.340 23:47:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:41.340 23:47:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:41.340 23:47:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:41.340 23:47:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:41.340 23:47:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:41.340 23:47:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:41.340 23:47:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:41.340 23:47:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:13:41.340 23:47:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:41.340 23:47:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:41.340 23:47:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:41.340 23:47:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:41.340 23:47:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:41.340 23:47:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:41.340 23:47:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:41.340 23:47:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:13:41.340 23:47:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:41.340 23:47:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:41.340 23:47:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:41.340 23:47:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:41.340 23:47:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:41.340 23:47:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:41.340 23:47:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:41.340 23:47:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:41.340 23:47:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:13:41.340 23:47:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:41.340 23:47:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:41.599 23:47:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:41.599 23:47:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:41.599 23:47:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:41.599 23:47:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:13:41.599 23:47:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:41.599 23:47:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:41.599 [2024-11-26 23:47:29.498777] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:13:41.599 [2024-11-26 23:47:29.498844] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:13:41.599 [2024-11-26 23:47:29.498917] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:13:41.599 [2024-11-26 23:47:29.499166] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:13:41.599 [2024-11-26 23:47:29.499183] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name Existed_Raid, state offline 00:13:41.599 23:47:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:41.599 23:47:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 92920 00:13:41.599 23:47:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@954 -- # '[' -z 92920 ']' 00:13:41.599 23:47:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@958 -- # kill -0 92920 00:13:41.599 23:47:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@959 -- # uname 00:13:41.599 23:47:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:13:41.599 23:47:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 92920 00:13:41.599 23:47:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:13:41.599 23:47:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:13:41.600 23:47:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 92920' 00:13:41.600 killing process with pid 92920 00:13:41.600 23:47:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@973 -- # kill 92920 00:13:41.600 [2024-11-26 23:47:29.543788] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:13:41.600 23:47:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@978 -- # wait 92920 00:13:41.600 [2024-11-26 23:47:29.583078] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:13:41.859 23:47:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:13:41.859 00:13:41.859 real 0m9.183s 00:13:41.859 user 0m15.708s 00:13:41.859 sys 0m1.978s 00:13:41.859 23:47:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:13:41.859 23:47:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:41.859 ************************************ 00:13:41.859 END TEST raid5f_state_function_test 00:13:41.859 ************************************ 00:13:41.859 23:47:29 bdev_raid -- bdev/bdev_raid.sh@987 -- # run_test raid5f_state_function_test_sb raid_state_function_test raid5f 4 true 00:13:41.859 23:47:29 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:13:41.859 23:47:29 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:13:41.859 23:47:29 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:13:41.859 ************************************ 00:13:41.859 START TEST raid5f_state_function_test_sb 00:13:41.859 ************************************ 00:13:41.859 23:47:29 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@1129 -- # raid_state_function_test raid5f 4 true 00:13:41.859 23:47:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=raid5f 00:13:41.859 23:47:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=4 00:13:41.859 23:47:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:13:41.859 23:47:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:13:41.859 23:47:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:13:41.859 23:47:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:13:41.859 23:47:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:13:41.859 23:47:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:13:41.859 23:47:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:13:41.859 23:47:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:13:41.859 23:47:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:13:41.859 23:47:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:13:41.859 23:47:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:13:41.859 23:47:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:13:41.859 23:47:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:13:41.859 23:47:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev4 00:13:41.859 23:47:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:13:41.859 23:47:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:13:41.859 23:47:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:13:41.859 23:47:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:13:41.859 23:47:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:13:41.859 23:47:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:13:41.859 23:47:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:13:41.859 23:47:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:13:41.859 23:47:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' raid5f '!=' raid1 ']' 00:13:41.859 23:47:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:13:41.859 23:47:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:13:41.859 23:47:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:13:41.859 23:47:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:13:41.859 23:47:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=93564 00:13:41.859 23:47:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:13:41.859 23:47:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 93564' 00:13:41.859 Process raid pid: 93564 00:13:41.859 23:47:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 93564 00:13:41.859 23:47:29 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@835 -- # '[' -z 93564 ']' 00:13:41.859 23:47:29 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:13:41.859 23:47:29 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@840 -- # local max_retries=100 00:13:41.859 23:47:29 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:13:41.859 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:13:41.859 23:47:29 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@844 -- # xtrace_disable 00:13:41.859 23:47:29 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:41.860 [2024-11-26 23:47:29.961871] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:13:41.860 [2024-11-26 23:47:29.962000] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:13:42.119 [2024-11-26 23:47:30.117834] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:13:42.119 [2024-11-26 23:47:30.142783] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:13:42.119 [2024-11-26 23:47:30.183442] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:13:42.119 [2024-11-26 23:47:30.183546] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:13:42.687 23:47:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:13:42.687 23:47:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@868 -- # return 0 00:13:42.687 23:47:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:13:42.687 23:47:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:42.687 23:47:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:42.688 [2024-11-26 23:47:30.777014] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:13:42.688 [2024-11-26 23:47:30.777066] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:13:42.688 [2024-11-26 23:47:30.777076] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:13:42.688 [2024-11-26 23:47:30.777085] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:13:42.688 [2024-11-26 23:47:30.777091] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:13:42.688 [2024-11-26 23:47:30.777102] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:13:42.688 [2024-11-26 23:47:30.777107] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:13:42.688 [2024-11-26 23:47:30.777115] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:13:42.688 23:47:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:42.688 23:47:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:13:42.688 23:47:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:42.688 23:47:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:42.688 23:47:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:42.688 23:47:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:42.688 23:47:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:13:42.688 23:47:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:42.688 23:47:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:42.688 23:47:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:42.688 23:47:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:42.688 23:47:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:42.688 23:47:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:42.688 23:47:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:42.688 23:47:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:42.688 23:47:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:42.947 23:47:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:42.947 "name": "Existed_Raid", 00:13:42.947 "uuid": "481f117c-76bc-4967-8a49-f23666fa1bc6", 00:13:42.947 "strip_size_kb": 64, 00:13:42.947 "state": "configuring", 00:13:42.947 "raid_level": "raid5f", 00:13:42.947 "superblock": true, 00:13:42.947 "num_base_bdevs": 4, 00:13:42.947 "num_base_bdevs_discovered": 0, 00:13:42.947 "num_base_bdevs_operational": 4, 00:13:42.947 "base_bdevs_list": [ 00:13:42.947 { 00:13:42.947 "name": "BaseBdev1", 00:13:42.947 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:42.947 "is_configured": false, 00:13:42.947 "data_offset": 0, 00:13:42.947 "data_size": 0 00:13:42.947 }, 00:13:42.948 { 00:13:42.948 "name": "BaseBdev2", 00:13:42.948 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:42.948 "is_configured": false, 00:13:42.948 "data_offset": 0, 00:13:42.948 "data_size": 0 00:13:42.948 }, 00:13:42.948 { 00:13:42.948 "name": "BaseBdev3", 00:13:42.948 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:42.948 "is_configured": false, 00:13:42.948 "data_offset": 0, 00:13:42.948 "data_size": 0 00:13:42.948 }, 00:13:42.948 { 00:13:42.948 "name": "BaseBdev4", 00:13:42.948 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:42.948 "is_configured": false, 00:13:42.948 "data_offset": 0, 00:13:42.948 "data_size": 0 00:13:42.948 } 00:13:42.948 ] 00:13:42.948 }' 00:13:42.948 23:47:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:42.948 23:47:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:43.208 23:47:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:13:43.208 23:47:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:43.208 23:47:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:43.208 [2024-11-26 23:47:31.220161] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:13:43.208 [2024-11-26 23:47:31.220252] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:13:43.208 23:47:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:43.208 23:47:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:13:43.208 23:47:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:43.208 23:47:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:43.208 [2024-11-26 23:47:31.232165] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:13:43.208 [2024-11-26 23:47:31.232245] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:13:43.208 [2024-11-26 23:47:31.232272] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:13:43.208 [2024-11-26 23:47:31.232293] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:13:43.208 [2024-11-26 23:47:31.232310] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:13:43.208 [2024-11-26 23:47:31.232330] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:13:43.208 [2024-11-26 23:47:31.232359] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:13:43.208 [2024-11-26 23:47:31.232380] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:13:43.208 23:47:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:43.208 23:47:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:13:43.208 23:47:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:43.208 23:47:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:43.208 [2024-11-26 23:47:31.252671] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:13:43.208 BaseBdev1 00:13:43.208 23:47:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:43.208 23:47:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:13:43.208 23:47:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:13:43.208 23:47:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:13:43.208 23:47:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:13:43.208 23:47:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:13:43.208 23:47:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:13:43.208 23:47:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:13:43.208 23:47:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:43.208 23:47:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:43.208 23:47:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:43.208 23:47:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:13:43.208 23:47:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:43.208 23:47:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:43.208 [ 00:13:43.208 { 00:13:43.208 "name": "BaseBdev1", 00:13:43.208 "aliases": [ 00:13:43.208 "9a479c4a-4e1f-4b18-89ab-1c27099eebe8" 00:13:43.208 ], 00:13:43.208 "product_name": "Malloc disk", 00:13:43.208 "block_size": 512, 00:13:43.208 "num_blocks": 65536, 00:13:43.208 "uuid": "9a479c4a-4e1f-4b18-89ab-1c27099eebe8", 00:13:43.208 "assigned_rate_limits": { 00:13:43.208 "rw_ios_per_sec": 0, 00:13:43.208 "rw_mbytes_per_sec": 0, 00:13:43.208 "r_mbytes_per_sec": 0, 00:13:43.208 "w_mbytes_per_sec": 0 00:13:43.208 }, 00:13:43.208 "claimed": true, 00:13:43.208 "claim_type": "exclusive_write", 00:13:43.208 "zoned": false, 00:13:43.208 "supported_io_types": { 00:13:43.208 "read": true, 00:13:43.208 "write": true, 00:13:43.208 "unmap": true, 00:13:43.208 "flush": true, 00:13:43.208 "reset": true, 00:13:43.208 "nvme_admin": false, 00:13:43.208 "nvme_io": false, 00:13:43.208 "nvme_io_md": false, 00:13:43.208 "write_zeroes": true, 00:13:43.208 "zcopy": true, 00:13:43.208 "get_zone_info": false, 00:13:43.208 "zone_management": false, 00:13:43.208 "zone_append": false, 00:13:43.208 "compare": false, 00:13:43.208 "compare_and_write": false, 00:13:43.208 "abort": true, 00:13:43.208 "seek_hole": false, 00:13:43.208 "seek_data": false, 00:13:43.208 "copy": true, 00:13:43.208 "nvme_iov_md": false 00:13:43.208 }, 00:13:43.209 "memory_domains": [ 00:13:43.209 { 00:13:43.209 "dma_device_id": "system", 00:13:43.209 "dma_device_type": 1 00:13:43.209 }, 00:13:43.209 { 00:13:43.209 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:43.209 "dma_device_type": 2 00:13:43.209 } 00:13:43.209 ], 00:13:43.209 "driver_specific": {} 00:13:43.209 } 00:13:43.209 ] 00:13:43.209 23:47:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:43.209 23:47:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:13:43.209 23:47:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:13:43.209 23:47:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:43.209 23:47:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:43.209 23:47:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:43.209 23:47:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:43.209 23:47:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:13:43.209 23:47:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:43.209 23:47:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:43.209 23:47:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:43.209 23:47:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:43.209 23:47:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:43.209 23:47:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:43.209 23:47:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:43.209 23:47:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:43.209 23:47:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:43.469 23:47:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:43.469 "name": "Existed_Raid", 00:13:43.469 "uuid": "3ba9932d-954a-40b2-a30b-60a22df8ef70", 00:13:43.469 "strip_size_kb": 64, 00:13:43.469 "state": "configuring", 00:13:43.469 "raid_level": "raid5f", 00:13:43.469 "superblock": true, 00:13:43.469 "num_base_bdevs": 4, 00:13:43.469 "num_base_bdevs_discovered": 1, 00:13:43.469 "num_base_bdevs_operational": 4, 00:13:43.469 "base_bdevs_list": [ 00:13:43.469 { 00:13:43.469 "name": "BaseBdev1", 00:13:43.469 "uuid": "9a479c4a-4e1f-4b18-89ab-1c27099eebe8", 00:13:43.469 "is_configured": true, 00:13:43.469 "data_offset": 2048, 00:13:43.469 "data_size": 63488 00:13:43.469 }, 00:13:43.469 { 00:13:43.469 "name": "BaseBdev2", 00:13:43.469 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:43.469 "is_configured": false, 00:13:43.469 "data_offset": 0, 00:13:43.469 "data_size": 0 00:13:43.469 }, 00:13:43.469 { 00:13:43.469 "name": "BaseBdev3", 00:13:43.469 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:43.469 "is_configured": false, 00:13:43.469 "data_offset": 0, 00:13:43.469 "data_size": 0 00:13:43.469 }, 00:13:43.469 { 00:13:43.469 "name": "BaseBdev4", 00:13:43.469 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:43.469 "is_configured": false, 00:13:43.469 "data_offset": 0, 00:13:43.469 "data_size": 0 00:13:43.469 } 00:13:43.469 ] 00:13:43.469 }' 00:13:43.469 23:47:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:43.469 23:47:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:43.728 23:47:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:13:43.728 23:47:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:43.728 23:47:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:43.728 [2024-11-26 23:47:31.707907] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:13:43.728 [2024-11-26 23:47:31.707946] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:13:43.728 23:47:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:43.728 23:47:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:13:43.728 23:47:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:43.728 23:47:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:43.728 [2024-11-26 23:47:31.719929] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:13:43.728 [2024-11-26 23:47:31.721757] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:13:43.728 [2024-11-26 23:47:31.721796] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:13:43.728 [2024-11-26 23:47:31.721805] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:13:43.728 [2024-11-26 23:47:31.721813] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:13:43.728 [2024-11-26 23:47:31.721819] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:13:43.728 [2024-11-26 23:47:31.721827] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:13:43.728 23:47:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:43.728 23:47:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:13:43.728 23:47:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:13:43.728 23:47:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:13:43.728 23:47:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:43.728 23:47:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:43.728 23:47:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:43.728 23:47:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:43.728 23:47:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:13:43.728 23:47:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:43.728 23:47:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:43.728 23:47:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:43.728 23:47:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:43.728 23:47:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:43.728 23:47:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:43.728 23:47:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:43.728 23:47:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:43.728 23:47:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:43.728 23:47:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:43.728 "name": "Existed_Raid", 00:13:43.728 "uuid": "90d55369-d0bc-45ea-afd7-c3c4867c9766", 00:13:43.728 "strip_size_kb": 64, 00:13:43.728 "state": "configuring", 00:13:43.728 "raid_level": "raid5f", 00:13:43.728 "superblock": true, 00:13:43.728 "num_base_bdevs": 4, 00:13:43.728 "num_base_bdevs_discovered": 1, 00:13:43.728 "num_base_bdevs_operational": 4, 00:13:43.728 "base_bdevs_list": [ 00:13:43.728 { 00:13:43.728 "name": "BaseBdev1", 00:13:43.728 "uuid": "9a479c4a-4e1f-4b18-89ab-1c27099eebe8", 00:13:43.728 "is_configured": true, 00:13:43.728 "data_offset": 2048, 00:13:43.728 "data_size": 63488 00:13:43.728 }, 00:13:43.728 { 00:13:43.728 "name": "BaseBdev2", 00:13:43.728 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:43.728 "is_configured": false, 00:13:43.728 "data_offset": 0, 00:13:43.728 "data_size": 0 00:13:43.728 }, 00:13:43.728 { 00:13:43.728 "name": "BaseBdev3", 00:13:43.728 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:43.728 "is_configured": false, 00:13:43.728 "data_offset": 0, 00:13:43.728 "data_size": 0 00:13:43.728 }, 00:13:43.728 { 00:13:43.728 "name": "BaseBdev4", 00:13:43.728 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:43.728 "is_configured": false, 00:13:43.728 "data_offset": 0, 00:13:43.728 "data_size": 0 00:13:43.728 } 00:13:43.728 ] 00:13:43.728 }' 00:13:43.728 23:47:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:43.728 23:47:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:44.296 23:47:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:13:44.296 23:47:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:44.296 23:47:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:44.296 [2024-11-26 23:47:32.133908] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:13:44.296 BaseBdev2 00:13:44.296 23:47:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:44.296 23:47:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:13:44.296 23:47:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:13:44.296 23:47:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:13:44.296 23:47:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:13:44.296 23:47:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:13:44.296 23:47:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:13:44.296 23:47:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:13:44.296 23:47:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:44.296 23:47:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:44.296 23:47:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:44.296 23:47:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:13:44.296 23:47:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:44.296 23:47:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:44.296 [ 00:13:44.296 { 00:13:44.296 "name": "BaseBdev2", 00:13:44.296 "aliases": [ 00:13:44.296 "e9070e18-a25e-4409-8311-9570e61d7262" 00:13:44.296 ], 00:13:44.296 "product_name": "Malloc disk", 00:13:44.296 "block_size": 512, 00:13:44.296 "num_blocks": 65536, 00:13:44.296 "uuid": "e9070e18-a25e-4409-8311-9570e61d7262", 00:13:44.296 "assigned_rate_limits": { 00:13:44.296 "rw_ios_per_sec": 0, 00:13:44.296 "rw_mbytes_per_sec": 0, 00:13:44.296 "r_mbytes_per_sec": 0, 00:13:44.296 "w_mbytes_per_sec": 0 00:13:44.296 }, 00:13:44.296 "claimed": true, 00:13:44.296 "claim_type": "exclusive_write", 00:13:44.296 "zoned": false, 00:13:44.296 "supported_io_types": { 00:13:44.296 "read": true, 00:13:44.296 "write": true, 00:13:44.296 "unmap": true, 00:13:44.296 "flush": true, 00:13:44.296 "reset": true, 00:13:44.296 "nvme_admin": false, 00:13:44.296 "nvme_io": false, 00:13:44.296 "nvme_io_md": false, 00:13:44.296 "write_zeroes": true, 00:13:44.296 "zcopy": true, 00:13:44.296 "get_zone_info": false, 00:13:44.296 "zone_management": false, 00:13:44.296 "zone_append": false, 00:13:44.296 "compare": false, 00:13:44.296 "compare_and_write": false, 00:13:44.296 "abort": true, 00:13:44.296 "seek_hole": false, 00:13:44.296 "seek_data": false, 00:13:44.296 "copy": true, 00:13:44.296 "nvme_iov_md": false 00:13:44.296 }, 00:13:44.296 "memory_domains": [ 00:13:44.296 { 00:13:44.296 "dma_device_id": "system", 00:13:44.296 "dma_device_type": 1 00:13:44.296 }, 00:13:44.296 { 00:13:44.296 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:44.296 "dma_device_type": 2 00:13:44.296 } 00:13:44.296 ], 00:13:44.296 "driver_specific": {} 00:13:44.296 } 00:13:44.296 ] 00:13:44.296 23:47:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:44.296 23:47:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:13:44.296 23:47:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:13:44.296 23:47:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:13:44.296 23:47:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:13:44.296 23:47:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:44.297 23:47:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:44.297 23:47:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:44.297 23:47:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:44.297 23:47:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:13:44.297 23:47:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:44.297 23:47:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:44.297 23:47:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:44.297 23:47:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:44.297 23:47:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:44.297 23:47:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:44.297 23:47:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:44.297 23:47:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:44.297 23:47:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:44.297 23:47:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:44.297 "name": "Existed_Raid", 00:13:44.297 "uuid": "90d55369-d0bc-45ea-afd7-c3c4867c9766", 00:13:44.297 "strip_size_kb": 64, 00:13:44.297 "state": "configuring", 00:13:44.297 "raid_level": "raid5f", 00:13:44.297 "superblock": true, 00:13:44.297 "num_base_bdevs": 4, 00:13:44.297 "num_base_bdevs_discovered": 2, 00:13:44.297 "num_base_bdevs_operational": 4, 00:13:44.297 "base_bdevs_list": [ 00:13:44.297 { 00:13:44.297 "name": "BaseBdev1", 00:13:44.297 "uuid": "9a479c4a-4e1f-4b18-89ab-1c27099eebe8", 00:13:44.297 "is_configured": true, 00:13:44.297 "data_offset": 2048, 00:13:44.297 "data_size": 63488 00:13:44.297 }, 00:13:44.297 { 00:13:44.297 "name": "BaseBdev2", 00:13:44.297 "uuid": "e9070e18-a25e-4409-8311-9570e61d7262", 00:13:44.297 "is_configured": true, 00:13:44.297 "data_offset": 2048, 00:13:44.297 "data_size": 63488 00:13:44.297 }, 00:13:44.297 { 00:13:44.297 "name": "BaseBdev3", 00:13:44.297 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:44.297 "is_configured": false, 00:13:44.297 "data_offset": 0, 00:13:44.297 "data_size": 0 00:13:44.297 }, 00:13:44.297 { 00:13:44.297 "name": "BaseBdev4", 00:13:44.297 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:44.297 "is_configured": false, 00:13:44.297 "data_offset": 0, 00:13:44.297 "data_size": 0 00:13:44.297 } 00:13:44.297 ] 00:13:44.297 }' 00:13:44.297 23:47:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:44.297 23:47:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:44.556 23:47:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:13:44.556 23:47:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:44.556 23:47:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:44.556 [2024-11-26 23:47:32.593875] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:13:44.556 BaseBdev3 00:13:44.556 23:47:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:44.556 23:47:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:13:44.556 23:47:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:13:44.556 23:47:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:13:44.556 23:47:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:13:44.556 23:47:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:13:44.556 23:47:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:13:44.556 23:47:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:13:44.556 23:47:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:44.556 23:47:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:44.556 23:47:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:44.556 23:47:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:13:44.556 23:47:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:44.556 23:47:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:44.556 [ 00:13:44.556 { 00:13:44.556 "name": "BaseBdev3", 00:13:44.556 "aliases": [ 00:13:44.556 "2184baa7-6763-4dee-8b7d-dd82b800d41e" 00:13:44.556 ], 00:13:44.556 "product_name": "Malloc disk", 00:13:44.556 "block_size": 512, 00:13:44.556 "num_blocks": 65536, 00:13:44.556 "uuid": "2184baa7-6763-4dee-8b7d-dd82b800d41e", 00:13:44.556 "assigned_rate_limits": { 00:13:44.556 "rw_ios_per_sec": 0, 00:13:44.556 "rw_mbytes_per_sec": 0, 00:13:44.556 "r_mbytes_per_sec": 0, 00:13:44.556 "w_mbytes_per_sec": 0 00:13:44.556 }, 00:13:44.556 "claimed": true, 00:13:44.557 "claim_type": "exclusive_write", 00:13:44.557 "zoned": false, 00:13:44.557 "supported_io_types": { 00:13:44.557 "read": true, 00:13:44.557 "write": true, 00:13:44.557 "unmap": true, 00:13:44.557 "flush": true, 00:13:44.557 "reset": true, 00:13:44.557 "nvme_admin": false, 00:13:44.557 "nvme_io": false, 00:13:44.557 "nvme_io_md": false, 00:13:44.557 "write_zeroes": true, 00:13:44.557 "zcopy": true, 00:13:44.557 "get_zone_info": false, 00:13:44.557 "zone_management": false, 00:13:44.557 "zone_append": false, 00:13:44.557 "compare": false, 00:13:44.557 "compare_and_write": false, 00:13:44.557 "abort": true, 00:13:44.557 "seek_hole": false, 00:13:44.557 "seek_data": false, 00:13:44.557 "copy": true, 00:13:44.557 "nvme_iov_md": false 00:13:44.557 }, 00:13:44.557 "memory_domains": [ 00:13:44.557 { 00:13:44.557 "dma_device_id": "system", 00:13:44.557 "dma_device_type": 1 00:13:44.557 }, 00:13:44.557 { 00:13:44.557 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:44.557 "dma_device_type": 2 00:13:44.557 } 00:13:44.557 ], 00:13:44.557 "driver_specific": {} 00:13:44.557 } 00:13:44.557 ] 00:13:44.557 23:47:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:44.557 23:47:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:13:44.557 23:47:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:13:44.557 23:47:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:13:44.557 23:47:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:13:44.557 23:47:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:44.557 23:47:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:44.557 23:47:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:44.557 23:47:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:44.557 23:47:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:13:44.557 23:47:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:44.557 23:47:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:44.557 23:47:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:44.557 23:47:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:44.557 23:47:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:44.557 23:47:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:44.557 23:47:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:44.557 23:47:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:44.557 23:47:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:44.816 23:47:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:44.816 "name": "Existed_Raid", 00:13:44.816 "uuid": "90d55369-d0bc-45ea-afd7-c3c4867c9766", 00:13:44.816 "strip_size_kb": 64, 00:13:44.816 "state": "configuring", 00:13:44.816 "raid_level": "raid5f", 00:13:44.816 "superblock": true, 00:13:44.816 "num_base_bdevs": 4, 00:13:44.816 "num_base_bdevs_discovered": 3, 00:13:44.816 "num_base_bdevs_operational": 4, 00:13:44.816 "base_bdevs_list": [ 00:13:44.816 { 00:13:44.816 "name": "BaseBdev1", 00:13:44.816 "uuid": "9a479c4a-4e1f-4b18-89ab-1c27099eebe8", 00:13:44.816 "is_configured": true, 00:13:44.816 "data_offset": 2048, 00:13:44.816 "data_size": 63488 00:13:44.816 }, 00:13:44.816 { 00:13:44.816 "name": "BaseBdev2", 00:13:44.816 "uuid": "e9070e18-a25e-4409-8311-9570e61d7262", 00:13:44.816 "is_configured": true, 00:13:44.816 "data_offset": 2048, 00:13:44.816 "data_size": 63488 00:13:44.816 }, 00:13:44.816 { 00:13:44.816 "name": "BaseBdev3", 00:13:44.816 "uuid": "2184baa7-6763-4dee-8b7d-dd82b800d41e", 00:13:44.816 "is_configured": true, 00:13:44.816 "data_offset": 2048, 00:13:44.816 "data_size": 63488 00:13:44.816 }, 00:13:44.816 { 00:13:44.816 "name": "BaseBdev4", 00:13:44.816 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:44.816 "is_configured": false, 00:13:44.816 "data_offset": 0, 00:13:44.816 "data_size": 0 00:13:44.816 } 00:13:44.816 ] 00:13:44.816 }' 00:13:44.816 23:47:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:44.816 23:47:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:45.076 23:47:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:13:45.076 23:47:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:45.076 23:47:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:45.076 [2024-11-26 23:47:33.063767] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:13:45.076 [2024-11-26 23:47:33.064038] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:13:45.076 [2024-11-26 23:47:33.064077] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:13:45.076 BaseBdev4 00:13:45.076 [2024-11-26 23:47:33.064387] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:13:45.076 [2024-11-26 23:47:33.064863] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:13:45.076 [2024-11-26 23:47:33.064925] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:13:45.076 [2024-11-26 23:47:33.065100] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:45.076 23:47:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:45.076 23:47:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev4 00:13:45.076 23:47:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:13:45.076 23:47:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:13:45.076 23:47:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:13:45.076 23:47:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:13:45.076 23:47:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:13:45.076 23:47:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:13:45.076 23:47:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:45.076 23:47:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:45.076 23:47:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:45.076 23:47:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:13:45.076 23:47:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:45.076 23:47:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:45.076 [ 00:13:45.076 { 00:13:45.076 "name": "BaseBdev4", 00:13:45.076 "aliases": [ 00:13:45.076 "a87d2fcd-3b6c-4a2c-b846-a3c5a3666206" 00:13:45.076 ], 00:13:45.076 "product_name": "Malloc disk", 00:13:45.076 "block_size": 512, 00:13:45.076 "num_blocks": 65536, 00:13:45.076 "uuid": "a87d2fcd-3b6c-4a2c-b846-a3c5a3666206", 00:13:45.076 "assigned_rate_limits": { 00:13:45.076 "rw_ios_per_sec": 0, 00:13:45.076 "rw_mbytes_per_sec": 0, 00:13:45.076 "r_mbytes_per_sec": 0, 00:13:45.076 "w_mbytes_per_sec": 0 00:13:45.076 }, 00:13:45.076 "claimed": true, 00:13:45.076 "claim_type": "exclusive_write", 00:13:45.076 "zoned": false, 00:13:45.076 "supported_io_types": { 00:13:45.076 "read": true, 00:13:45.076 "write": true, 00:13:45.076 "unmap": true, 00:13:45.076 "flush": true, 00:13:45.076 "reset": true, 00:13:45.076 "nvme_admin": false, 00:13:45.076 "nvme_io": false, 00:13:45.076 "nvme_io_md": false, 00:13:45.076 "write_zeroes": true, 00:13:45.076 "zcopy": true, 00:13:45.076 "get_zone_info": false, 00:13:45.076 "zone_management": false, 00:13:45.076 "zone_append": false, 00:13:45.076 "compare": false, 00:13:45.076 "compare_and_write": false, 00:13:45.076 "abort": true, 00:13:45.076 "seek_hole": false, 00:13:45.076 "seek_data": false, 00:13:45.076 "copy": true, 00:13:45.076 "nvme_iov_md": false 00:13:45.076 }, 00:13:45.076 "memory_domains": [ 00:13:45.076 { 00:13:45.076 "dma_device_id": "system", 00:13:45.076 "dma_device_type": 1 00:13:45.076 }, 00:13:45.076 { 00:13:45.076 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:45.076 "dma_device_type": 2 00:13:45.076 } 00:13:45.076 ], 00:13:45.076 "driver_specific": {} 00:13:45.076 } 00:13:45.076 ] 00:13:45.076 23:47:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:45.076 23:47:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:13:45.076 23:47:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:13:45.076 23:47:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:13:45.076 23:47:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 4 00:13:45.076 23:47:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:45.076 23:47:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:45.076 23:47:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:45.076 23:47:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:45.076 23:47:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:13:45.076 23:47:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:45.076 23:47:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:45.076 23:47:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:45.077 23:47:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:45.077 23:47:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:45.077 23:47:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:45.077 23:47:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:45.077 23:47:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:45.077 23:47:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:45.077 23:47:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:45.077 "name": "Existed_Raid", 00:13:45.077 "uuid": "90d55369-d0bc-45ea-afd7-c3c4867c9766", 00:13:45.077 "strip_size_kb": 64, 00:13:45.077 "state": "online", 00:13:45.077 "raid_level": "raid5f", 00:13:45.077 "superblock": true, 00:13:45.077 "num_base_bdevs": 4, 00:13:45.077 "num_base_bdevs_discovered": 4, 00:13:45.077 "num_base_bdevs_operational": 4, 00:13:45.077 "base_bdevs_list": [ 00:13:45.077 { 00:13:45.077 "name": "BaseBdev1", 00:13:45.077 "uuid": "9a479c4a-4e1f-4b18-89ab-1c27099eebe8", 00:13:45.077 "is_configured": true, 00:13:45.077 "data_offset": 2048, 00:13:45.077 "data_size": 63488 00:13:45.077 }, 00:13:45.077 { 00:13:45.077 "name": "BaseBdev2", 00:13:45.077 "uuid": "e9070e18-a25e-4409-8311-9570e61d7262", 00:13:45.077 "is_configured": true, 00:13:45.077 "data_offset": 2048, 00:13:45.077 "data_size": 63488 00:13:45.077 }, 00:13:45.077 { 00:13:45.077 "name": "BaseBdev3", 00:13:45.077 "uuid": "2184baa7-6763-4dee-8b7d-dd82b800d41e", 00:13:45.077 "is_configured": true, 00:13:45.077 "data_offset": 2048, 00:13:45.077 "data_size": 63488 00:13:45.077 }, 00:13:45.077 { 00:13:45.077 "name": "BaseBdev4", 00:13:45.077 "uuid": "a87d2fcd-3b6c-4a2c-b846-a3c5a3666206", 00:13:45.077 "is_configured": true, 00:13:45.077 "data_offset": 2048, 00:13:45.077 "data_size": 63488 00:13:45.077 } 00:13:45.077 ] 00:13:45.077 }' 00:13:45.077 23:47:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:45.077 23:47:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:45.646 23:47:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:13:45.646 23:47:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:13:45.646 23:47:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:13:45.646 23:47:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:13:45.646 23:47:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:13:45.646 23:47:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:13:45.646 23:47:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:13:45.646 23:47:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:45.646 23:47:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:45.646 23:47:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:13:45.646 [2024-11-26 23:47:33.555247] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:13:45.646 23:47:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:45.646 23:47:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:13:45.646 "name": "Existed_Raid", 00:13:45.646 "aliases": [ 00:13:45.646 "90d55369-d0bc-45ea-afd7-c3c4867c9766" 00:13:45.646 ], 00:13:45.646 "product_name": "Raid Volume", 00:13:45.646 "block_size": 512, 00:13:45.646 "num_blocks": 190464, 00:13:45.646 "uuid": "90d55369-d0bc-45ea-afd7-c3c4867c9766", 00:13:45.646 "assigned_rate_limits": { 00:13:45.646 "rw_ios_per_sec": 0, 00:13:45.646 "rw_mbytes_per_sec": 0, 00:13:45.646 "r_mbytes_per_sec": 0, 00:13:45.646 "w_mbytes_per_sec": 0 00:13:45.646 }, 00:13:45.646 "claimed": false, 00:13:45.646 "zoned": false, 00:13:45.646 "supported_io_types": { 00:13:45.646 "read": true, 00:13:45.646 "write": true, 00:13:45.646 "unmap": false, 00:13:45.646 "flush": false, 00:13:45.646 "reset": true, 00:13:45.646 "nvme_admin": false, 00:13:45.646 "nvme_io": false, 00:13:45.646 "nvme_io_md": false, 00:13:45.646 "write_zeroes": true, 00:13:45.646 "zcopy": false, 00:13:45.646 "get_zone_info": false, 00:13:45.646 "zone_management": false, 00:13:45.646 "zone_append": false, 00:13:45.646 "compare": false, 00:13:45.646 "compare_and_write": false, 00:13:45.646 "abort": false, 00:13:45.646 "seek_hole": false, 00:13:45.646 "seek_data": false, 00:13:45.646 "copy": false, 00:13:45.646 "nvme_iov_md": false 00:13:45.646 }, 00:13:45.646 "driver_specific": { 00:13:45.646 "raid": { 00:13:45.646 "uuid": "90d55369-d0bc-45ea-afd7-c3c4867c9766", 00:13:45.646 "strip_size_kb": 64, 00:13:45.646 "state": "online", 00:13:45.646 "raid_level": "raid5f", 00:13:45.646 "superblock": true, 00:13:45.646 "num_base_bdevs": 4, 00:13:45.646 "num_base_bdevs_discovered": 4, 00:13:45.646 "num_base_bdevs_operational": 4, 00:13:45.646 "base_bdevs_list": [ 00:13:45.646 { 00:13:45.646 "name": "BaseBdev1", 00:13:45.646 "uuid": "9a479c4a-4e1f-4b18-89ab-1c27099eebe8", 00:13:45.646 "is_configured": true, 00:13:45.646 "data_offset": 2048, 00:13:45.646 "data_size": 63488 00:13:45.646 }, 00:13:45.646 { 00:13:45.646 "name": "BaseBdev2", 00:13:45.646 "uuid": "e9070e18-a25e-4409-8311-9570e61d7262", 00:13:45.646 "is_configured": true, 00:13:45.646 "data_offset": 2048, 00:13:45.646 "data_size": 63488 00:13:45.646 }, 00:13:45.646 { 00:13:45.646 "name": "BaseBdev3", 00:13:45.646 "uuid": "2184baa7-6763-4dee-8b7d-dd82b800d41e", 00:13:45.646 "is_configured": true, 00:13:45.646 "data_offset": 2048, 00:13:45.646 "data_size": 63488 00:13:45.646 }, 00:13:45.646 { 00:13:45.646 "name": "BaseBdev4", 00:13:45.646 "uuid": "a87d2fcd-3b6c-4a2c-b846-a3c5a3666206", 00:13:45.646 "is_configured": true, 00:13:45.646 "data_offset": 2048, 00:13:45.646 "data_size": 63488 00:13:45.646 } 00:13:45.646 ] 00:13:45.646 } 00:13:45.646 } 00:13:45.646 }' 00:13:45.646 23:47:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:13:45.646 23:47:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:13:45.646 BaseBdev2 00:13:45.646 BaseBdev3 00:13:45.646 BaseBdev4' 00:13:45.646 23:47:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:45.646 23:47:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:13:45.646 23:47:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:45.646 23:47:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:45.646 23:47:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:13:45.646 23:47:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:45.646 23:47:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:45.646 23:47:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:45.646 23:47:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:45.646 23:47:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:45.646 23:47:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:45.646 23:47:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:45.646 23:47:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:13:45.646 23:47:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:45.646 23:47:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:45.646 23:47:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:45.646 23:47:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:45.647 23:47:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:45.647 23:47:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:45.647 23:47:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:13:45.647 23:47:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:45.647 23:47:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:45.647 23:47:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:45.647 23:47:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:45.908 23:47:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:45.908 23:47:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:45.908 23:47:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:45.909 23:47:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:13:45.909 23:47:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:45.909 23:47:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:45.909 23:47:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:45.909 23:47:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:45.909 23:47:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:45.909 23:47:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:45.909 23:47:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:13:45.909 23:47:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:45.909 23:47:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:45.909 [2024-11-26 23:47:33.850602] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:13:45.909 23:47:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:45.909 23:47:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:13:45.909 23:47:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy raid5f 00:13:45.909 23:47:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:13:45.909 23:47:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@199 -- # return 0 00:13:45.909 23:47:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:13:45.909 23:47:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 3 00:13:45.909 23:47:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:45.909 23:47:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:45.909 23:47:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:45.909 23:47:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:45.909 23:47:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:45.909 23:47:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:45.909 23:47:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:45.909 23:47:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:45.909 23:47:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:45.909 23:47:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:45.909 23:47:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:45.909 23:47:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:45.909 23:47:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:45.909 23:47:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:45.909 23:47:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:45.909 "name": "Existed_Raid", 00:13:45.909 "uuid": "90d55369-d0bc-45ea-afd7-c3c4867c9766", 00:13:45.909 "strip_size_kb": 64, 00:13:45.909 "state": "online", 00:13:45.909 "raid_level": "raid5f", 00:13:45.909 "superblock": true, 00:13:45.909 "num_base_bdevs": 4, 00:13:45.909 "num_base_bdevs_discovered": 3, 00:13:45.909 "num_base_bdevs_operational": 3, 00:13:45.909 "base_bdevs_list": [ 00:13:45.909 { 00:13:45.909 "name": null, 00:13:45.909 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:45.909 "is_configured": false, 00:13:45.909 "data_offset": 0, 00:13:45.909 "data_size": 63488 00:13:45.909 }, 00:13:45.909 { 00:13:45.909 "name": "BaseBdev2", 00:13:45.909 "uuid": "e9070e18-a25e-4409-8311-9570e61d7262", 00:13:45.909 "is_configured": true, 00:13:45.909 "data_offset": 2048, 00:13:45.909 "data_size": 63488 00:13:45.909 }, 00:13:45.909 { 00:13:45.909 "name": "BaseBdev3", 00:13:45.909 "uuid": "2184baa7-6763-4dee-8b7d-dd82b800d41e", 00:13:45.909 "is_configured": true, 00:13:45.909 "data_offset": 2048, 00:13:45.909 "data_size": 63488 00:13:45.909 }, 00:13:45.909 { 00:13:45.909 "name": "BaseBdev4", 00:13:45.909 "uuid": "a87d2fcd-3b6c-4a2c-b846-a3c5a3666206", 00:13:45.909 "is_configured": true, 00:13:45.909 "data_offset": 2048, 00:13:45.909 "data_size": 63488 00:13:45.909 } 00:13:45.909 ] 00:13:45.909 }' 00:13:45.909 23:47:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:45.909 23:47:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:46.166 23:47:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:13:46.166 23:47:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:13:46.166 23:47:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:46.166 23:47:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:46.166 23:47:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:46.166 23:47:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:13:46.425 23:47:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:46.425 23:47:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:13:46.425 23:47:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:13:46.425 23:47:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:13:46.425 23:47:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:46.425 23:47:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:46.425 [2024-11-26 23:47:34.325053] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:13:46.425 [2024-11-26 23:47:34.325190] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:13:46.425 [2024-11-26 23:47:34.336112] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:13:46.425 23:47:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:46.425 23:47:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:13:46.425 23:47:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:13:46.425 23:47:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:46.425 23:47:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:13:46.425 23:47:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:46.425 23:47:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:46.425 23:47:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:46.425 23:47:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:13:46.425 23:47:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:13:46.425 23:47:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:13:46.425 23:47:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:46.425 23:47:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:46.425 [2024-11-26 23:47:34.396038] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:13:46.425 23:47:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:46.425 23:47:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:13:46.425 23:47:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:13:46.425 23:47:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:13:46.425 23:47:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:46.425 23:47:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:46.425 23:47:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:46.425 23:47:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:46.425 23:47:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:13:46.425 23:47:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:13:46.425 23:47:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev4 00:13:46.425 23:47:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:46.425 23:47:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:46.425 [2024-11-26 23:47:34.462806] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev4 00:13:46.425 [2024-11-26 23:47:34.462895] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:13:46.425 23:47:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:46.425 23:47:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:13:46.425 23:47:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:13:46.425 23:47:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:46.425 23:47:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:46.425 23:47:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:46.425 23:47:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:13:46.425 23:47:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:46.425 23:47:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:13:46.425 23:47:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:13:46.425 23:47:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 4 -gt 2 ']' 00:13:46.425 23:47:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:13:46.425 23:47:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:13:46.425 23:47:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:13:46.425 23:47:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:46.425 23:47:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:46.425 BaseBdev2 00:13:46.425 23:47:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:46.425 23:47:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:13:46.425 23:47:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:13:46.425 23:47:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:13:46.425 23:47:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:13:46.425 23:47:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:13:46.425 23:47:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:13:46.425 23:47:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:13:46.425 23:47:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:46.425 23:47:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:46.684 23:47:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:46.684 23:47:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:13:46.684 23:47:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:46.684 23:47:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:46.684 [ 00:13:46.684 { 00:13:46.684 "name": "BaseBdev2", 00:13:46.684 "aliases": [ 00:13:46.684 "907b7fa0-7aee-494b-9b09-6d264cbfa894" 00:13:46.684 ], 00:13:46.684 "product_name": "Malloc disk", 00:13:46.684 "block_size": 512, 00:13:46.684 "num_blocks": 65536, 00:13:46.684 "uuid": "907b7fa0-7aee-494b-9b09-6d264cbfa894", 00:13:46.684 "assigned_rate_limits": { 00:13:46.684 "rw_ios_per_sec": 0, 00:13:46.684 "rw_mbytes_per_sec": 0, 00:13:46.684 "r_mbytes_per_sec": 0, 00:13:46.684 "w_mbytes_per_sec": 0 00:13:46.684 }, 00:13:46.684 "claimed": false, 00:13:46.684 "zoned": false, 00:13:46.684 "supported_io_types": { 00:13:46.684 "read": true, 00:13:46.684 "write": true, 00:13:46.684 "unmap": true, 00:13:46.684 "flush": true, 00:13:46.684 "reset": true, 00:13:46.684 "nvme_admin": false, 00:13:46.684 "nvme_io": false, 00:13:46.684 "nvme_io_md": false, 00:13:46.684 "write_zeroes": true, 00:13:46.684 "zcopy": true, 00:13:46.684 "get_zone_info": false, 00:13:46.684 "zone_management": false, 00:13:46.684 "zone_append": false, 00:13:46.684 "compare": false, 00:13:46.684 "compare_and_write": false, 00:13:46.684 "abort": true, 00:13:46.684 "seek_hole": false, 00:13:46.684 "seek_data": false, 00:13:46.684 "copy": true, 00:13:46.684 "nvme_iov_md": false 00:13:46.684 }, 00:13:46.684 "memory_domains": [ 00:13:46.684 { 00:13:46.684 "dma_device_id": "system", 00:13:46.684 "dma_device_type": 1 00:13:46.684 }, 00:13:46.684 { 00:13:46.684 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:46.684 "dma_device_type": 2 00:13:46.684 } 00:13:46.684 ], 00:13:46.684 "driver_specific": {} 00:13:46.684 } 00:13:46.684 ] 00:13:46.684 23:47:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:46.684 23:47:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:13:46.684 23:47:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:13:46.684 23:47:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:13:46.684 23:47:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:13:46.684 23:47:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:46.684 23:47:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:46.684 BaseBdev3 00:13:46.684 23:47:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:46.684 23:47:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:13:46.684 23:47:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:13:46.684 23:47:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:13:46.684 23:47:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:13:46.684 23:47:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:13:46.684 23:47:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:13:46.684 23:47:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:13:46.684 23:47:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:46.684 23:47:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:46.684 23:47:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:46.684 23:47:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:13:46.684 23:47:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:46.684 23:47:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:46.684 [ 00:13:46.684 { 00:13:46.684 "name": "BaseBdev3", 00:13:46.684 "aliases": [ 00:13:46.684 "fa2ce605-4511-4edc-b079-6d3eeb395f64" 00:13:46.684 ], 00:13:46.684 "product_name": "Malloc disk", 00:13:46.684 "block_size": 512, 00:13:46.684 "num_blocks": 65536, 00:13:46.684 "uuid": "fa2ce605-4511-4edc-b079-6d3eeb395f64", 00:13:46.684 "assigned_rate_limits": { 00:13:46.684 "rw_ios_per_sec": 0, 00:13:46.684 "rw_mbytes_per_sec": 0, 00:13:46.684 "r_mbytes_per_sec": 0, 00:13:46.684 "w_mbytes_per_sec": 0 00:13:46.684 }, 00:13:46.684 "claimed": false, 00:13:46.684 "zoned": false, 00:13:46.684 "supported_io_types": { 00:13:46.684 "read": true, 00:13:46.684 "write": true, 00:13:46.684 "unmap": true, 00:13:46.684 "flush": true, 00:13:46.684 "reset": true, 00:13:46.684 "nvme_admin": false, 00:13:46.684 "nvme_io": false, 00:13:46.684 "nvme_io_md": false, 00:13:46.684 "write_zeroes": true, 00:13:46.684 "zcopy": true, 00:13:46.684 "get_zone_info": false, 00:13:46.684 "zone_management": false, 00:13:46.684 "zone_append": false, 00:13:46.684 "compare": false, 00:13:46.684 "compare_and_write": false, 00:13:46.684 "abort": true, 00:13:46.684 "seek_hole": false, 00:13:46.684 "seek_data": false, 00:13:46.684 "copy": true, 00:13:46.684 "nvme_iov_md": false 00:13:46.684 }, 00:13:46.684 "memory_domains": [ 00:13:46.684 { 00:13:46.684 "dma_device_id": "system", 00:13:46.684 "dma_device_type": 1 00:13:46.684 }, 00:13:46.684 { 00:13:46.684 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:46.684 "dma_device_type": 2 00:13:46.684 } 00:13:46.684 ], 00:13:46.684 "driver_specific": {} 00:13:46.684 } 00:13:46.684 ] 00:13:46.684 23:47:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:46.684 23:47:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:13:46.684 23:47:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:13:46.684 23:47:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:13:46.684 23:47:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:13:46.684 23:47:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:46.684 23:47:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:46.684 BaseBdev4 00:13:46.684 23:47:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:46.684 23:47:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev4 00:13:46.684 23:47:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:13:46.684 23:47:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:13:46.684 23:47:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:13:46.684 23:47:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:13:46.685 23:47:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:13:46.685 23:47:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:13:46.685 23:47:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:46.685 23:47:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:46.685 23:47:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:46.685 23:47:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:13:46.685 23:47:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:46.685 23:47:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:46.685 [ 00:13:46.685 { 00:13:46.685 "name": "BaseBdev4", 00:13:46.685 "aliases": [ 00:13:46.685 "36a4ce93-eac5-4d5b-8659-364179f27ebf" 00:13:46.685 ], 00:13:46.685 "product_name": "Malloc disk", 00:13:46.685 "block_size": 512, 00:13:46.685 "num_blocks": 65536, 00:13:46.685 "uuid": "36a4ce93-eac5-4d5b-8659-364179f27ebf", 00:13:46.685 "assigned_rate_limits": { 00:13:46.685 "rw_ios_per_sec": 0, 00:13:46.685 "rw_mbytes_per_sec": 0, 00:13:46.685 "r_mbytes_per_sec": 0, 00:13:46.685 "w_mbytes_per_sec": 0 00:13:46.685 }, 00:13:46.685 "claimed": false, 00:13:46.685 "zoned": false, 00:13:46.685 "supported_io_types": { 00:13:46.685 "read": true, 00:13:46.685 "write": true, 00:13:46.685 "unmap": true, 00:13:46.685 "flush": true, 00:13:46.685 "reset": true, 00:13:46.685 "nvme_admin": false, 00:13:46.685 "nvme_io": false, 00:13:46.685 "nvme_io_md": false, 00:13:46.685 "write_zeroes": true, 00:13:46.685 "zcopy": true, 00:13:46.685 "get_zone_info": false, 00:13:46.685 "zone_management": false, 00:13:46.685 "zone_append": false, 00:13:46.685 "compare": false, 00:13:46.685 "compare_and_write": false, 00:13:46.685 "abort": true, 00:13:46.685 "seek_hole": false, 00:13:46.685 "seek_data": false, 00:13:46.685 "copy": true, 00:13:46.685 "nvme_iov_md": false 00:13:46.685 }, 00:13:46.685 "memory_domains": [ 00:13:46.685 { 00:13:46.685 "dma_device_id": "system", 00:13:46.685 "dma_device_type": 1 00:13:46.685 }, 00:13:46.685 { 00:13:46.685 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:46.685 "dma_device_type": 2 00:13:46.685 } 00:13:46.685 ], 00:13:46.685 "driver_specific": {} 00:13:46.685 } 00:13:46.685 ] 00:13:46.685 23:47:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:46.685 23:47:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:13:46.685 23:47:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:13:46.685 23:47:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:13:46.685 23:47:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:13:46.685 23:47:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:46.685 23:47:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:46.685 [2024-11-26 23:47:34.693652] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:13:46.685 [2024-11-26 23:47:34.693738] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:13:46.685 [2024-11-26 23:47:34.693784] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:13:46.685 [2024-11-26 23:47:34.695534] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:13:46.685 [2024-11-26 23:47:34.695617] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:13:46.685 23:47:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:46.685 23:47:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:13:46.685 23:47:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:46.685 23:47:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:46.685 23:47:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:46.685 23:47:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:46.685 23:47:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:13:46.685 23:47:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:46.685 23:47:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:46.685 23:47:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:46.685 23:47:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:46.685 23:47:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:46.685 23:47:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:46.685 23:47:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:46.685 23:47:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:46.685 23:47:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:46.685 23:47:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:46.685 "name": "Existed_Raid", 00:13:46.685 "uuid": "a3cd89e1-9962-4fbf-8435-bb351e5ae63a", 00:13:46.685 "strip_size_kb": 64, 00:13:46.685 "state": "configuring", 00:13:46.685 "raid_level": "raid5f", 00:13:46.685 "superblock": true, 00:13:46.685 "num_base_bdevs": 4, 00:13:46.685 "num_base_bdevs_discovered": 3, 00:13:46.685 "num_base_bdevs_operational": 4, 00:13:46.685 "base_bdevs_list": [ 00:13:46.685 { 00:13:46.685 "name": "BaseBdev1", 00:13:46.685 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:46.685 "is_configured": false, 00:13:46.685 "data_offset": 0, 00:13:46.685 "data_size": 0 00:13:46.685 }, 00:13:46.685 { 00:13:46.685 "name": "BaseBdev2", 00:13:46.685 "uuid": "907b7fa0-7aee-494b-9b09-6d264cbfa894", 00:13:46.685 "is_configured": true, 00:13:46.685 "data_offset": 2048, 00:13:46.685 "data_size": 63488 00:13:46.685 }, 00:13:46.685 { 00:13:46.685 "name": "BaseBdev3", 00:13:46.685 "uuid": "fa2ce605-4511-4edc-b079-6d3eeb395f64", 00:13:46.685 "is_configured": true, 00:13:46.685 "data_offset": 2048, 00:13:46.685 "data_size": 63488 00:13:46.685 }, 00:13:46.685 { 00:13:46.685 "name": "BaseBdev4", 00:13:46.685 "uuid": "36a4ce93-eac5-4d5b-8659-364179f27ebf", 00:13:46.685 "is_configured": true, 00:13:46.685 "data_offset": 2048, 00:13:46.685 "data_size": 63488 00:13:46.685 } 00:13:46.685 ] 00:13:46.685 }' 00:13:46.685 23:47:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:46.685 23:47:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:47.251 23:47:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:13:47.251 23:47:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:47.251 23:47:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:47.251 [2024-11-26 23:47:35.084932] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:13:47.251 23:47:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:47.251 23:47:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:13:47.251 23:47:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:47.251 23:47:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:47.251 23:47:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:47.251 23:47:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:47.251 23:47:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:13:47.251 23:47:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:47.251 23:47:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:47.252 23:47:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:47.252 23:47:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:47.252 23:47:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:47.252 23:47:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:47.252 23:47:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:47.252 23:47:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:47.252 23:47:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:47.252 23:47:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:47.252 "name": "Existed_Raid", 00:13:47.252 "uuid": "a3cd89e1-9962-4fbf-8435-bb351e5ae63a", 00:13:47.252 "strip_size_kb": 64, 00:13:47.252 "state": "configuring", 00:13:47.252 "raid_level": "raid5f", 00:13:47.252 "superblock": true, 00:13:47.252 "num_base_bdevs": 4, 00:13:47.252 "num_base_bdevs_discovered": 2, 00:13:47.252 "num_base_bdevs_operational": 4, 00:13:47.252 "base_bdevs_list": [ 00:13:47.252 { 00:13:47.252 "name": "BaseBdev1", 00:13:47.252 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:47.252 "is_configured": false, 00:13:47.252 "data_offset": 0, 00:13:47.252 "data_size": 0 00:13:47.252 }, 00:13:47.252 { 00:13:47.252 "name": null, 00:13:47.252 "uuid": "907b7fa0-7aee-494b-9b09-6d264cbfa894", 00:13:47.252 "is_configured": false, 00:13:47.252 "data_offset": 0, 00:13:47.252 "data_size": 63488 00:13:47.252 }, 00:13:47.252 { 00:13:47.252 "name": "BaseBdev3", 00:13:47.252 "uuid": "fa2ce605-4511-4edc-b079-6d3eeb395f64", 00:13:47.252 "is_configured": true, 00:13:47.252 "data_offset": 2048, 00:13:47.252 "data_size": 63488 00:13:47.252 }, 00:13:47.252 { 00:13:47.252 "name": "BaseBdev4", 00:13:47.252 "uuid": "36a4ce93-eac5-4d5b-8659-364179f27ebf", 00:13:47.252 "is_configured": true, 00:13:47.252 "data_offset": 2048, 00:13:47.252 "data_size": 63488 00:13:47.252 } 00:13:47.252 ] 00:13:47.252 }' 00:13:47.252 23:47:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:47.252 23:47:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:47.509 23:47:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:47.509 23:47:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:13:47.509 23:47:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:47.509 23:47:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:47.509 23:47:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:47.509 23:47:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:13:47.509 23:47:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:13:47.509 23:47:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:47.509 23:47:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:47.509 [2024-11-26 23:47:35.542872] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:13:47.509 BaseBdev1 00:13:47.509 23:47:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:47.509 23:47:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:13:47.509 23:47:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:13:47.509 23:47:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:13:47.509 23:47:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:13:47.509 23:47:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:13:47.509 23:47:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:13:47.509 23:47:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:13:47.509 23:47:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:47.509 23:47:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:47.509 23:47:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:47.509 23:47:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:13:47.509 23:47:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:47.509 23:47:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:47.509 [ 00:13:47.509 { 00:13:47.509 "name": "BaseBdev1", 00:13:47.509 "aliases": [ 00:13:47.509 "c176b910-f12f-4f78-90bd-7dbcbab364cb" 00:13:47.509 ], 00:13:47.509 "product_name": "Malloc disk", 00:13:47.509 "block_size": 512, 00:13:47.509 "num_blocks": 65536, 00:13:47.509 "uuid": "c176b910-f12f-4f78-90bd-7dbcbab364cb", 00:13:47.509 "assigned_rate_limits": { 00:13:47.509 "rw_ios_per_sec": 0, 00:13:47.509 "rw_mbytes_per_sec": 0, 00:13:47.509 "r_mbytes_per_sec": 0, 00:13:47.509 "w_mbytes_per_sec": 0 00:13:47.509 }, 00:13:47.509 "claimed": true, 00:13:47.509 "claim_type": "exclusive_write", 00:13:47.509 "zoned": false, 00:13:47.509 "supported_io_types": { 00:13:47.509 "read": true, 00:13:47.509 "write": true, 00:13:47.509 "unmap": true, 00:13:47.509 "flush": true, 00:13:47.509 "reset": true, 00:13:47.509 "nvme_admin": false, 00:13:47.509 "nvme_io": false, 00:13:47.509 "nvme_io_md": false, 00:13:47.509 "write_zeroes": true, 00:13:47.509 "zcopy": true, 00:13:47.509 "get_zone_info": false, 00:13:47.509 "zone_management": false, 00:13:47.509 "zone_append": false, 00:13:47.509 "compare": false, 00:13:47.509 "compare_and_write": false, 00:13:47.509 "abort": true, 00:13:47.509 "seek_hole": false, 00:13:47.509 "seek_data": false, 00:13:47.509 "copy": true, 00:13:47.509 "nvme_iov_md": false 00:13:47.509 }, 00:13:47.509 "memory_domains": [ 00:13:47.509 { 00:13:47.509 "dma_device_id": "system", 00:13:47.509 "dma_device_type": 1 00:13:47.509 }, 00:13:47.509 { 00:13:47.509 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:47.509 "dma_device_type": 2 00:13:47.509 } 00:13:47.509 ], 00:13:47.509 "driver_specific": {} 00:13:47.509 } 00:13:47.509 ] 00:13:47.509 23:47:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:47.509 23:47:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:13:47.509 23:47:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:13:47.509 23:47:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:47.509 23:47:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:47.509 23:47:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:47.509 23:47:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:47.509 23:47:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:13:47.509 23:47:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:47.509 23:47:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:47.509 23:47:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:47.509 23:47:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:47.509 23:47:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:47.509 23:47:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:47.509 23:47:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:47.509 23:47:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:47.509 23:47:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:47.509 23:47:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:47.509 "name": "Existed_Raid", 00:13:47.509 "uuid": "a3cd89e1-9962-4fbf-8435-bb351e5ae63a", 00:13:47.509 "strip_size_kb": 64, 00:13:47.509 "state": "configuring", 00:13:47.509 "raid_level": "raid5f", 00:13:47.509 "superblock": true, 00:13:47.509 "num_base_bdevs": 4, 00:13:47.509 "num_base_bdevs_discovered": 3, 00:13:47.509 "num_base_bdevs_operational": 4, 00:13:47.509 "base_bdevs_list": [ 00:13:47.509 { 00:13:47.509 "name": "BaseBdev1", 00:13:47.509 "uuid": "c176b910-f12f-4f78-90bd-7dbcbab364cb", 00:13:47.509 "is_configured": true, 00:13:47.509 "data_offset": 2048, 00:13:47.509 "data_size": 63488 00:13:47.509 }, 00:13:47.509 { 00:13:47.509 "name": null, 00:13:47.509 "uuid": "907b7fa0-7aee-494b-9b09-6d264cbfa894", 00:13:47.509 "is_configured": false, 00:13:47.509 "data_offset": 0, 00:13:47.509 "data_size": 63488 00:13:47.509 }, 00:13:47.509 { 00:13:47.509 "name": "BaseBdev3", 00:13:47.509 "uuid": "fa2ce605-4511-4edc-b079-6d3eeb395f64", 00:13:47.509 "is_configured": true, 00:13:47.509 "data_offset": 2048, 00:13:47.509 "data_size": 63488 00:13:47.509 }, 00:13:47.509 { 00:13:47.509 "name": "BaseBdev4", 00:13:47.509 "uuid": "36a4ce93-eac5-4d5b-8659-364179f27ebf", 00:13:47.509 "is_configured": true, 00:13:47.509 "data_offset": 2048, 00:13:47.509 "data_size": 63488 00:13:47.509 } 00:13:47.509 ] 00:13:47.509 }' 00:13:47.509 23:47:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:47.509 23:47:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:48.075 23:47:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:48.075 23:47:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:48.075 23:47:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:48.075 23:47:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:13:48.075 23:47:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:48.075 23:47:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:13:48.075 23:47:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:13:48.075 23:47:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:48.075 23:47:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:48.075 [2024-11-26 23:47:36.058102] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:13:48.075 23:47:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:48.075 23:47:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:13:48.075 23:47:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:48.075 23:47:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:48.075 23:47:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:48.075 23:47:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:48.075 23:47:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:13:48.075 23:47:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:48.075 23:47:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:48.075 23:47:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:48.075 23:47:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:48.075 23:47:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:48.075 23:47:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:48.075 23:47:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:48.075 23:47:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:48.075 23:47:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:48.075 23:47:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:48.075 "name": "Existed_Raid", 00:13:48.075 "uuid": "a3cd89e1-9962-4fbf-8435-bb351e5ae63a", 00:13:48.075 "strip_size_kb": 64, 00:13:48.075 "state": "configuring", 00:13:48.075 "raid_level": "raid5f", 00:13:48.075 "superblock": true, 00:13:48.075 "num_base_bdevs": 4, 00:13:48.075 "num_base_bdevs_discovered": 2, 00:13:48.075 "num_base_bdevs_operational": 4, 00:13:48.075 "base_bdevs_list": [ 00:13:48.075 { 00:13:48.075 "name": "BaseBdev1", 00:13:48.075 "uuid": "c176b910-f12f-4f78-90bd-7dbcbab364cb", 00:13:48.075 "is_configured": true, 00:13:48.075 "data_offset": 2048, 00:13:48.075 "data_size": 63488 00:13:48.075 }, 00:13:48.075 { 00:13:48.075 "name": null, 00:13:48.075 "uuid": "907b7fa0-7aee-494b-9b09-6d264cbfa894", 00:13:48.075 "is_configured": false, 00:13:48.075 "data_offset": 0, 00:13:48.075 "data_size": 63488 00:13:48.075 }, 00:13:48.075 { 00:13:48.075 "name": null, 00:13:48.075 "uuid": "fa2ce605-4511-4edc-b079-6d3eeb395f64", 00:13:48.075 "is_configured": false, 00:13:48.075 "data_offset": 0, 00:13:48.075 "data_size": 63488 00:13:48.075 }, 00:13:48.075 { 00:13:48.075 "name": "BaseBdev4", 00:13:48.075 "uuid": "36a4ce93-eac5-4d5b-8659-364179f27ebf", 00:13:48.075 "is_configured": true, 00:13:48.075 "data_offset": 2048, 00:13:48.075 "data_size": 63488 00:13:48.075 } 00:13:48.075 ] 00:13:48.075 }' 00:13:48.075 23:47:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:48.075 23:47:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:48.642 23:47:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:48.642 23:47:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:48.642 23:47:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:48.642 23:47:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:13:48.642 23:47:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:48.642 23:47:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:13:48.642 23:47:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:13:48.642 23:47:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:48.642 23:47:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:48.642 [2024-11-26 23:47:36.525322] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:13:48.642 23:47:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:48.642 23:47:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:13:48.642 23:47:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:48.642 23:47:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:48.642 23:47:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:48.642 23:47:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:48.642 23:47:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:13:48.642 23:47:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:48.642 23:47:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:48.642 23:47:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:48.642 23:47:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:48.642 23:47:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:48.642 23:47:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:48.642 23:47:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:48.642 23:47:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:48.642 23:47:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:48.642 23:47:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:48.642 "name": "Existed_Raid", 00:13:48.642 "uuid": "a3cd89e1-9962-4fbf-8435-bb351e5ae63a", 00:13:48.642 "strip_size_kb": 64, 00:13:48.642 "state": "configuring", 00:13:48.642 "raid_level": "raid5f", 00:13:48.642 "superblock": true, 00:13:48.642 "num_base_bdevs": 4, 00:13:48.642 "num_base_bdevs_discovered": 3, 00:13:48.642 "num_base_bdevs_operational": 4, 00:13:48.642 "base_bdevs_list": [ 00:13:48.642 { 00:13:48.642 "name": "BaseBdev1", 00:13:48.642 "uuid": "c176b910-f12f-4f78-90bd-7dbcbab364cb", 00:13:48.642 "is_configured": true, 00:13:48.642 "data_offset": 2048, 00:13:48.642 "data_size": 63488 00:13:48.642 }, 00:13:48.642 { 00:13:48.642 "name": null, 00:13:48.642 "uuid": "907b7fa0-7aee-494b-9b09-6d264cbfa894", 00:13:48.642 "is_configured": false, 00:13:48.642 "data_offset": 0, 00:13:48.642 "data_size": 63488 00:13:48.642 }, 00:13:48.642 { 00:13:48.642 "name": "BaseBdev3", 00:13:48.642 "uuid": "fa2ce605-4511-4edc-b079-6d3eeb395f64", 00:13:48.643 "is_configured": true, 00:13:48.643 "data_offset": 2048, 00:13:48.643 "data_size": 63488 00:13:48.643 }, 00:13:48.643 { 00:13:48.643 "name": "BaseBdev4", 00:13:48.643 "uuid": "36a4ce93-eac5-4d5b-8659-364179f27ebf", 00:13:48.643 "is_configured": true, 00:13:48.643 "data_offset": 2048, 00:13:48.643 "data_size": 63488 00:13:48.643 } 00:13:48.643 ] 00:13:48.643 }' 00:13:48.643 23:47:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:48.643 23:47:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:48.901 23:47:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:48.901 23:47:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:48.901 23:47:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:48.901 23:47:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:13:48.901 23:47:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:48.901 23:47:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:13:48.901 23:47:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:13:48.901 23:47:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:48.902 23:47:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:48.902 [2024-11-26 23:47:37.000549] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:13:48.902 23:47:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:48.902 23:47:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:13:48.902 23:47:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:48.902 23:47:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:48.902 23:47:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:48.902 23:47:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:48.902 23:47:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:13:48.902 23:47:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:48.902 23:47:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:48.902 23:47:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:48.902 23:47:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:48.902 23:47:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:48.902 23:47:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:48.902 23:47:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:48.902 23:47:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:49.160 23:47:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:49.160 23:47:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:49.160 "name": "Existed_Raid", 00:13:49.160 "uuid": "a3cd89e1-9962-4fbf-8435-bb351e5ae63a", 00:13:49.160 "strip_size_kb": 64, 00:13:49.161 "state": "configuring", 00:13:49.161 "raid_level": "raid5f", 00:13:49.161 "superblock": true, 00:13:49.161 "num_base_bdevs": 4, 00:13:49.161 "num_base_bdevs_discovered": 2, 00:13:49.161 "num_base_bdevs_operational": 4, 00:13:49.161 "base_bdevs_list": [ 00:13:49.161 { 00:13:49.161 "name": null, 00:13:49.161 "uuid": "c176b910-f12f-4f78-90bd-7dbcbab364cb", 00:13:49.161 "is_configured": false, 00:13:49.161 "data_offset": 0, 00:13:49.161 "data_size": 63488 00:13:49.161 }, 00:13:49.161 { 00:13:49.161 "name": null, 00:13:49.161 "uuid": "907b7fa0-7aee-494b-9b09-6d264cbfa894", 00:13:49.161 "is_configured": false, 00:13:49.161 "data_offset": 0, 00:13:49.161 "data_size": 63488 00:13:49.161 }, 00:13:49.161 { 00:13:49.161 "name": "BaseBdev3", 00:13:49.161 "uuid": "fa2ce605-4511-4edc-b079-6d3eeb395f64", 00:13:49.161 "is_configured": true, 00:13:49.161 "data_offset": 2048, 00:13:49.161 "data_size": 63488 00:13:49.161 }, 00:13:49.161 { 00:13:49.161 "name": "BaseBdev4", 00:13:49.161 "uuid": "36a4ce93-eac5-4d5b-8659-364179f27ebf", 00:13:49.161 "is_configured": true, 00:13:49.161 "data_offset": 2048, 00:13:49.161 "data_size": 63488 00:13:49.161 } 00:13:49.161 ] 00:13:49.161 }' 00:13:49.161 23:47:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:49.161 23:47:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:49.419 23:47:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:13:49.419 23:47:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:49.419 23:47:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:49.419 23:47:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:49.419 23:47:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:49.419 23:47:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:13:49.419 23:47:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:13:49.419 23:47:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:49.419 23:47:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:49.419 [2024-11-26 23:47:37.458016] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:13:49.419 23:47:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:49.419 23:47:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:13:49.419 23:47:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:49.419 23:47:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:49.419 23:47:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:49.419 23:47:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:49.419 23:47:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:13:49.419 23:47:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:49.419 23:47:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:49.419 23:47:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:49.419 23:47:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:49.419 23:47:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:49.419 23:47:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:49.419 23:47:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:49.419 23:47:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:49.419 23:47:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:49.419 23:47:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:49.419 "name": "Existed_Raid", 00:13:49.419 "uuid": "a3cd89e1-9962-4fbf-8435-bb351e5ae63a", 00:13:49.419 "strip_size_kb": 64, 00:13:49.419 "state": "configuring", 00:13:49.419 "raid_level": "raid5f", 00:13:49.419 "superblock": true, 00:13:49.419 "num_base_bdevs": 4, 00:13:49.419 "num_base_bdevs_discovered": 3, 00:13:49.419 "num_base_bdevs_operational": 4, 00:13:49.419 "base_bdevs_list": [ 00:13:49.419 { 00:13:49.419 "name": null, 00:13:49.419 "uuid": "c176b910-f12f-4f78-90bd-7dbcbab364cb", 00:13:49.419 "is_configured": false, 00:13:49.419 "data_offset": 0, 00:13:49.419 "data_size": 63488 00:13:49.419 }, 00:13:49.419 { 00:13:49.419 "name": "BaseBdev2", 00:13:49.419 "uuid": "907b7fa0-7aee-494b-9b09-6d264cbfa894", 00:13:49.419 "is_configured": true, 00:13:49.419 "data_offset": 2048, 00:13:49.419 "data_size": 63488 00:13:49.419 }, 00:13:49.419 { 00:13:49.419 "name": "BaseBdev3", 00:13:49.419 "uuid": "fa2ce605-4511-4edc-b079-6d3eeb395f64", 00:13:49.419 "is_configured": true, 00:13:49.419 "data_offset": 2048, 00:13:49.419 "data_size": 63488 00:13:49.419 }, 00:13:49.419 { 00:13:49.419 "name": "BaseBdev4", 00:13:49.419 "uuid": "36a4ce93-eac5-4d5b-8659-364179f27ebf", 00:13:49.419 "is_configured": true, 00:13:49.419 "data_offset": 2048, 00:13:49.419 "data_size": 63488 00:13:49.419 } 00:13:49.419 ] 00:13:49.419 }' 00:13:49.419 23:47:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:49.419 23:47:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:49.986 23:47:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:49.986 23:47:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:49.986 23:47:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:49.986 23:47:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:13:49.986 23:47:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:49.986 23:47:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:13:49.986 23:47:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:49.986 23:47:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:13:49.986 23:47:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:49.986 23:47:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:49.986 23:47:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:49.986 23:47:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u c176b910-f12f-4f78-90bd-7dbcbab364cb 00:13:49.986 23:47:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:49.986 23:47:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:49.986 [2024-11-26 23:47:38.003920] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:13:49.986 [2024-11-26 23:47:38.004175] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:13:49.986 [2024-11-26 23:47:38.004223] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:13:49.986 [2024-11-26 23:47:38.004503] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002a10 00:13:49.986 NewBaseBdev 00:13:49.986 [2024-11-26 23:47:38.004971] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:13:49.986 [2024-11-26 23:47:38.005023] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001c80 00:13:49.986 23:47:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:49.986 [2024-11-26 23:47:38.005153] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:49.986 23:47:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:13:49.986 23:47:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:13:49.986 23:47:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:13:49.986 23:47:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:13:49.986 23:47:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:13:49.986 23:47:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:13:49.986 23:47:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:13:49.986 23:47:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:49.986 23:47:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:49.986 23:47:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:49.986 23:47:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:13:49.986 23:47:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:49.986 23:47:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:49.986 [ 00:13:49.986 { 00:13:49.986 "name": "NewBaseBdev", 00:13:49.986 "aliases": [ 00:13:49.986 "c176b910-f12f-4f78-90bd-7dbcbab364cb" 00:13:49.986 ], 00:13:49.986 "product_name": "Malloc disk", 00:13:49.986 "block_size": 512, 00:13:49.986 "num_blocks": 65536, 00:13:49.986 "uuid": "c176b910-f12f-4f78-90bd-7dbcbab364cb", 00:13:49.986 "assigned_rate_limits": { 00:13:49.986 "rw_ios_per_sec": 0, 00:13:49.986 "rw_mbytes_per_sec": 0, 00:13:49.986 "r_mbytes_per_sec": 0, 00:13:49.986 "w_mbytes_per_sec": 0 00:13:49.986 }, 00:13:49.986 "claimed": true, 00:13:49.986 "claim_type": "exclusive_write", 00:13:49.986 "zoned": false, 00:13:49.986 "supported_io_types": { 00:13:49.986 "read": true, 00:13:49.986 "write": true, 00:13:49.986 "unmap": true, 00:13:49.986 "flush": true, 00:13:49.986 "reset": true, 00:13:49.986 "nvme_admin": false, 00:13:49.986 "nvme_io": false, 00:13:49.986 "nvme_io_md": false, 00:13:49.986 "write_zeroes": true, 00:13:49.986 "zcopy": true, 00:13:49.986 "get_zone_info": false, 00:13:49.986 "zone_management": false, 00:13:49.986 "zone_append": false, 00:13:49.986 "compare": false, 00:13:49.986 "compare_and_write": false, 00:13:49.986 "abort": true, 00:13:49.986 "seek_hole": false, 00:13:49.986 "seek_data": false, 00:13:49.986 "copy": true, 00:13:49.986 "nvme_iov_md": false 00:13:49.986 }, 00:13:49.986 "memory_domains": [ 00:13:49.986 { 00:13:49.986 "dma_device_id": "system", 00:13:49.986 "dma_device_type": 1 00:13:49.986 }, 00:13:49.986 { 00:13:49.986 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:49.986 "dma_device_type": 2 00:13:49.986 } 00:13:49.986 ], 00:13:49.986 "driver_specific": {} 00:13:49.986 } 00:13:49.986 ] 00:13:49.986 23:47:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:49.986 23:47:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:13:49.986 23:47:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 4 00:13:49.986 23:47:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:49.986 23:47:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:49.986 23:47:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:49.986 23:47:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:49.986 23:47:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:13:49.986 23:47:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:49.986 23:47:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:49.986 23:47:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:49.986 23:47:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:49.986 23:47:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:49.986 23:47:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:49.986 23:47:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:49.986 23:47:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:49.986 23:47:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:49.986 23:47:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:49.986 "name": "Existed_Raid", 00:13:49.986 "uuid": "a3cd89e1-9962-4fbf-8435-bb351e5ae63a", 00:13:49.986 "strip_size_kb": 64, 00:13:49.986 "state": "online", 00:13:49.986 "raid_level": "raid5f", 00:13:49.986 "superblock": true, 00:13:49.986 "num_base_bdevs": 4, 00:13:49.986 "num_base_bdevs_discovered": 4, 00:13:49.986 "num_base_bdevs_operational": 4, 00:13:49.986 "base_bdevs_list": [ 00:13:49.986 { 00:13:49.986 "name": "NewBaseBdev", 00:13:49.986 "uuid": "c176b910-f12f-4f78-90bd-7dbcbab364cb", 00:13:49.986 "is_configured": true, 00:13:49.986 "data_offset": 2048, 00:13:49.986 "data_size": 63488 00:13:49.986 }, 00:13:49.986 { 00:13:49.986 "name": "BaseBdev2", 00:13:49.986 "uuid": "907b7fa0-7aee-494b-9b09-6d264cbfa894", 00:13:49.986 "is_configured": true, 00:13:49.986 "data_offset": 2048, 00:13:49.986 "data_size": 63488 00:13:49.986 }, 00:13:49.986 { 00:13:49.986 "name": "BaseBdev3", 00:13:49.986 "uuid": "fa2ce605-4511-4edc-b079-6d3eeb395f64", 00:13:49.986 "is_configured": true, 00:13:49.986 "data_offset": 2048, 00:13:49.986 "data_size": 63488 00:13:49.986 }, 00:13:49.986 { 00:13:49.986 "name": "BaseBdev4", 00:13:49.986 "uuid": "36a4ce93-eac5-4d5b-8659-364179f27ebf", 00:13:49.986 "is_configured": true, 00:13:49.986 "data_offset": 2048, 00:13:49.986 "data_size": 63488 00:13:49.986 } 00:13:49.986 ] 00:13:49.986 }' 00:13:49.986 23:47:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:49.986 23:47:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:50.554 23:47:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:13:50.554 23:47:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:13:50.554 23:47:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:13:50.554 23:47:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:13:50.554 23:47:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:13:50.554 23:47:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:13:50.554 23:47:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:13:50.554 23:47:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:13:50.554 23:47:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:50.554 23:47:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:50.554 [2024-11-26 23:47:38.411431] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:13:50.554 23:47:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:50.554 23:47:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:13:50.554 "name": "Existed_Raid", 00:13:50.554 "aliases": [ 00:13:50.554 "a3cd89e1-9962-4fbf-8435-bb351e5ae63a" 00:13:50.554 ], 00:13:50.554 "product_name": "Raid Volume", 00:13:50.554 "block_size": 512, 00:13:50.554 "num_blocks": 190464, 00:13:50.554 "uuid": "a3cd89e1-9962-4fbf-8435-bb351e5ae63a", 00:13:50.554 "assigned_rate_limits": { 00:13:50.554 "rw_ios_per_sec": 0, 00:13:50.554 "rw_mbytes_per_sec": 0, 00:13:50.554 "r_mbytes_per_sec": 0, 00:13:50.554 "w_mbytes_per_sec": 0 00:13:50.554 }, 00:13:50.554 "claimed": false, 00:13:50.554 "zoned": false, 00:13:50.554 "supported_io_types": { 00:13:50.554 "read": true, 00:13:50.554 "write": true, 00:13:50.554 "unmap": false, 00:13:50.554 "flush": false, 00:13:50.554 "reset": true, 00:13:50.554 "nvme_admin": false, 00:13:50.554 "nvme_io": false, 00:13:50.554 "nvme_io_md": false, 00:13:50.554 "write_zeroes": true, 00:13:50.554 "zcopy": false, 00:13:50.554 "get_zone_info": false, 00:13:50.554 "zone_management": false, 00:13:50.554 "zone_append": false, 00:13:50.554 "compare": false, 00:13:50.554 "compare_and_write": false, 00:13:50.554 "abort": false, 00:13:50.554 "seek_hole": false, 00:13:50.554 "seek_data": false, 00:13:50.554 "copy": false, 00:13:50.554 "nvme_iov_md": false 00:13:50.554 }, 00:13:50.554 "driver_specific": { 00:13:50.554 "raid": { 00:13:50.554 "uuid": "a3cd89e1-9962-4fbf-8435-bb351e5ae63a", 00:13:50.554 "strip_size_kb": 64, 00:13:50.554 "state": "online", 00:13:50.554 "raid_level": "raid5f", 00:13:50.554 "superblock": true, 00:13:50.554 "num_base_bdevs": 4, 00:13:50.554 "num_base_bdevs_discovered": 4, 00:13:50.554 "num_base_bdevs_operational": 4, 00:13:50.554 "base_bdevs_list": [ 00:13:50.554 { 00:13:50.554 "name": "NewBaseBdev", 00:13:50.554 "uuid": "c176b910-f12f-4f78-90bd-7dbcbab364cb", 00:13:50.554 "is_configured": true, 00:13:50.554 "data_offset": 2048, 00:13:50.554 "data_size": 63488 00:13:50.554 }, 00:13:50.554 { 00:13:50.554 "name": "BaseBdev2", 00:13:50.554 "uuid": "907b7fa0-7aee-494b-9b09-6d264cbfa894", 00:13:50.554 "is_configured": true, 00:13:50.554 "data_offset": 2048, 00:13:50.554 "data_size": 63488 00:13:50.554 }, 00:13:50.554 { 00:13:50.554 "name": "BaseBdev3", 00:13:50.554 "uuid": "fa2ce605-4511-4edc-b079-6d3eeb395f64", 00:13:50.554 "is_configured": true, 00:13:50.554 "data_offset": 2048, 00:13:50.554 "data_size": 63488 00:13:50.554 }, 00:13:50.554 { 00:13:50.554 "name": "BaseBdev4", 00:13:50.554 "uuid": "36a4ce93-eac5-4d5b-8659-364179f27ebf", 00:13:50.554 "is_configured": true, 00:13:50.554 "data_offset": 2048, 00:13:50.554 "data_size": 63488 00:13:50.554 } 00:13:50.554 ] 00:13:50.554 } 00:13:50.554 } 00:13:50.554 }' 00:13:50.554 23:47:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:13:50.554 23:47:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:13:50.554 BaseBdev2 00:13:50.554 BaseBdev3 00:13:50.554 BaseBdev4' 00:13:50.554 23:47:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:50.554 23:47:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:13:50.554 23:47:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:50.554 23:47:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:13:50.554 23:47:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:50.554 23:47:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:50.554 23:47:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:50.554 23:47:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:50.554 23:47:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:50.554 23:47:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:50.554 23:47:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:50.554 23:47:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:13:50.554 23:47:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:50.554 23:47:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:50.554 23:47:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:50.554 23:47:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:50.554 23:47:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:50.554 23:47:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:50.554 23:47:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:50.554 23:47:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:13:50.554 23:47:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:50.554 23:47:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:50.554 23:47:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:50.554 23:47:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:50.554 23:47:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:50.555 23:47:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:50.555 23:47:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:50.555 23:47:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:50.555 23:47:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:13:50.555 23:47:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:50.555 23:47:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:50.813 23:47:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:50.813 23:47:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:50.813 23:47:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:50.813 23:47:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:13:50.813 23:47:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:50.813 23:47:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:50.813 [2024-11-26 23:47:38.694752] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:13:50.813 [2024-11-26 23:47:38.694776] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:13:50.813 [2024-11-26 23:47:38.694835] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:13:50.813 [2024-11-26 23:47:38.695073] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:13:50.813 [2024-11-26 23:47:38.695083] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name Existed_Raid, state offline 00:13:50.813 23:47:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:50.813 23:47:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 93564 00:13:50.813 23:47:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@954 -- # '[' -z 93564 ']' 00:13:50.813 23:47:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@958 -- # kill -0 93564 00:13:50.813 23:47:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@959 -- # uname 00:13:50.813 23:47:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:13:50.813 23:47:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 93564 00:13:50.813 killing process with pid 93564 00:13:50.813 23:47:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:13:50.813 23:47:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:13:50.813 23:47:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@972 -- # echo 'killing process with pid 93564' 00:13:50.813 23:47:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@973 -- # kill 93564 00:13:50.813 [2024-11-26 23:47:38.736746] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:13:50.813 23:47:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@978 -- # wait 93564 00:13:50.813 [2024-11-26 23:47:38.775815] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:13:51.073 23:47:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:13:51.073 00:13:51.073 real 0m9.106s 00:13:51.073 user 0m15.593s 00:13:51.073 sys 0m1.925s 00:13:51.073 23:47:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@1130 -- # xtrace_disable 00:13:51.073 ************************************ 00:13:51.073 END TEST raid5f_state_function_test_sb 00:13:51.073 ************************************ 00:13:51.073 23:47:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:51.073 23:47:39 bdev_raid -- bdev/bdev_raid.sh@988 -- # run_test raid5f_superblock_test raid_superblock_test raid5f 4 00:13:51.073 23:47:39 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:13:51.073 23:47:39 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:13:51.073 23:47:39 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:13:51.073 ************************************ 00:13:51.073 START TEST raid5f_superblock_test 00:13:51.073 ************************************ 00:13:51.073 23:47:39 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@1129 -- # raid_superblock_test raid5f 4 00:13:51.073 23:47:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=raid5f 00:13:51.073 23:47:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=4 00:13:51.073 23:47:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:13:51.073 23:47:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:13:51.073 23:47:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:13:51.073 23:47:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:13:51.073 23:47:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:13:51.073 23:47:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:13:51.073 23:47:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:13:51.073 23:47:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:13:51.073 23:47:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:13:51.073 23:47:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:13:51.073 23:47:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:13:51.073 23:47:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' raid5f '!=' raid1 ']' 00:13:51.073 23:47:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@405 -- # strip_size=64 00:13:51.073 23:47:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@406 -- # strip_size_create_arg='-z 64' 00:13:51.073 23:47:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=94212 00:13:51.073 23:47:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:13:51.073 23:47:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 94212 00:13:51.073 23:47:39 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@835 -- # '[' -z 94212 ']' 00:13:51.073 23:47:39 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:13:51.073 23:47:39 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:13:51.073 23:47:39 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:13:51.073 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:13:51.073 23:47:39 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:13:51.073 23:47:39 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:51.073 [2024-11-26 23:47:39.136470] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:13:51.073 [2024-11-26 23:47:39.136676] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid94212 ] 00:13:51.330 [2024-11-26 23:47:39.292809] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:13:51.330 [2024-11-26 23:47:39.316726] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:13:51.330 [2024-11-26 23:47:39.357401] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:13:51.330 [2024-11-26 23:47:39.357515] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:13:51.897 23:47:39 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:13:51.898 23:47:39 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@868 -- # return 0 00:13:51.898 23:47:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:13:51.898 23:47:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:13:51.898 23:47:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:13:51.898 23:47:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:13:51.898 23:47:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:13:51.898 23:47:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:13:51.898 23:47:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:13:51.898 23:47:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:13:51.898 23:47:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:13:51.898 23:47:39 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:51.898 23:47:39 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:51.898 malloc1 00:13:51.898 23:47:39 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:51.898 23:47:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:13:51.898 23:47:39 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:51.898 23:47:39 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:51.898 [2024-11-26 23:47:39.971599] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:13:51.898 [2024-11-26 23:47:39.971704] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:51.898 [2024-11-26 23:47:39.971760] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:13:51.898 [2024-11-26 23:47:39.971798] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:51.898 [2024-11-26 23:47:39.973811] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:51.898 [2024-11-26 23:47:39.973881] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:13:51.898 pt1 00:13:51.898 23:47:39 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:51.898 23:47:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:13:51.898 23:47:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:13:51.898 23:47:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:13:51.898 23:47:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:13:51.898 23:47:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:13:51.898 23:47:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:13:51.898 23:47:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:13:51.898 23:47:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:13:51.898 23:47:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:13:51.898 23:47:39 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:51.898 23:47:39 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:51.898 malloc2 00:13:51.898 23:47:39 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:51.898 23:47:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:13:51.898 23:47:39 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:51.898 23:47:39 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:51.898 [2024-11-26 23:47:40.003902] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:13:51.898 [2024-11-26 23:47:40.003997] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:51.898 [2024-11-26 23:47:40.004034] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:13:51.898 [2024-11-26 23:47:40.004064] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:51.898 [2024-11-26 23:47:40.006082] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:51.898 [2024-11-26 23:47:40.006154] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:13:51.898 pt2 00:13:51.898 23:47:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:51.898 23:47:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:13:51.898 23:47:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:13:51.898 23:47:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc3 00:13:51.898 23:47:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt3 00:13:51.898 23:47:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000003 00:13:51.898 23:47:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:13:51.898 23:47:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:13:51.898 23:47:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:13:51.898 23:47:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc3 00:13:51.898 23:47:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:51.898 23:47:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:51.898 malloc3 00:13:51.898 23:47:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:51.898 23:47:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:13:51.898 23:47:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:51.898 23:47:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:52.157 [2024-11-26 23:47:40.032127] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:13:52.157 [2024-11-26 23:47:40.032222] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:52.157 [2024-11-26 23:47:40.032264] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:13:52.157 [2024-11-26 23:47:40.032296] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:52.157 [2024-11-26 23:47:40.034265] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:52.157 [2024-11-26 23:47:40.034337] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:13:52.157 pt3 00:13:52.157 23:47:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:52.157 23:47:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:13:52.157 23:47:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:13:52.157 23:47:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc4 00:13:52.157 23:47:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt4 00:13:52.157 23:47:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000004 00:13:52.157 23:47:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:13:52.157 23:47:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:13:52.157 23:47:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:13:52.157 23:47:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc4 00:13:52.157 23:47:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:52.157 23:47:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:52.157 malloc4 00:13:52.157 23:47:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:52.157 23:47:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:13:52.157 23:47:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:52.157 23:47:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:52.157 [2024-11-26 23:47:40.076116] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:13:52.157 [2024-11-26 23:47:40.076168] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:52.157 [2024-11-26 23:47:40.076184] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:13:52.157 [2024-11-26 23:47:40.076196] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:52.157 [2024-11-26 23:47:40.078209] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:52.157 [2024-11-26 23:47:40.078248] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:13:52.157 pt4 00:13:52.157 23:47:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:52.157 23:47:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:13:52.157 23:47:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:13:52.157 23:47:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''pt1 pt2 pt3 pt4'\''' -n raid_bdev1 -s 00:13:52.157 23:47:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:52.157 23:47:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:52.157 [2024-11-26 23:47:40.088124] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:13:52.157 [2024-11-26 23:47:40.089844] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:13:52.157 [2024-11-26 23:47:40.089954] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:13:52.157 [2024-11-26 23:47:40.090022] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:13:52.158 [2024-11-26 23:47:40.090174] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:13:52.158 [2024-11-26 23:47:40.090192] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:13:52.158 [2024-11-26 23:47:40.090433] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:13:52.158 [2024-11-26 23:47:40.090872] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:13:52.158 [2024-11-26 23:47:40.090888] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:13:52.158 [2024-11-26 23:47:40.091019] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:52.158 23:47:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:52.158 23:47:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 4 00:13:52.158 23:47:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:52.158 23:47:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:52.158 23:47:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:52.158 23:47:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:52.158 23:47:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:13:52.158 23:47:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:52.158 23:47:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:52.158 23:47:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:52.158 23:47:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:52.158 23:47:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:52.158 23:47:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:52.158 23:47:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:52.158 23:47:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:52.158 23:47:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:52.158 23:47:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:52.158 "name": "raid_bdev1", 00:13:52.158 "uuid": "43d4b963-4cf1-4328-82ae-2ff51a753272", 00:13:52.158 "strip_size_kb": 64, 00:13:52.158 "state": "online", 00:13:52.158 "raid_level": "raid5f", 00:13:52.158 "superblock": true, 00:13:52.158 "num_base_bdevs": 4, 00:13:52.158 "num_base_bdevs_discovered": 4, 00:13:52.158 "num_base_bdevs_operational": 4, 00:13:52.158 "base_bdevs_list": [ 00:13:52.158 { 00:13:52.158 "name": "pt1", 00:13:52.158 "uuid": "00000000-0000-0000-0000-000000000001", 00:13:52.158 "is_configured": true, 00:13:52.158 "data_offset": 2048, 00:13:52.158 "data_size": 63488 00:13:52.158 }, 00:13:52.158 { 00:13:52.158 "name": "pt2", 00:13:52.158 "uuid": "00000000-0000-0000-0000-000000000002", 00:13:52.158 "is_configured": true, 00:13:52.158 "data_offset": 2048, 00:13:52.158 "data_size": 63488 00:13:52.158 }, 00:13:52.158 { 00:13:52.158 "name": "pt3", 00:13:52.158 "uuid": "00000000-0000-0000-0000-000000000003", 00:13:52.158 "is_configured": true, 00:13:52.158 "data_offset": 2048, 00:13:52.158 "data_size": 63488 00:13:52.158 }, 00:13:52.158 { 00:13:52.158 "name": "pt4", 00:13:52.158 "uuid": "00000000-0000-0000-0000-000000000004", 00:13:52.158 "is_configured": true, 00:13:52.158 "data_offset": 2048, 00:13:52.158 "data_size": 63488 00:13:52.158 } 00:13:52.158 ] 00:13:52.158 }' 00:13:52.158 23:47:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:52.158 23:47:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:52.726 23:47:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:13:52.726 23:47:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:13:52.726 23:47:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:13:52.726 23:47:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:13:52.726 23:47:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:13:52.726 23:47:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:13:52.726 23:47:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:13:52.726 23:47:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:13:52.726 23:47:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:52.726 23:47:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:52.726 [2024-11-26 23:47:40.576002] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:13:52.726 23:47:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:52.726 23:47:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:13:52.726 "name": "raid_bdev1", 00:13:52.726 "aliases": [ 00:13:52.726 "43d4b963-4cf1-4328-82ae-2ff51a753272" 00:13:52.726 ], 00:13:52.726 "product_name": "Raid Volume", 00:13:52.726 "block_size": 512, 00:13:52.726 "num_blocks": 190464, 00:13:52.726 "uuid": "43d4b963-4cf1-4328-82ae-2ff51a753272", 00:13:52.726 "assigned_rate_limits": { 00:13:52.726 "rw_ios_per_sec": 0, 00:13:52.726 "rw_mbytes_per_sec": 0, 00:13:52.726 "r_mbytes_per_sec": 0, 00:13:52.726 "w_mbytes_per_sec": 0 00:13:52.726 }, 00:13:52.726 "claimed": false, 00:13:52.726 "zoned": false, 00:13:52.726 "supported_io_types": { 00:13:52.726 "read": true, 00:13:52.726 "write": true, 00:13:52.726 "unmap": false, 00:13:52.726 "flush": false, 00:13:52.726 "reset": true, 00:13:52.726 "nvme_admin": false, 00:13:52.726 "nvme_io": false, 00:13:52.726 "nvme_io_md": false, 00:13:52.726 "write_zeroes": true, 00:13:52.726 "zcopy": false, 00:13:52.726 "get_zone_info": false, 00:13:52.726 "zone_management": false, 00:13:52.726 "zone_append": false, 00:13:52.726 "compare": false, 00:13:52.726 "compare_and_write": false, 00:13:52.726 "abort": false, 00:13:52.726 "seek_hole": false, 00:13:52.726 "seek_data": false, 00:13:52.726 "copy": false, 00:13:52.726 "nvme_iov_md": false 00:13:52.726 }, 00:13:52.726 "driver_specific": { 00:13:52.726 "raid": { 00:13:52.726 "uuid": "43d4b963-4cf1-4328-82ae-2ff51a753272", 00:13:52.726 "strip_size_kb": 64, 00:13:52.726 "state": "online", 00:13:52.726 "raid_level": "raid5f", 00:13:52.726 "superblock": true, 00:13:52.726 "num_base_bdevs": 4, 00:13:52.726 "num_base_bdevs_discovered": 4, 00:13:52.726 "num_base_bdevs_operational": 4, 00:13:52.726 "base_bdevs_list": [ 00:13:52.726 { 00:13:52.726 "name": "pt1", 00:13:52.726 "uuid": "00000000-0000-0000-0000-000000000001", 00:13:52.726 "is_configured": true, 00:13:52.726 "data_offset": 2048, 00:13:52.726 "data_size": 63488 00:13:52.726 }, 00:13:52.726 { 00:13:52.726 "name": "pt2", 00:13:52.726 "uuid": "00000000-0000-0000-0000-000000000002", 00:13:52.726 "is_configured": true, 00:13:52.726 "data_offset": 2048, 00:13:52.726 "data_size": 63488 00:13:52.726 }, 00:13:52.726 { 00:13:52.726 "name": "pt3", 00:13:52.726 "uuid": "00000000-0000-0000-0000-000000000003", 00:13:52.726 "is_configured": true, 00:13:52.726 "data_offset": 2048, 00:13:52.726 "data_size": 63488 00:13:52.726 }, 00:13:52.726 { 00:13:52.726 "name": "pt4", 00:13:52.726 "uuid": "00000000-0000-0000-0000-000000000004", 00:13:52.726 "is_configured": true, 00:13:52.726 "data_offset": 2048, 00:13:52.726 "data_size": 63488 00:13:52.726 } 00:13:52.726 ] 00:13:52.726 } 00:13:52.726 } 00:13:52.726 }' 00:13:52.727 23:47:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:13:52.727 23:47:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:13:52.727 pt2 00:13:52.727 pt3 00:13:52.727 pt4' 00:13:52.727 23:47:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:52.727 23:47:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:13:52.727 23:47:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:52.727 23:47:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:52.727 23:47:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:13:52.727 23:47:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:52.727 23:47:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:52.727 23:47:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:52.727 23:47:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:52.727 23:47:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:52.727 23:47:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:52.727 23:47:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:52.727 23:47:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:13:52.727 23:47:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:52.727 23:47:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:52.727 23:47:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:52.727 23:47:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:52.727 23:47:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:52.727 23:47:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:52.727 23:47:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:52.727 23:47:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:13:52.727 23:47:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:52.727 23:47:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:52.727 23:47:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:52.727 23:47:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:52.727 23:47:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:52.727 23:47:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:52.727 23:47:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt4 00:13:52.727 23:47:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:52.727 23:47:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:52.727 23:47:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:52.986 23:47:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:52.986 23:47:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:52.986 23:47:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:52.986 23:47:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:13:52.986 23:47:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:13:52.986 23:47:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:52.986 23:47:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:52.986 [2024-11-26 23:47:40.907445] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:13:52.986 23:47:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:52.986 23:47:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=43d4b963-4cf1-4328-82ae-2ff51a753272 00:13:52.986 23:47:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z 43d4b963-4cf1-4328-82ae-2ff51a753272 ']' 00:13:52.986 23:47:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:13:52.986 23:47:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:52.986 23:47:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:52.986 [2024-11-26 23:47:40.935230] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:13:52.986 [2024-11-26 23:47:40.935255] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:13:52.986 [2024-11-26 23:47:40.935313] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:13:52.986 [2024-11-26 23:47:40.935402] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:13:52.986 [2024-11-26 23:47:40.935418] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:13:52.986 23:47:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:52.986 23:47:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:13:52.987 23:47:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:52.987 23:47:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:52.987 23:47:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:52.987 23:47:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:52.987 23:47:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:13:52.987 23:47:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:13:52.987 23:47:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:13:52.987 23:47:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:13:52.987 23:47:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:52.987 23:47:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:52.987 23:47:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:52.987 23:47:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:13:52.987 23:47:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:13:52.987 23:47:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:52.987 23:47:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:52.987 23:47:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:52.987 23:47:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:13:52.987 23:47:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt3 00:13:52.987 23:47:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:52.987 23:47:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:52.987 23:47:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:52.987 23:47:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:13:52.987 23:47:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt4 00:13:52.987 23:47:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:52.987 23:47:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:52.987 23:47:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:52.987 23:47:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:13:52.987 23:47:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:13:52.987 23:47:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:52.987 23:47:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:52.987 23:47:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:52.987 23:47:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:13:52.987 23:47:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:13:52.987 23:47:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@652 -- # local es=0 00:13:52.987 23:47:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:13:52.987 23:47:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:13:52.987 23:47:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:13:52.987 23:47:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:13:52.987 23:47:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:13:52.987 23:47:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:13:52.987 23:47:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:52.987 23:47:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:52.987 [2024-11-26 23:47:41.083003] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:13:52.987 [2024-11-26 23:47:41.084816] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:13:52.987 [2024-11-26 23:47:41.084860] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc3 is claimed 00:13:52.987 [2024-11-26 23:47:41.084887] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc4 is claimed 00:13:52.987 [2024-11-26 23:47:41.084926] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:13:52.987 [2024-11-26 23:47:41.084975] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:13:52.987 [2024-11-26 23:47:41.084994] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc3 00:13:52.987 [2024-11-26 23:47:41.085009] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc4 00:13:52.987 [2024-11-26 23:47:41.085022] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:13:52.987 [2024-11-26 23:47:41.085031] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state configuring 00:13:52.987 request: 00:13:52.987 { 00:13:52.987 "name": "raid_bdev1", 00:13:52.987 "raid_level": "raid5f", 00:13:52.987 "base_bdevs": [ 00:13:52.987 "malloc1", 00:13:52.987 "malloc2", 00:13:52.987 "malloc3", 00:13:52.987 "malloc4" 00:13:52.987 ], 00:13:52.987 "strip_size_kb": 64, 00:13:52.987 "superblock": false, 00:13:52.987 "method": "bdev_raid_create", 00:13:52.987 "req_id": 1 00:13:52.987 } 00:13:52.987 Got JSON-RPC error response 00:13:52.987 response: 00:13:52.987 { 00:13:52.987 "code": -17, 00:13:52.987 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:13:52.987 } 00:13:52.987 23:47:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:13:52.987 23:47:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@655 -- # es=1 00:13:52.987 23:47:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:13:52.987 23:47:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:13:52.987 23:47:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:13:52.987 23:47:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:52.987 23:47:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:52.987 23:47:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:52.987 23:47:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:13:52.987 23:47:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:53.246 23:47:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:13:53.246 23:47:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:13:53.246 23:47:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:13:53.246 23:47:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:53.246 23:47:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:53.246 [2024-11-26 23:47:41.146868] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:13:53.246 [2024-11-26 23:47:41.146960] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:53.246 [2024-11-26 23:47:41.147000] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:13:53.246 [2024-11-26 23:47:41.147028] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:53.246 [2024-11-26 23:47:41.149054] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:53.246 [2024-11-26 23:47:41.149123] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:13:53.246 [2024-11-26 23:47:41.149212] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:13:53.246 [2024-11-26 23:47:41.149262] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:13:53.246 pt1 00:13:53.246 23:47:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:53.246 23:47:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid5f 64 4 00:13:53.246 23:47:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:53.246 23:47:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:53.246 23:47:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:53.246 23:47:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:53.246 23:47:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:13:53.246 23:47:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:53.246 23:47:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:53.246 23:47:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:53.246 23:47:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:53.246 23:47:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:53.246 23:47:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:53.246 23:47:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:53.246 23:47:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:53.246 23:47:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:53.246 23:47:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:53.246 "name": "raid_bdev1", 00:13:53.246 "uuid": "43d4b963-4cf1-4328-82ae-2ff51a753272", 00:13:53.246 "strip_size_kb": 64, 00:13:53.246 "state": "configuring", 00:13:53.246 "raid_level": "raid5f", 00:13:53.246 "superblock": true, 00:13:53.246 "num_base_bdevs": 4, 00:13:53.246 "num_base_bdevs_discovered": 1, 00:13:53.246 "num_base_bdevs_operational": 4, 00:13:53.246 "base_bdevs_list": [ 00:13:53.246 { 00:13:53.246 "name": "pt1", 00:13:53.246 "uuid": "00000000-0000-0000-0000-000000000001", 00:13:53.246 "is_configured": true, 00:13:53.246 "data_offset": 2048, 00:13:53.246 "data_size": 63488 00:13:53.246 }, 00:13:53.246 { 00:13:53.246 "name": null, 00:13:53.246 "uuid": "00000000-0000-0000-0000-000000000002", 00:13:53.246 "is_configured": false, 00:13:53.247 "data_offset": 2048, 00:13:53.247 "data_size": 63488 00:13:53.247 }, 00:13:53.247 { 00:13:53.247 "name": null, 00:13:53.247 "uuid": "00000000-0000-0000-0000-000000000003", 00:13:53.247 "is_configured": false, 00:13:53.247 "data_offset": 2048, 00:13:53.247 "data_size": 63488 00:13:53.247 }, 00:13:53.247 { 00:13:53.247 "name": null, 00:13:53.247 "uuid": "00000000-0000-0000-0000-000000000004", 00:13:53.247 "is_configured": false, 00:13:53.247 "data_offset": 2048, 00:13:53.247 "data_size": 63488 00:13:53.247 } 00:13:53.247 ] 00:13:53.247 }' 00:13:53.247 23:47:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:53.247 23:47:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:53.506 23:47:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 4 -gt 2 ']' 00:13:53.506 23:47:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@472 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:13:53.506 23:47:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:53.506 23:47:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:53.506 [2024-11-26 23:47:41.586106] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:13:53.506 [2024-11-26 23:47:41.586158] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:53.506 [2024-11-26 23:47:41.586177] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009c80 00:13:53.506 [2024-11-26 23:47:41.586185] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:53.506 [2024-11-26 23:47:41.586542] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:53.506 [2024-11-26 23:47:41.586558] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:13:53.506 [2024-11-26 23:47:41.586616] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:13:53.506 [2024-11-26 23:47:41.586635] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:13:53.506 pt2 00:13:53.506 23:47:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:53.506 23:47:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@473 -- # rpc_cmd bdev_passthru_delete pt2 00:13:53.506 23:47:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:53.506 23:47:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:53.506 [2024-11-26 23:47:41.598119] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt2 00:13:53.506 23:47:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:53.506 23:47:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@474 -- # verify_raid_bdev_state raid_bdev1 configuring raid5f 64 4 00:13:53.506 23:47:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:53.506 23:47:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:53.506 23:47:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:53.506 23:47:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:53.506 23:47:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:13:53.506 23:47:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:53.506 23:47:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:53.506 23:47:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:53.506 23:47:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:53.506 23:47:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:53.506 23:47:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:53.506 23:47:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:53.506 23:47:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:53.506 23:47:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:53.765 23:47:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:53.765 "name": "raid_bdev1", 00:13:53.765 "uuid": "43d4b963-4cf1-4328-82ae-2ff51a753272", 00:13:53.765 "strip_size_kb": 64, 00:13:53.765 "state": "configuring", 00:13:53.765 "raid_level": "raid5f", 00:13:53.765 "superblock": true, 00:13:53.765 "num_base_bdevs": 4, 00:13:53.765 "num_base_bdevs_discovered": 1, 00:13:53.765 "num_base_bdevs_operational": 4, 00:13:53.765 "base_bdevs_list": [ 00:13:53.765 { 00:13:53.765 "name": "pt1", 00:13:53.765 "uuid": "00000000-0000-0000-0000-000000000001", 00:13:53.765 "is_configured": true, 00:13:53.765 "data_offset": 2048, 00:13:53.765 "data_size": 63488 00:13:53.765 }, 00:13:53.765 { 00:13:53.765 "name": null, 00:13:53.765 "uuid": "00000000-0000-0000-0000-000000000002", 00:13:53.765 "is_configured": false, 00:13:53.765 "data_offset": 0, 00:13:53.765 "data_size": 63488 00:13:53.765 }, 00:13:53.765 { 00:13:53.765 "name": null, 00:13:53.765 "uuid": "00000000-0000-0000-0000-000000000003", 00:13:53.765 "is_configured": false, 00:13:53.765 "data_offset": 2048, 00:13:53.765 "data_size": 63488 00:13:53.765 }, 00:13:53.765 { 00:13:53.765 "name": null, 00:13:53.765 "uuid": "00000000-0000-0000-0000-000000000004", 00:13:53.765 "is_configured": false, 00:13:53.765 "data_offset": 2048, 00:13:53.765 "data_size": 63488 00:13:53.765 } 00:13:53.765 ] 00:13:53.765 }' 00:13:53.765 23:47:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:53.765 23:47:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:54.025 23:47:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:13:54.025 23:47:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:13:54.025 23:47:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:13:54.025 23:47:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:54.025 23:47:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:54.025 [2024-11-26 23:47:42.025421] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:13:54.025 [2024-11-26 23:47:42.025512] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:54.025 [2024-11-26 23:47:42.025542] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009f80 00:13:54.025 [2024-11-26 23:47:42.025570] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:54.025 [2024-11-26 23:47:42.025901] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:54.025 [2024-11-26 23:47:42.025956] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:13:54.025 [2024-11-26 23:47:42.026034] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:13:54.025 [2024-11-26 23:47:42.026088] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:13:54.025 pt2 00:13:54.025 23:47:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:54.025 23:47:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:13:54.025 23:47:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:13:54.025 23:47:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:13:54.025 23:47:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:54.025 23:47:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:54.025 [2024-11-26 23:47:42.037377] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:13:54.025 [2024-11-26 23:47:42.037452] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:54.025 [2024-11-26 23:47:42.037482] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:13:54.025 [2024-11-26 23:47:42.037509] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:54.025 [2024-11-26 23:47:42.037833] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:54.025 [2024-11-26 23:47:42.037885] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:13:54.025 [2024-11-26 23:47:42.037957] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:13:54.025 [2024-11-26 23:47:42.038002] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:13:54.025 pt3 00:13:54.025 23:47:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:54.025 23:47:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:13:54.025 23:47:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:13:54.025 23:47:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:13:54.025 23:47:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:54.025 23:47:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:54.025 [2024-11-26 23:47:42.049373] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:13:54.025 [2024-11-26 23:47:42.049452] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:54.025 [2024-11-26 23:47:42.049479] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a580 00:13:54.025 [2024-11-26 23:47:42.049506] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:54.025 [2024-11-26 23:47:42.049789] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:54.025 [2024-11-26 23:47:42.049856] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:13:54.025 [2024-11-26 23:47:42.049928] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt4 00:13:54.025 [2024-11-26 23:47:42.049976] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:13:54.025 [2024-11-26 23:47:42.050086] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:13:54.025 [2024-11-26 23:47:42.050125] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:13:54.025 [2024-11-26 23:47:42.050361] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:13:54.025 [2024-11-26 23:47:42.050840] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:13:54.025 [2024-11-26 23:47:42.050886] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:13:54.025 [2024-11-26 23:47:42.051008] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:54.025 pt4 00:13:54.025 23:47:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:54.025 23:47:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:13:54.025 23:47:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:13:54.025 23:47:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 4 00:13:54.025 23:47:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:54.025 23:47:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:54.025 23:47:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:54.025 23:47:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:54.025 23:47:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:13:54.025 23:47:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:54.025 23:47:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:54.025 23:47:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:54.025 23:47:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:54.025 23:47:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:54.025 23:47:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:54.025 23:47:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:54.025 23:47:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:54.026 23:47:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:54.026 23:47:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:54.026 "name": "raid_bdev1", 00:13:54.026 "uuid": "43d4b963-4cf1-4328-82ae-2ff51a753272", 00:13:54.026 "strip_size_kb": 64, 00:13:54.026 "state": "online", 00:13:54.026 "raid_level": "raid5f", 00:13:54.026 "superblock": true, 00:13:54.026 "num_base_bdevs": 4, 00:13:54.026 "num_base_bdevs_discovered": 4, 00:13:54.026 "num_base_bdevs_operational": 4, 00:13:54.026 "base_bdevs_list": [ 00:13:54.026 { 00:13:54.026 "name": "pt1", 00:13:54.026 "uuid": "00000000-0000-0000-0000-000000000001", 00:13:54.026 "is_configured": true, 00:13:54.026 "data_offset": 2048, 00:13:54.026 "data_size": 63488 00:13:54.026 }, 00:13:54.026 { 00:13:54.026 "name": "pt2", 00:13:54.026 "uuid": "00000000-0000-0000-0000-000000000002", 00:13:54.026 "is_configured": true, 00:13:54.026 "data_offset": 2048, 00:13:54.026 "data_size": 63488 00:13:54.026 }, 00:13:54.026 { 00:13:54.026 "name": "pt3", 00:13:54.026 "uuid": "00000000-0000-0000-0000-000000000003", 00:13:54.026 "is_configured": true, 00:13:54.026 "data_offset": 2048, 00:13:54.026 "data_size": 63488 00:13:54.026 }, 00:13:54.026 { 00:13:54.026 "name": "pt4", 00:13:54.026 "uuid": "00000000-0000-0000-0000-000000000004", 00:13:54.026 "is_configured": true, 00:13:54.026 "data_offset": 2048, 00:13:54.026 "data_size": 63488 00:13:54.026 } 00:13:54.026 ] 00:13:54.026 }' 00:13:54.026 23:47:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:54.026 23:47:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:54.595 23:47:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:13:54.595 23:47:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:13:54.595 23:47:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:13:54.595 23:47:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:13:54.595 23:47:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:13:54.595 23:47:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:13:54.595 23:47:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:13:54.595 23:47:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:13:54.595 23:47:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:54.595 23:47:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:54.595 [2024-11-26 23:47:42.492791] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:13:54.595 23:47:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:54.595 23:47:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:13:54.595 "name": "raid_bdev1", 00:13:54.595 "aliases": [ 00:13:54.595 "43d4b963-4cf1-4328-82ae-2ff51a753272" 00:13:54.595 ], 00:13:54.595 "product_name": "Raid Volume", 00:13:54.595 "block_size": 512, 00:13:54.595 "num_blocks": 190464, 00:13:54.595 "uuid": "43d4b963-4cf1-4328-82ae-2ff51a753272", 00:13:54.595 "assigned_rate_limits": { 00:13:54.595 "rw_ios_per_sec": 0, 00:13:54.595 "rw_mbytes_per_sec": 0, 00:13:54.595 "r_mbytes_per_sec": 0, 00:13:54.595 "w_mbytes_per_sec": 0 00:13:54.595 }, 00:13:54.595 "claimed": false, 00:13:54.595 "zoned": false, 00:13:54.595 "supported_io_types": { 00:13:54.595 "read": true, 00:13:54.595 "write": true, 00:13:54.595 "unmap": false, 00:13:54.595 "flush": false, 00:13:54.595 "reset": true, 00:13:54.595 "nvme_admin": false, 00:13:54.595 "nvme_io": false, 00:13:54.595 "nvme_io_md": false, 00:13:54.595 "write_zeroes": true, 00:13:54.595 "zcopy": false, 00:13:54.595 "get_zone_info": false, 00:13:54.595 "zone_management": false, 00:13:54.595 "zone_append": false, 00:13:54.595 "compare": false, 00:13:54.595 "compare_and_write": false, 00:13:54.595 "abort": false, 00:13:54.595 "seek_hole": false, 00:13:54.595 "seek_data": false, 00:13:54.595 "copy": false, 00:13:54.595 "nvme_iov_md": false 00:13:54.595 }, 00:13:54.595 "driver_specific": { 00:13:54.595 "raid": { 00:13:54.595 "uuid": "43d4b963-4cf1-4328-82ae-2ff51a753272", 00:13:54.595 "strip_size_kb": 64, 00:13:54.595 "state": "online", 00:13:54.595 "raid_level": "raid5f", 00:13:54.595 "superblock": true, 00:13:54.595 "num_base_bdevs": 4, 00:13:54.595 "num_base_bdevs_discovered": 4, 00:13:54.595 "num_base_bdevs_operational": 4, 00:13:54.595 "base_bdevs_list": [ 00:13:54.595 { 00:13:54.595 "name": "pt1", 00:13:54.595 "uuid": "00000000-0000-0000-0000-000000000001", 00:13:54.595 "is_configured": true, 00:13:54.595 "data_offset": 2048, 00:13:54.595 "data_size": 63488 00:13:54.595 }, 00:13:54.595 { 00:13:54.595 "name": "pt2", 00:13:54.595 "uuid": "00000000-0000-0000-0000-000000000002", 00:13:54.595 "is_configured": true, 00:13:54.595 "data_offset": 2048, 00:13:54.595 "data_size": 63488 00:13:54.595 }, 00:13:54.595 { 00:13:54.595 "name": "pt3", 00:13:54.595 "uuid": "00000000-0000-0000-0000-000000000003", 00:13:54.595 "is_configured": true, 00:13:54.595 "data_offset": 2048, 00:13:54.595 "data_size": 63488 00:13:54.595 }, 00:13:54.595 { 00:13:54.595 "name": "pt4", 00:13:54.595 "uuid": "00000000-0000-0000-0000-000000000004", 00:13:54.595 "is_configured": true, 00:13:54.595 "data_offset": 2048, 00:13:54.595 "data_size": 63488 00:13:54.595 } 00:13:54.595 ] 00:13:54.595 } 00:13:54.595 } 00:13:54.595 }' 00:13:54.595 23:47:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:13:54.595 23:47:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:13:54.595 pt2 00:13:54.595 pt3 00:13:54.595 pt4' 00:13:54.595 23:47:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:54.595 23:47:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:13:54.595 23:47:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:54.595 23:47:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:13:54.595 23:47:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:54.595 23:47:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:54.595 23:47:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:54.595 23:47:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:54.595 23:47:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:54.595 23:47:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:54.595 23:47:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:54.595 23:47:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:54.595 23:47:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:13:54.595 23:47:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:54.595 23:47:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:54.595 23:47:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:54.595 23:47:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:54.595 23:47:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:54.595 23:47:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:54.595 23:47:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:13:54.595 23:47:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:54.595 23:47:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:54.595 23:47:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:54.855 23:47:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:54.855 23:47:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:54.855 23:47:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:54.855 23:47:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:54.855 23:47:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt4 00:13:54.855 23:47:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:54.855 23:47:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:54.855 23:47:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:54.855 23:47:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:54.855 23:47:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:54.855 23:47:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:54.855 23:47:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:13:54.855 23:47:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:54.855 23:47:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:54.855 23:47:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:13:54.855 [2024-11-26 23:47:42.796233] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:13:54.855 23:47:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:54.855 23:47:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' 43d4b963-4cf1-4328-82ae-2ff51a753272 '!=' 43d4b963-4cf1-4328-82ae-2ff51a753272 ']' 00:13:54.855 23:47:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy raid5f 00:13:54.855 23:47:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:13:54.855 23:47:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@199 -- # return 0 00:13:54.855 23:47:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@493 -- # rpc_cmd bdev_passthru_delete pt1 00:13:54.855 23:47:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:54.855 23:47:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:54.855 [2024-11-26 23:47:42.844028] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt1 00:13:54.855 23:47:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:54.855 23:47:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@496 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:13:54.855 23:47:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:54.855 23:47:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:54.855 23:47:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:54.855 23:47:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:54.855 23:47:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:54.855 23:47:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:54.855 23:47:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:54.855 23:47:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:54.855 23:47:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:54.855 23:47:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:54.855 23:47:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:54.855 23:47:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:54.855 23:47:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:54.855 23:47:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:54.855 23:47:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:54.855 "name": "raid_bdev1", 00:13:54.855 "uuid": "43d4b963-4cf1-4328-82ae-2ff51a753272", 00:13:54.855 "strip_size_kb": 64, 00:13:54.855 "state": "online", 00:13:54.855 "raid_level": "raid5f", 00:13:54.855 "superblock": true, 00:13:54.855 "num_base_bdevs": 4, 00:13:54.855 "num_base_bdevs_discovered": 3, 00:13:54.855 "num_base_bdevs_operational": 3, 00:13:54.855 "base_bdevs_list": [ 00:13:54.856 { 00:13:54.856 "name": null, 00:13:54.856 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:54.856 "is_configured": false, 00:13:54.856 "data_offset": 0, 00:13:54.856 "data_size": 63488 00:13:54.856 }, 00:13:54.856 { 00:13:54.856 "name": "pt2", 00:13:54.856 "uuid": "00000000-0000-0000-0000-000000000002", 00:13:54.856 "is_configured": true, 00:13:54.856 "data_offset": 2048, 00:13:54.856 "data_size": 63488 00:13:54.856 }, 00:13:54.856 { 00:13:54.856 "name": "pt3", 00:13:54.856 "uuid": "00000000-0000-0000-0000-000000000003", 00:13:54.856 "is_configured": true, 00:13:54.856 "data_offset": 2048, 00:13:54.856 "data_size": 63488 00:13:54.856 }, 00:13:54.856 { 00:13:54.856 "name": "pt4", 00:13:54.856 "uuid": "00000000-0000-0000-0000-000000000004", 00:13:54.856 "is_configured": true, 00:13:54.856 "data_offset": 2048, 00:13:54.856 "data_size": 63488 00:13:54.856 } 00:13:54.856 ] 00:13:54.856 }' 00:13:54.856 23:47:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:54.856 23:47:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:55.115 23:47:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@499 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:13:55.115 23:47:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:55.115 23:47:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:55.115 [2024-11-26 23:47:43.235300] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:13:55.115 [2024-11-26 23:47:43.235384] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:13:55.115 [2024-11-26 23:47:43.235467] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:13:55.115 [2024-11-26 23:47:43.235549] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:13:55.115 [2024-11-26 23:47:43.235596] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:13:55.115 23:47:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:55.115 23:47:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@500 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:55.115 23:47:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:55.115 23:47:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:55.374 23:47:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@500 -- # jq -r '.[]' 00:13:55.374 23:47:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:55.374 23:47:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@500 -- # raid_bdev= 00:13:55.374 23:47:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@501 -- # '[' -n '' ']' 00:13:55.374 23:47:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i = 1 )) 00:13:55.374 23:47:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:13:55.374 23:47:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt2 00:13:55.374 23:47:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:55.374 23:47:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:55.374 23:47:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:55.374 23:47:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:13:55.374 23:47:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:13:55.374 23:47:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt3 00:13:55.374 23:47:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:55.374 23:47:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:55.374 23:47:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:55.374 23:47:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:13:55.374 23:47:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:13:55.374 23:47:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt4 00:13:55.374 23:47:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:55.374 23:47:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:55.374 23:47:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:55.374 23:47:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:13:55.374 23:47:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:13:55.374 23:47:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i = 1 )) 00:13:55.374 23:47:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:13:55.374 23:47:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@512 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:13:55.374 23:47:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:55.374 23:47:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:55.374 [2024-11-26 23:47:43.335131] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:13:55.374 [2024-11-26 23:47:43.335184] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:55.374 [2024-11-26 23:47:43.335198] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a880 00:13:55.374 [2024-11-26 23:47:43.335207] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:55.374 [2024-11-26 23:47:43.337200] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:55.374 [2024-11-26 23:47:43.337239] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:13:55.374 [2024-11-26 23:47:43.337302] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:13:55.374 [2024-11-26 23:47:43.337336] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:13:55.374 pt2 00:13:55.374 23:47:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:55.374 23:47:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@515 -- # verify_raid_bdev_state raid_bdev1 configuring raid5f 64 3 00:13:55.374 23:47:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:55.374 23:47:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:55.374 23:47:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:55.374 23:47:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:55.374 23:47:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:55.374 23:47:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:55.374 23:47:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:55.374 23:47:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:55.374 23:47:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:55.374 23:47:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:55.374 23:47:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:55.374 23:47:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:55.374 23:47:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:55.374 23:47:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:55.374 23:47:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:55.374 "name": "raid_bdev1", 00:13:55.374 "uuid": "43d4b963-4cf1-4328-82ae-2ff51a753272", 00:13:55.374 "strip_size_kb": 64, 00:13:55.374 "state": "configuring", 00:13:55.374 "raid_level": "raid5f", 00:13:55.374 "superblock": true, 00:13:55.374 "num_base_bdevs": 4, 00:13:55.374 "num_base_bdevs_discovered": 1, 00:13:55.374 "num_base_bdevs_operational": 3, 00:13:55.374 "base_bdevs_list": [ 00:13:55.374 { 00:13:55.374 "name": null, 00:13:55.374 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:55.374 "is_configured": false, 00:13:55.374 "data_offset": 2048, 00:13:55.374 "data_size": 63488 00:13:55.374 }, 00:13:55.374 { 00:13:55.374 "name": "pt2", 00:13:55.374 "uuid": "00000000-0000-0000-0000-000000000002", 00:13:55.374 "is_configured": true, 00:13:55.374 "data_offset": 2048, 00:13:55.374 "data_size": 63488 00:13:55.374 }, 00:13:55.374 { 00:13:55.374 "name": null, 00:13:55.374 "uuid": "00000000-0000-0000-0000-000000000003", 00:13:55.374 "is_configured": false, 00:13:55.374 "data_offset": 2048, 00:13:55.374 "data_size": 63488 00:13:55.374 }, 00:13:55.374 { 00:13:55.374 "name": null, 00:13:55.374 "uuid": "00000000-0000-0000-0000-000000000004", 00:13:55.374 "is_configured": false, 00:13:55.374 "data_offset": 2048, 00:13:55.374 "data_size": 63488 00:13:55.374 } 00:13:55.374 ] 00:13:55.374 }' 00:13:55.374 23:47:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:55.374 23:47:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:55.634 23:47:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i++ )) 00:13:55.634 23:47:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:13:55.634 23:47:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@512 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:13:55.634 23:47:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:55.634 23:47:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:55.634 [2024-11-26 23:47:43.750556] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:13:55.634 [2024-11-26 23:47:43.750650] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:55.634 [2024-11-26 23:47:43.750684] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ae80 00:13:55.634 [2024-11-26 23:47:43.750715] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:55.634 [2024-11-26 23:47:43.751058] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:55.634 [2024-11-26 23:47:43.751114] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:13:55.634 [2024-11-26 23:47:43.751203] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:13:55.634 [2024-11-26 23:47:43.751257] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:13:55.634 pt3 00:13:55.634 23:47:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:55.634 23:47:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@515 -- # verify_raid_bdev_state raid_bdev1 configuring raid5f 64 3 00:13:55.634 23:47:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:55.634 23:47:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:55.634 23:47:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:55.634 23:47:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:55.634 23:47:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:55.634 23:47:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:55.634 23:47:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:55.634 23:47:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:55.634 23:47:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:55.634 23:47:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:55.634 23:47:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:55.893 23:47:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:55.893 23:47:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:55.893 23:47:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:55.893 23:47:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:55.893 "name": "raid_bdev1", 00:13:55.893 "uuid": "43d4b963-4cf1-4328-82ae-2ff51a753272", 00:13:55.893 "strip_size_kb": 64, 00:13:55.893 "state": "configuring", 00:13:55.893 "raid_level": "raid5f", 00:13:55.893 "superblock": true, 00:13:55.893 "num_base_bdevs": 4, 00:13:55.893 "num_base_bdevs_discovered": 2, 00:13:55.893 "num_base_bdevs_operational": 3, 00:13:55.893 "base_bdevs_list": [ 00:13:55.893 { 00:13:55.893 "name": null, 00:13:55.893 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:55.893 "is_configured": false, 00:13:55.893 "data_offset": 2048, 00:13:55.893 "data_size": 63488 00:13:55.893 }, 00:13:55.893 { 00:13:55.893 "name": "pt2", 00:13:55.893 "uuid": "00000000-0000-0000-0000-000000000002", 00:13:55.893 "is_configured": true, 00:13:55.893 "data_offset": 2048, 00:13:55.893 "data_size": 63488 00:13:55.893 }, 00:13:55.893 { 00:13:55.893 "name": "pt3", 00:13:55.893 "uuid": "00000000-0000-0000-0000-000000000003", 00:13:55.893 "is_configured": true, 00:13:55.893 "data_offset": 2048, 00:13:55.893 "data_size": 63488 00:13:55.893 }, 00:13:55.893 { 00:13:55.893 "name": null, 00:13:55.893 "uuid": "00000000-0000-0000-0000-000000000004", 00:13:55.893 "is_configured": false, 00:13:55.893 "data_offset": 2048, 00:13:55.893 "data_size": 63488 00:13:55.893 } 00:13:55.893 ] 00:13:55.893 }' 00:13:55.893 23:47:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:55.893 23:47:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:56.152 23:47:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i++ )) 00:13:56.152 23:47:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:13:56.153 23:47:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@519 -- # i=3 00:13:56.153 23:47:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@520 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:13:56.153 23:47:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:56.153 23:47:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:56.153 [2024-11-26 23:47:44.185751] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:13:56.153 [2024-11-26 23:47:44.185802] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:56.153 [2024-11-26 23:47:44.185818] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b180 00:13:56.153 [2024-11-26 23:47:44.185828] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:56.153 [2024-11-26 23:47:44.186139] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:56.153 [2024-11-26 23:47:44.186157] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:13:56.153 [2024-11-26 23:47:44.186211] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt4 00:13:56.153 [2024-11-26 23:47:44.186230] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:13:56.153 [2024-11-26 23:47:44.186313] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:13:56.153 [2024-11-26 23:47:44.186323] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:13:56.153 [2024-11-26 23:47:44.186564] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000026d0 00:13:56.153 [2024-11-26 23:47:44.187069] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:13:56.153 [2024-11-26 23:47:44.187090] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001c80 00:13:56.153 [2024-11-26 23:47:44.187278] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:56.153 pt4 00:13:56.153 23:47:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:56.153 23:47:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@523 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:13:56.153 23:47:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:56.153 23:47:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:56.153 23:47:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:56.153 23:47:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:56.153 23:47:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:56.153 23:47:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:56.153 23:47:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:56.153 23:47:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:56.153 23:47:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:56.153 23:47:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:56.153 23:47:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:56.153 23:47:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:56.153 23:47:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:56.153 23:47:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:56.153 23:47:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:56.153 "name": "raid_bdev1", 00:13:56.153 "uuid": "43d4b963-4cf1-4328-82ae-2ff51a753272", 00:13:56.153 "strip_size_kb": 64, 00:13:56.153 "state": "online", 00:13:56.153 "raid_level": "raid5f", 00:13:56.153 "superblock": true, 00:13:56.153 "num_base_bdevs": 4, 00:13:56.153 "num_base_bdevs_discovered": 3, 00:13:56.153 "num_base_bdevs_operational": 3, 00:13:56.153 "base_bdevs_list": [ 00:13:56.153 { 00:13:56.153 "name": null, 00:13:56.153 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:56.153 "is_configured": false, 00:13:56.153 "data_offset": 2048, 00:13:56.153 "data_size": 63488 00:13:56.153 }, 00:13:56.153 { 00:13:56.153 "name": "pt2", 00:13:56.153 "uuid": "00000000-0000-0000-0000-000000000002", 00:13:56.153 "is_configured": true, 00:13:56.153 "data_offset": 2048, 00:13:56.153 "data_size": 63488 00:13:56.153 }, 00:13:56.153 { 00:13:56.153 "name": "pt3", 00:13:56.153 "uuid": "00000000-0000-0000-0000-000000000003", 00:13:56.153 "is_configured": true, 00:13:56.153 "data_offset": 2048, 00:13:56.153 "data_size": 63488 00:13:56.153 }, 00:13:56.153 { 00:13:56.153 "name": "pt4", 00:13:56.153 "uuid": "00000000-0000-0000-0000-000000000004", 00:13:56.153 "is_configured": true, 00:13:56.153 "data_offset": 2048, 00:13:56.153 "data_size": 63488 00:13:56.153 } 00:13:56.153 ] 00:13:56.153 }' 00:13:56.153 23:47:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:56.153 23:47:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:56.723 23:47:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@526 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:13:56.723 23:47:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:56.723 23:47:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:56.723 [2024-11-26 23:47:44.656980] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:13:56.723 [2024-11-26 23:47:44.657009] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:13:56.723 [2024-11-26 23:47:44.657070] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:13:56.723 [2024-11-26 23:47:44.657137] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:13:56.723 [2024-11-26 23:47:44.657146] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name raid_bdev1, state offline 00:13:56.723 23:47:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:56.723 23:47:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@527 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:56.723 23:47:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@527 -- # jq -r '.[]' 00:13:56.723 23:47:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:56.723 23:47:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:56.723 23:47:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:56.723 23:47:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@527 -- # raid_bdev= 00:13:56.723 23:47:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@528 -- # '[' -n '' ']' 00:13:56.723 23:47:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@532 -- # '[' 4 -gt 2 ']' 00:13:56.723 23:47:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@534 -- # i=3 00:13:56.723 23:47:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@535 -- # rpc_cmd bdev_passthru_delete pt4 00:13:56.723 23:47:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:56.723 23:47:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:56.723 23:47:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:56.723 23:47:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@540 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:13:56.723 23:47:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:56.723 23:47:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:56.723 [2024-11-26 23:47:44.732851] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:13:56.723 [2024-11-26 23:47:44.732906] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:56.723 [2024-11-26 23:47:44.732926] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b480 00:13:56.723 [2024-11-26 23:47:44.732933] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:56.723 [2024-11-26 23:47:44.735006] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:56.723 [2024-11-26 23:47:44.735042] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:13:56.723 [2024-11-26 23:47:44.735105] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:13:56.723 [2024-11-26 23:47:44.735136] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:13:56.723 [2024-11-26 23:47:44.735221] bdev_raid.c:3685:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev pt2 (4) greater than existing raid bdev raid_bdev1 (2) 00:13:56.723 [2024-11-26 23:47:44.735232] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:13:56.723 [2024-11-26 23:47:44.735256] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002000 name raid_bdev1, state configuring 00:13:56.723 [2024-11-26 23:47:44.735287] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:13:56.723 [2024-11-26 23:47:44.735403] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:13:56.723 pt1 00:13:56.723 23:47:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:56.723 23:47:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@542 -- # '[' 4 -gt 2 ']' 00:13:56.723 23:47:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@545 -- # verify_raid_bdev_state raid_bdev1 configuring raid5f 64 3 00:13:56.723 23:47:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:56.723 23:47:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:56.723 23:47:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:56.723 23:47:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:56.723 23:47:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:56.723 23:47:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:56.723 23:47:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:56.723 23:47:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:56.723 23:47:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:56.723 23:47:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:56.723 23:47:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:56.723 23:47:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:56.723 23:47:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:56.723 23:47:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:56.723 23:47:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:56.723 "name": "raid_bdev1", 00:13:56.723 "uuid": "43d4b963-4cf1-4328-82ae-2ff51a753272", 00:13:56.723 "strip_size_kb": 64, 00:13:56.723 "state": "configuring", 00:13:56.723 "raid_level": "raid5f", 00:13:56.723 "superblock": true, 00:13:56.723 "num_base_bdevs": 4, 00:13:56.723 "num_base_bdevs_discovered": 2, 00:13:56.723 "num_base_bdevs_operational": 3, 00:13:56.723 "base_bdevs_list": [ 00:13:56.723 { 00:13:56.723 "name": null, 00:13:56.723 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:56.723 "is_configured": false, 00:13:56.723 "data_offset": 2048, 00:13:56.723 "data_size": 63488 00:13:56.723 }, 00:13:56.723 { 00:13:56.723 "name": "pt2", 00:13:56.723 "uuid": "00000000-0000-0000-0000-000000000002", 00:13:56.723 "is_configured": true, 00:13:56.723 "data_offset": 2048, 00:13:56.723 "data_size": 63488 00:13:56.723 }, 00:13:56.723 { 00:13:56.723 "name": "pt3", 00:13:56.723 "uuid": "00000000-0000-0000-0000-000000000003", 00:13:56.723 "is_configured": true, 00:13:56.723 "data_offset": 2048, 00:13:56.723 "data_size": 63488 00:13:56.723 }, 00:13:56.723 { 00:13:56.723 "name": null, 00:13:56.723 "uuid": "00000000-0000-0000-0000-000000000004", 00:13:56.723 "is_configured": false, 00:13:56.723 "data_offset": 2048, 00:13:56.724 "data_size": 63488 00:13:56.724 } 00:13:56.724 ] 00:13:56.724 }' 00:13:56.724 23:47:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:56.724 23:47:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:57.294 23:47:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@546 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:13:57.294 23:47:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@546 -- # rpc_cmd bdev_raid_get_bdevs configuring 00:13:57.294 23:47:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:57.294 23:47:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:57.294 23:47:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:57.294 23:47:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@546 -- # [[ false == \f\a\l\s\e ]] 00:13:57.294 23:47:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@549 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:13:57.294 23:47:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:57.294 23:47:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:57.294 [2024-11-26 23:47:45.208040] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:13:57.294 [2024-11-26 23:47:45.208142] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:57.294 [2024-11-26 23:47:45.208183] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ba80 00:13:57.294 [2024-11-26 23:47:45.208214] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:57.294 [2024-11-26 23:47:45.208589] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:57.294 [2024-11-26 23:47:45.208647] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:13:57.294 [2024-11-26 23:47:45.208732] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt4 00:13:57.294 [2024-11-26 23:47:45.208784] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:13:57.294 [2024-11-26 23:47:45.208899] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000002380 00:13:57.294 [2024-11-26 23:47:45.208943] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:13:57.294 [2024-11-26 23:47:45.209177] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000027a0 00:13:57.294 [2024-11-26 23:47:45.209758] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000002380 00:13:57.295 [2024-11-26 23:47:45.209814] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000002380 00:13:57.295 [2024-11-26 23:47:45.210027] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:57.295 pt4 00:13:57.295 23:47:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:57.295 23:47:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@554 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:13:57.295 23:47:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:57.295 23:47:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:57.295 23:47:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:57.295 23:47:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:57.295 23:47:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:57.295 23:47:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:57.295 23:47:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:57.295 23:47:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:57.295 23:47:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:57.295 23:47:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:57.295 23:47:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:57.295 23:47:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:57.295 23:47:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:57.295 23:47:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:57.295 23:47:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:57.295 "name": "raid_bdev1", 00:13:57.295 "uuid": "43d4b963-4cf1-4328-82ae-2ff51a753272", 00:13:57.295 "strip_size_kb": 64, 00:13:57.295 "state": "online", 00:13:57.295 "raid_level": "raid5f", 00:13:57.295 "superblock": true, 00:13:57.295 "num_base_bdevs": 4, 00:13:57.295 "num_base_bdevs_discovered": 3, 00:13:57.295 "num_base_bdevs_operational": 3, 00:13:57.295 "base_bdevs_list": [ 00:13:57.295 { 00:13:57.295 "name": null, 00:13:57.295 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:57.295 "is_configured": false, 00:13:57.295 "data_offset": 2048, 00:13:57.295 "data_size": 63488 00:13:57.295 }, 00:13:57.295 { 00:13:57.295 "name": "pt2", 00:13:57.295 "uuid": "00000000-0000-0000-0000-000000000002", 00:13:57.295 "is_configured": true, 00:13:57.295 "data_offset": 2048, 00:13:57.295 "data_size": 63488 00:13:57.295 }, 00:13:57.295 { 00:13:57.295 "name": "pt3", 00:13:57.295 "uuid": "00000000-0000-0000-0000-000000000003", 00:13:57.295 "is_configured": true, 00:13:57.295 "data_offset": 2048, 00:13:57.295 "data_size": 63488 00:13:57.295 }, 00:13:57.295 { 00:13:57.295 "name": "pt4", 00:13:57.295 "uuid": "00000000-0000-0000-0000-000000000004", 00:13:57.295 "is_configured": true, 00:13:57.295 "data_offset": 2048, 00:13:57.295 "data_size": 63488 00:13:57.295 } 00:13:57.295 ] 00:13:57.295 }' 00:13:57.295 23:47:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:57.295 23:47:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:57.555 23:47:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@555 -- # rpc_cmd bdev_raid_get_bdevs online 00:13:57.555 23:47:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:57.555 23:47:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:57.813 23:47:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@555 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:13:57.813 23:47:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:57.813 23:47:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@555 -- # [[ false == \f\a\l\s\e ]] 00:13:57.813 23:47:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@558 -- # jq -r '.[] | .uuid' 00:13:57.813 23:47:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@558 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:13:57.813 23:47:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:57.813 23:47:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:57.813 [2024-11-26 23:47:45.743289] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:13:57.813 23:47:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:57.813 23:47:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@558 -- # '[' 43d4b963-4cf1-4328-82ae-2ff51a753272 '!=' 43d4b963-4cf1-4328-82ae-2ff51a753272 ']' 00:13:57.813 23:47:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 94212 00:13:57.813 23:47:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@954 -- # '[' -z 94212 ']' 00:13:57.813 23:47:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@958 -- # kill -0 94212 00:13:57.813 23:47:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@959 -- # uname 00:13:57.813 23:47:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:13:57.813 23:47:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 94212 00:13:57.813 23:47:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:13:57.813 23:47:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:13:57.813 23:47:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 94212' 00:13:57.813 killing process with pid 94212 00:13:57.813 23:47:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@973 -- # kill 94212 00:13:57.813 [2024-11-26 23:47:45.824712] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:13:57.813 [2024-11-26 23:47:45.824787] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:13:57.813 [2024-11-26 23:47:45.824856] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:13:57.813 [2024-11-26 23:47:45.824865] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002380 name raid_bdev1, state offline 00:13:57.813 23:47:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@978 -- # wait 94212 00:13:57.813 [2024-11-26 23:47:45.866752] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:13:58.071 23:47:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:13:58.071 00:13:58.071 real 0m7.029s 00:13:58.071 user 0m11.891s 00:13:58.071 sys 0m1.444s 00:13:58.071 23:47:46 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:13:58.071 ************************************ 00:13:58.071 END TEST raid5f_superblock_test 00:13:58.071 ************************************ 00:13:58.071 23:47:46 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:58.071 23:47:46 bdev_raid -- bdev/bdev_raid.sh@989 -- # '[' true = true ']' 00:13:58.071 23:47:46 bdev_raid -- bdev/bdev_raid.sh@990 -- # run_test raid5f_rebuild_test raid_rebuild_test raid5f 4 false false true 00:13:58.071 23:47:46 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:13:58.071 23:47:46 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:13:58.071 23:47:46 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:13:58.071 ************************************ 00:13:58.071 START TEST raid5f_rebuild_test 00:13:58.071 ************************************ 00:13:58.071 23:47:46 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@1129 -- # raid_rebuild_test raid5f 4 false false true 00:13:58.071 23:47:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@569 -- # local raid_level=raid5f 00:13:58.072 23:47:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=4 00:13:58.072 23:47:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@571 -- # local superblock=false 00:13:58.072 23:47:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:13:58.072 23:47:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@573 -- # local verify=true 00:13:58.072 23:47:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:13:58.072 23:47:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:13:58.072 23:47:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:13:58.072 23:47:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:13:58.072 23:47:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:13:58.072 23:47:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:13:58.072 23:47:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:13:58.072 23:47:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:13:58.072 23:47:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev3 00:13:58.072 23:47:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:13:58.072 23:47:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:13:58.072 23:47:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev4 00:13:58.072 23:47:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:13:58.072 23:47:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:13:58.072 23:47:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:13:58.072 23:47:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:13:58.072 23:47:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:13:58.072 23:47:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@576 -- # local strip_size 00:13:58.072 23:47:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@577 -- # local create_arg 00:13:58.072 23:47:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:13:58.072 23:47:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@579 -- # local data_offset 00:13:58.072 23:47:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@581 -- # '[' raid5f '!=' raid1 ']' 00:13:58.072 23:47:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@582 -- # '[' false = true ']' 00:13:58.072 23:47:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@586 -- # strip_size=64 00:13:58.072 23:47:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@587 -- # create_arg+=' -z 64' 00:13:58.072 23:47:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@592 -- # '[' false = true ']' 00:13:58.072 23:47:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@597 -- # raid_pid=94682 00:13:58.072 23:47:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:13:58.072 23:47:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@598 -- # waitforlisten 94682 00:13:58.072 23:47:46 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@835 -- # '[' -z 94682 ']' 00:13:58.072 23:47:46 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:13:58.072 23:47:46 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:13:58.072 23:47:46 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:13:58.072 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:13:58.072 23:47:46 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:13:58.072 23:47:46 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:58.331 [2024-11-26 23:47:46.246055] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:13:58.331 [2024-11-26 23:47:46.246250] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.ealI/O size of 3145728 is greater than zero copy threshold (65536). 00:13:58.331 Zero copy mechanism will not be used. 00:13:58.331 :6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid94682 ] 00:13:58.331 [2024-11-26 23:47:46.398585] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:13:58.331 [2024-11-26 23:47:46.422639] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:13:58.591 [2024-11-26 23:47:46.463832] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:13:58.591 [2024-11-26 23:47:46.463936] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:13:59.180 23:47:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:13:59.180 23:47:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@868 -- # return 0 00:13:59.180 23:47:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:13:59.180 23:47:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:13:59.180 23:47:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:59.180 23:47:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:59.180 BaseBdev1_malloc 00:13:59.180 23:47:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:59.180 23:47:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:13:59.180 23:47:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:59.180 23:47:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:59.180 [2024-11-26 23:47:47.082207] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:13:59.180 [2024-11-26 23:47:47.082262] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:59.180 [2024-11-26 23:47:47.082307] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:13:59.180 [2024-11-26 23:47:47.082325] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:59.180 [2024-11-26 23:47:47.084304] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:59.180 [2024-11-26 23:47:47.084353] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:13:59.180 BaseBdev1 00:13:59.180 23:47:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:59.180 23:47:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:13:59.180 23:47:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:13:59.180 23:47:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:59.180 23:47:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:59.180 BaseBdev2_malloc 00:13:59.180 23:47:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:59.180 23:47:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:13:59.180 23:47:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:59.180 23:47:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:59.180 [2024-11-26 23:47:47.110434] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:13:59.180 [2024-11-26 23:47:47.110485] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:59.180 [2024-11-26 23:47:47.110515] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:13:59.180 [2024-11-26 23:47:47.110524] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:59.180 [2024-11-26 23:47:47.112511] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:59.180 [2024-11-26 23:47:47.112549] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:13:59.180 BaseBdev2 00:13:59.180 23:47:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:59.180 23:47:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:13:59.180 23:47:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:13:59.180 23:47:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:59.180 23:47:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:59.180 BaseBdev3_malloc 00:13:59.180 23:47:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:59.180 23:47:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev3_malloc -p BaseBdev3 00:13:59.180 23:47:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:59.180 23:47:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:59.180 [2024-11-26 23:47:47.138778] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev3_malloc 00:13:59.180 [2024-11-26 23:47:47.138828] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:59.180 [2024-11-26 23:47:47.138865] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:13:59.180 [2024-11-26 23:47:47.138873] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:59.180 [2024-11-26 23:47:47.140893] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:59.180 [2024-11-26 23:47:47.140928] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:13:59.180 BaseBdev3 00:13:59.180 23:47:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:59.180 23:47:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:13:59.180 23:47:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:13:59.180 23:47:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:59.180 23:47:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:59.180 BaseBdev4_malloc 00:13:59.180 23:47:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:59.180 23:47:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev4_malloc -p BaseBdev4 00:13:59.180 23:47:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:59.180 23:47:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:59.180 [2024-11-26 23:47:47.184708] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev4_malloc 00:13:59.180 [2024-11-26 23:47:47.184859] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:59.180 [2024-11-26 23:47:47.184905] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:13:59.180 [2024-11-26 23:47:47.184921] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:59.180 [2024-11-26 23:47:47.188603] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:59.180 [2024-11-26 23:47:47.188663] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:13:59.180 BaseBdev4 00:13:59.180 23:47:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:59.180 23:47:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:13:59.180 23:47:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:59.180 23:47:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:59.180 spare_malloc 00:13:59.181 23:47:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:59.181 23:47:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:13:59.181 23:47:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:59.181 23:47:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:59.181 spare_delay 00:13:59.181 23:47:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:59.181 23:47:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:13:59.181 23:47:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:59.181 23:47:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:59.181 [2024-11-26 23:47:47.225462] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:13:59.181 [2024-11-26 23:47:47.225505] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:59.181 [2024-11-26 23:47:47.225537] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009c80 00:13:59.181 [2024-11-26 23:47:47.225545] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:59.181 [2024-11-26 23:47:47.227498] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:59.181 [2024-11-26 23:47:47.227576] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:13:59.181 spare 00:13:59.181 23:47:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:59.181 23:47:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 00:13:59.181 23:47:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:59.181 23:47:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:59.181 [2024-11-26 23:47:47.237513] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:13:59.181 [2024-11-26 23:47:47.239236] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:13:59.181 [2024-11-26 23:47:47.239296] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:13:59.181 [2024-11-26 23:47:47.239342] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:13:59.181 [2024-11-26 23:47:47.239432] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:13:59.181 [2024-11-26 23:47:47.239441] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 196608, blocklen 512 00:13:59.181 [2024-11-26 23:47:47.239680] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:13:59.181 [2024-11-26 23:47:47.240129] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:13:59.181 [2024-11-26 23:47:47.240142] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:13:59.181 [2024-11-26 23:47:47.240249] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:59.181 23:47:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:59.181 23:47:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 4 00:13:59.181 23:47:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:59.181 23:47:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:59.181 23:47:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:59.181 23:47:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:59.181 23:47:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:13:59.181 23:47:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:59.181 23:47:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:59.181 23:47:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:59.181 23:47:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:59.181 23:47:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:59.181 23:47:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:59.181 23:47:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:59.181 23:47:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:59.181 23:47:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:59.181 23:47:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:59.181 "name": "raid_bdev1", 00:13:59.181 "uuid": "fb5fe94c-b292-46be-acce-5eed89b870f4", 00:13:59.181 "strip_size_kb": 64, 00:13:59.181 "state": "online", 00:13:59.181 "raid_level": "raid5f", 00:13:59.181 "superblock": false, 00:13:59.181 "num_base_bdevs": 4, 00:13:59.181 "num_base_bdevs_discovered": 4, 00:13:59.181 "num_base_bdevs_operational": 4, 00:13:59.181 "base_bdevs_list": [ 00:13:59.181 { 00:13:59.181 "name": "BaseBdev1", 00:13:59.181 "uuid": "154e258c-fae5-5667-86c0-4d54aaa70fe4", 00:13:59.181 "is_configured": true, 00:13:59.181 "data_offset": 0, 00:13:59.181 "data_size": 65536 00:13:59.181 }, 00:13:59.181 { 00:13:59.181 "name": "BaseBdev2", 00:13:59.181 "uuid": "eb03108c-ecda-51e0-9f0e-c54f21f1353d", 00:13:59.181 "is_configured": true, 00:13:59.181 "data_offset": 0, 00:13:59.181 "data_size": 65536 00:13:59.181 }, 00:13:59.181 { 00:13:59.181 "name": "BaseBdev3", 00:13:59.181 "uuid": "0f1041c7-edcf-5738-854d-b694da9829b6", 00:13:59.181 "is_configured": true, 00:13:59.181 "data_offset": 0, 00:13:59.181 "data_size": 65536 00:13:59.181 }, 00:13:59.181 { 00:13:59.181 "name": "BaseBdev4", 00:13:59.181 "uuid": "948dc191-0fba-551a-a044-ef9c4d004931", 00:13:59.181 "is_configured": true, 00:13:59.181 "data_offset": 0, 00:13:59.181 "data_size": 65536 00:13:59.181 } 00:13:59.181 ] 00:13:59.181 }' 00:13:59.181 23:47:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:59.181 23:47:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:59.764 23:47:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:13:59.764 23:47:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:13:59.764 23:47:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:59.764 23:47:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:59.764 [2024-11-26 23:47:47.645405] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:13:59.764 23:47:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:59.764 23:47:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=196608 00:13:59.764 23:47:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:59.764 23:47:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:13:59.764 23:47:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:59.764 23:47:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:59.764 23:47:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:59.764 23:47:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@619 -- # data_offset=0 00:13:59.764 23:47:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:13:59.764 23:47:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:13:59.764 23:47:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:13:59.764 23:47:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:13:59.764 23:47:47 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:13:59.764 23:47:47 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:13:59.764 23:47:47 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@10 -- # local bdev_list 00:13:59.764 23:47:47 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:13:59.764 23:47:47 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@11 -- # local nbd_list 00:13:59.764 23:47:47 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@12 -- # local i 00:13:59.764 23:47:47 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:13:59.764 23:47:47 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:13:59.764 23:47:47 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:14:00.024 [2024-11-26 23:47:47.908817] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000027a0 00:14:00.024 /dev/nbd0 00:14:00.024 23:47:47 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:14:00.024 23:47:47 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:14:00.024 23:47:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:14:00.024 23:47:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@873 -- # local i 00:14:00.024 23:47:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:14:00.024 23:47:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:14:00.024 23:47:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:14:00.024 23:47:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@877 -- # break 00:14:00.024 23:47:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:14:00.024 23:47:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:14:00.024 23:47:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:14:00.024 1+0 records in 00:14:00.024 1+0 records out 00:14:00.024 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000543857 s, 7.5 MB/s 00:14:00.024 23:47:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:00.024 23:47:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@890 -- # size=4096 00:14:00.024 23:47:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:00.024 23:47:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:14:00.024 23:47:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@893 -- # return 0 00:14:00.024 23:47:47 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:14:00.024 23:47:47 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:14:00.024 23:47:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@629 -- # '[' raid5f = raid5f ']' 00:14:00.024 23:47:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@630 -- # write_unit_size=384 00:14:00.024 23:47:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@631 -- # echo 192 00:14:00.024 23:47:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=196608 count=512 oflag=direct 00:14:00.283 512+0 records in 00:14:00.283 512+0 records out 00:14:00.283 100663296 bytes (101 MB, 96 MiB) copied, 0.391126 s, 257 MB/s 00:14:00.283 23:47:48 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:14:00.283 23:47:48 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:14:00.283 23:47:48 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:14:00.283 23:47:48 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@50 -- # local nbd_list 00:14:00.283 23:47:48 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@51 -- # local i 00:14:00.283 23:47:48 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:14:00.283 23:47:48 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:14:00.542 23:47:48 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:14:00.542 23:47:48 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:14:00.542 23:47:48 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:14:00.542 23:47:48 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:14:00.542 23:47:48 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:14:00.542 23:47:48 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:14:00.542 [2024-11-26 23:47:48.581960] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:00.542 23:47:48 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:14:00.542 23:47:48 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:14:00.542 23:47:48 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:14:00.542 23:47:48 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:00.542 23:47:48 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:00.542 [2024-11-26 23:47:48.594014] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:14:00.542 23:47:48 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:00.542 23:47:48 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:14:00.542 23:47:48 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:00.542 23:47:48 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:00.542 23:47:48 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:00.542 23:47:48 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:00.542 23:47:48 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:00.542 23:47:48 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:00.542 23:47:48 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:00.542 23:47:48 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:00.542 23:47:48 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:00.542 23:47:48 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:00.542 23:47:48 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:00.542 23:47:48 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:00.542 23:47:48 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:00.542 23:47:48 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:00.542 23:47:48 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:00.542 "name": "raid_bdev1", 00:14:00.542 "uuid": "fb5fe94c-b292-46be-acce-5eed89b870f4", 00:14:00.542 "strip_size_kb": 64, 00:14:00.542 "state": "online", 00:14:00.542 "raid_level": "raid5f", 00:14:00.542 "superblock": false, 00:14:00.542 "num_base_bdevs": 4, 00:14:00.542 "num_base_bdevs_discovered": 3, 00:14:00.542 "num_base_bdevs_operational": 3, 00:14:00.542 "base_bdevs_list": [ 00:14:00.542 { 00:14:00.542 "name": null, 00:14:00.542 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:00.542 "is_configured": false, 00:14:00.542 "data_offset": 0, 00:14:00.542 "data_size": 65536 00:14:00.542 }, 00:14:00.542 { 00:14:00.542 "name": "BaseBdev2", 00:14:00.542 "uuid": "eb03108c-ecda-51e0-9f0e-c54f21f1353d", 00:14:00.542 "is_configured": true, 00:14:00.542 "data_offset": 0, 00:14:00.542 "data_size": 65536 00:14:00.542 }, 00:14:00.542 { 00:14:00.542 "name": "BaseBdev3", 00:14:00.542 "uuid": "0f1041c7-edcf-5738-854d-b694da9829b6", 00:14:00.542 "is_configured": true, 00:14:00.542 "data_offset": 0, 00:14:00.542 "data_size": 65536 00:14:00.542 }, 00:14:00.542 { 00:14:00.542 "name": "BaseBdev4", 00:14:00.542 "uuid": "948dc191-0fba-551a-a044-ef9c4d004931", 00:14:00.542 "is_configured": true, 00:14:00.542 "data_offset": 0, 00:14:00.542 "data_size": 65536 00:14:00.542 } 00:14:00.542 ] 00:14:00.542 }' 00:14:00.542 23:47:48 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:00.542 23:47:48 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:01.111 23:47:48 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:14:01.111 23:47:49 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:01.111 23:47:49 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:01.111 [2024-11-26 23:47:49.009305] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:14:01.111 [2024-11-26 23:47:49.013377] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000027da0 00:14:01.111 23:47:49 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:01.111 23:47:49 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@647 -- # sleep 1 00:14:01.111 [2024-11-26 23:47:49.015530] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:14:02.050 23:47:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:02.050 23:47:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:02.050 23:47:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:02.050 23:47:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:02.050 23:47:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:02.050 23:47:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:02.050 23:47:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:02.050 23:47:50 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:02.050 23:47:50 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:02.050 23:47:50 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:02.051 23:47:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:02.051 "name": "raid_bdev1", 00:14:02.051 "uuid": "fb5fe94c-b292-46be-acce-5eed89b870f4", 00:14:02.051 "strip_size_kb": 64, 00:14:02.051 "state": "online", 00:14:02.051 "raid_level": "raid5f", 00:14:02.051 "superblock": false, 00:14:02.051 "num_base_bdevs": 4, 00:14:02.051 "num_base_bdevs_discovered": 4, 00:14:02.051 "num_base_bdevs_operational": 4, 00:14:02.051 "process": { 00:14:02.051 "type": "rebuild", 00:14:02.051 "target": "spare", 00:14:02.051 "progress": { 00:14:02.051 "blocks": 19200, 00:14:02.051 "percent": 9 00:14:02.051 } 00:14:02.051 }, 00:14:02.051 "base_bdevs_list": [ 00:14:02.051 { 00:14:02.051 "name": "spare", 00:14:02.051 "uuid": "83b2c087-3dc3-57fd-901b-49219a82421f", 00:14:02.051 "is_configured": true, 00:14:02.051 "data_offset": 0, 00:14:02.051 "data_size": 65536 00:14:02.051 }, 00:14:02.051 { 00:14:02.051 "name": "BaseBdev2", 00:14:02.051 "uuid": "eb03108c-ecda-51e0-9f0e-c54f21f1353d", 00:14:02.051 "is_configured": true, 00:14:02.051 "data_offset": 0, 00:14:02.051 "data_size": 65536 00:14:02.051 }, 00:14:02.051 { 00:14:02.051 "name": "BaseBdev3", 00:14:02.051 "uuid": "0f1041c7-edcf-5738-854d-b694da9829b6", 00:14:02.051 "is_configured": true, 00:14:02.051 "data_offset": 0, 00:14:02.051 "data_size": 65536 00:14:02.051 }, 00:14:02.051 { 00:14:02.051 "name": "BaseBdev4", 00:14:02.051 "uuid": "948dc191-0fba-551a-a044-ef9c4d004931", 00:14:02.051 "is_configured": true, 00:14:02.051 "data_offset": 0, 00:14:02.051 "data_size": 65536 00:14:02.051 } 00:14:02.051 ] 00:14:02.051 }' 00:14:02.051 23:47:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:02.051 23:47:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:02.051 23:47:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:02.051 23:47:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:02.051 23:47:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:14:02.051 23:47:50 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:02.051 23:47:50 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:02.051 [2024-11-26 23:47:50.152172] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:14:02.311 [2024-11-26 23:47:50.220971] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:14:02.311 [2024-11-26 23:47:50.221035] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:02.311 [2024-11-26 23:47:50.221054] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:14:02.311 [2024-11-26 23:47:50.221062] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:14:02.311 23:47:50 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:02.311 23:47:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:14:02.311 23:47:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:02.311 23:47:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:02.311 23:47:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:02.311 23:47:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:02.311 23:47:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:02.311 23:47:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:02.311 23:47:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:02.311 23:47:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:02.311 23:47:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:02.311 23:47:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:02.311 23:47:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:02.311 23:47:50 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:02.311 23:47:50 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:02.311 23:47:50 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:02.311 23:47:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:02.311 "name": "raid_bdev1", 00:14:02.311 "uuid": "fb5fe94c-b292-46be-acce-5eed89b870f4", 00:14:02.311 "strip_size_kb": 64, 00:14:02.311 "state": "online", 00:14:02.311 "raid_level": "raid5f", 00:14:02.311 "superblock": false, 00:14:02.311 "num_base_bdevs": 4, 00:14:02.311 "num_base_bdevs_discovered": 3, 00:14:02.311 "num_base_bdevs_operational": 3, 00:14:02.311 "base_bdevs_list": [ 00:14:02.311 { 00:14:02.311 "name": null, 00:14:02.311 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:02.311 "is_configured": false, 00:14:02.311 "data_offset": 0, 00:14:02.311 "data_size": 65536 00:14:02.311 }, 00:14:02.311 { 00:14:02.311 "name": "BaseBdev2", 00:14:02.311 "uuid": "eb03108c-ecda-51e0-9f0e-c54f21f1353d", 00:14:02.311 "is_configured": true, 00:14:02.311 "data_offset": 0, 00:14:02.311 "data_size": 65536 00:14:02.311 }, 00:14:02.311 { 00:14:02.311 "name": "BaseBdev3", 00:14:02.311 "uuid": "0f1041c7-edcf-5738-854d-b694da9829b6", 00:14:02.311 "is_configured": true, 00:14:02.311 "data_offset": 0, 00:14:02.311 "data_size": 65536 00:14:02.311 }, 00:14:02.311 { 00:14:02.311 "name": "BaseBdev4", 00:14:02.311 "uuid": "948dc191-0fba-551a-a044-ef9c4d004931", 00:14:02.311 "is_configured": true, 00:14:02.311 "data_offset": 0, 00:14:02.311 "data_size": 65536 00:14:02.311 } 00:14:02.311 ] 00:14:02.311 }' 00:14:02.311 23:47:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:02.311 23:47:50 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:02.570 23:47:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:14:02.570 23:47:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:02.570 23:47:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:14:02.570 23:47:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=none 00:14:02.570 23:47:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:02.570 23:47:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:02.570 23:47:50 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:02.570 23:47:50 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:02.570 23:47:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:02.570 23:47:50 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:02.830 23:47:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:02.830 "name": "raid_bdev1", 00:14:02.830 "uuid": "fb5fe94c-b292-46be-acce-5eed89b870f4", 00:14:02.830 "strip_size_kb": 64, 00:14:02.830 "state": "online", 00:14:02.830 "raid_level": "raid5f", 00:14:02.830 "superblock": false, 00:14:02.830 "num_base_bdevs": 4, 00:14:02.830 "num_base_bdevs_discovered": 3, 00:14:02.830 "num_base_bdevs_operational": 3, 00:14:02.830 "base_bdevs_list": [ 00:14:02.830 { 00:14:02.830 "name": null, 00:14:02.830 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:02.830 "is_configured": false, 00:14:02.830 "data_offset": 0, 00:14:02.830 "data_size": 65536 00:14:02.830 }, 00:14:02.830 { 00:14:02.830 "name": "BaseBdev2", 00:14:02.830 "uuid": "eb03108c-ecda-51e0-9f0e-c54f21f1353d", 00:14:02.830 "is_configured": true, 00:14:02.830 "data_offset": 0, 00:14:02.830 "data_size": 65536 00:14:02.830 }, 00:14:02.830 { 00:14:02.830 "name": "BaseBdev3", 00:14:02.830 "uuid": "0f1041c7-edcf-5738-854d-b694da9829b6", 00:14:02.830 "is_configured": true, 00:14:02.830 "data_offset": 0, 00:14:02.830 "data_size": 65536 00:14:02.830 }, 00:14:02.830 { 00:14:02.830 "name": "BaseBdev4", 00:14:02.830 "uuid": "948dc191-0fba-551a-a044-ef9c4d004931", 00:14:02.830 "is_configured": true, 00:14:02.830 "data_offset": 0, 00:14:02.830 "data_size": 65536 00:14:02.830 } 00:14:02.830 ] 00:14:02.830 }' 00:14:02.830 23:47:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:02.830 23:47:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:14:02.830 23:47:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:02.830 23:47:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:14:02.830 23:47:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:14:02.830 23:47:50 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:02.830 23:47:50 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:02.830 [2024-11-26 23:47:50.809777] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:14:02.830 [2024-11-26 23:47:50.813312] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000027e70 00:14:02.830 [2024-11-26 23:47:50.815362] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:14:02.830 23:47:50 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:02.830 23:47:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@663 -- # sleep 1 00:14:03.770 23:47:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:03.770 23:47:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:03.770 23:47:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:03.770 23:47:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:03.770 23:47:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:03.770 23:47:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:03.770 23:47:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:03.770 23:47:51 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:03.770 23:47:51 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:03.770 23:47:51 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:03.770 23:47:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:03.770 "name": "raid_bdev1", 00:14:03.770 "uuid": "fb5fe94c-b292-46be-acce-5eed89b870f4", 00:14:03.770 "strip_size_kb": 64, 00:14:03.770 "state": "online", 00:14:03.770 "raid_level": "raid5f", 00:14:03.770 "superblock": false, 00:14:03.770 "num_base_bdevs": 4, 00:14:03.770 "num_base_bdevs_discovered": 4, 00:14:03.770 "num_base_bdevs_operational": 4, 00:14:03.770 "process": { 00:14:03.770 "type": "rebuild", 00:14:03.770 "target": "spare", 00:14:03.770 "progress": { 00:14:03.770 "blocks": 19200, 00:14:03.770 "percent": 9 00:14:03.770 } 00:14:03.770 }, 00:14:03.770 "base_bdevs_list": [ 00:14:03.770 { 00:14:03.770 "name": "spare", 00:14:03.770 "uuid": "83b2c087-3dc3-57fd-901b-49219a82421f", 00:14:03.770 "is_configured": true, 00:14:03.770 "data_offset": 0, 00:14:03.770 "data_size": 65536 00:14:03.770 }, 00:14:03.770 { 00:14:03.770 "name": "BaseBdev2", 00:14:03.770 "uuid": "eb03108c-ecda-51e0-9f0e-c54f21f1353d", 00:14:03.770 "is_configured": true, 00:14:03.770 "data_offset": 0, 00:14:03.770 "data_size": 65536 00:14:03.770 }, 00:14:03.770 { 00:14:03.770 "name": "BaseBdev3", 00:14:03.770 "uuid": "0f1041c7-edcf-5738-854d-b694da9829b6", 00:14:03.770 "is_configured": true, 00:14:03.770 "data_offset": 0, 00:14:03.770 "data_size": 65536 00:14:03.770 }, 00:14:03.770 { 00:14:03.770 "name": "BaseBdev4", 00:14:03.770 "uuid": "948dc191-0fba-551a-a044-ef9c4d004931", 00:14:03.770 "is_configured": true, 00:14:03.770 "data_offset": 0, 00:14:03.770 "data_size": 65536 00:14:03.770 } 00:14:03.770 ] 00:14:03.770 }' 00:14:03.770 23:47:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:04.030 23:47:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:04.030 23:47:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:04.030 23:47:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:04.030 23:47:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@666 -- # '[' false = true ']' 00:14:04.030 23:47:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=4 00:14:04.030 23:47:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@693 -- # '[' raid5f = raid1 ']' 00:14:04.030 23:47:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@706 -- # local timeout=503 00:14:04.030 23:47:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:04.030 23:47:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:04.030 23:47:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:04.030 23:47:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:04.030 23:47:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:04.030 23:47:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:04.030 23:47:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:04.030 23:47:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:04.030 23:47:51 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:04.030 23:47:51 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:04.030 23:47:51 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:04.030 23:47:52 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:04.030 "name": "raid_bdev1", 00:14:04.030 "uuid": "fb5fe94c-b292-46be-acce-5eed89b870f4", 00:14:04.030 "strip_size_kb": 64, 00:14:04.030 "state": "online", 00:14:04.030 "raid_level": "raid5f", 00:14:04.030 "superblock": false, 00:14:04.030 "num_base_bdevs": 4, 00:14:04.030 "num_base_bdevs_discovered": 4, 00:14:04.030 "num_base_bdevs_operational": 4, 00:14:04.030 "process": { 00:14:04.030 "type": "rebuild", 00:14:04.030 "target": "spare", 00:14:04.030 "progress": { 00:14:04.030 "blocks": 21120, 00:14:04.030 "percent": 10 00:14:04.030 } 00:14:04.030 }, 00:14:04.030 "base_bdevs_list": [ 00:14:04.030 { 00:14:04.030 "name": "spare", 00:14:04.030 "uuid": "83b2c087-3dc3-57fd-901b-49219a82421f", 00:14:04.030 "is_configured": true, 00:14:04.030 "data_offset": 0, 00:14:04.030 "data_size": 65536 00:14:04.030 }, 00:14:04.030 { 00:14:04.030 "name": "BaseBdev2", 00:14:04.030 "uuid": "eb03108c-ecda-51e0-9f0e-c54f21f1353d", 00:14:04.030 "is_configured": true, 00:14:04.030 "data_offset": 0, 00:14:04.030 "data_size": 65536 00:14:04.030 }, 00:14:04.030 { 00:14:04.030 "name": "BaseBdev3", 00:14:04.030 "uuid": "0f1041c7-edcf-5738-854d-b694da9829b6", 00:14:04.030 "is_configured": true, 00:14:04.030 "data_offset": 0, 00:14:04.030 "data_size": 65536 00:14:04.030 }, 00:14:04.030 { 00:14:04.030 "name": "BaseBdev4", 00:14:04.030 "uuid": "948dc191-0fba-551a-a044-ef9c4d004931", 00:14:04.030 "is_configured": true, 00:14:04.030 "data_offset": 0, 00:14:04.030 "data_size": 65536 00:14:04.030 } 00:14:04.030 ] 00:14:04.030 }' 00:14:04.030 23:47:52 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:04.030 23:47:52 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:04.030 23:47:52 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:04.030 23:47:52 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:04.030 23:47:52 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:14:04.968 23:47:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:04.968 23:47:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:04.968 23:47:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:04.968 23:47:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:04.968 23:47:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:04.969 23:47:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:04.969 23:47:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:05.229 23:47:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:05.229 23:47:53 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:05.229 23:47:53 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:05.229 23:47:53 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:05.229 23:47:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:05.229 "name": "raid_bdev1", 00:14:05.229 "uuid": "fb5fe94c-b292-46be-acce-5eed89b870f4", 00:14:05.229 "strip_size_kb": 64, 00:14:05.229 "state": "online", 00:14:05.229 "raid_level": "raid5f", 00:14:05.229 "superblock": false, 00:14:05.229 "num_base_bdevs": 4, 00:14:05.229 "num_base_bdevs_discovered": 4, 00:14:05.229 "num_base_bdevs_operational": 4, 00:14:05.229 "process": { 00:14:05.229 "type": "rebuild", 00:14:05.229 "target": "spare", 00:14:05.229 "progress": { 00:14:05.229 "blocks": 42240, 00:14:05.229 "percent": 21 00:14:05.229 } 00:14:05.229 }, 00:14:05.229 "base_bdevs_list": [ 00:14:05.229 { 00:14:05.229 "name": "spare", 00:14:05.229 "uuid": "83b2c087-3dc3-57fd-901b-49219a82421f", 00:14:05.229 "is_configured": true, 00:14:05.229 "data_offset": 0, 00:14:05.229 "data_size": 65536 00:14:05.229 }, 00:14:05.229 { 00:14:05.229 "name": "BaseBdev2", 00:14:05.229 "uuid": "eb03108c-ecda-51e0-9f0e-c54f21f1353d", 00:14:05.229 "is_configured": true, 00:14:05.229 "data_offset": 0, 00:14:05.229 "data_size": 65536 00:14:05.229 }, 00:14:05.229 { 00:14:05.229 "name": "BaseBdev3", 00:14:05.229 "uuid": "0f1041c7-edcf-5738-854d-b694da9829b6", 00:14:05.229 "is_configured": true, 00:14:05.229 "data_offset": 0, 00:14:05.229 "data_size": 65536 00:14:05.229 }, 00:14:05.229 { 00:14:05.229 "name": "BaseBdev4", 00:14:05.229 "uuid": "948dc191-0fba-551a-a044-ef9c4d004931", 00:14:05.229 "is_configured": true, 00:14:05.229 "data_offset": 0, 00:14:05.229 "data_size": 65536 00:14:05.229 } 00:14:05.229 ] 00:14:05.229 }' 00:14:05.229 23:47:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:05.229 23:47:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:05.229 23:47:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:05.229 23:47:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:05.229 23:47:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:14:06.169 23:47:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:06.169 23:47:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:06.169 23:47:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:06.169 23:47:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:06.169 23:47:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:06.169 23:47:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:06.169 23:47:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:06.169 23:47:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:06.169 23:47:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:06.169 23:47:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:06.169 23:47:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:06.169 23:47:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:06.169 "name": "raid_bdev1", 00:14:06.169 "uuid": "fb5fe94c-b292-46be-acce-5eed89b870f4", 00:14:06.169 "strip_size_kb": 64, 00:14:06.169 "state": "online", 00:14:06.169 "raid_level": "raid5f", 00:14:06.169 "superblock": false, 00:14:06.169 "num_base_bdevs": 4, 00:14:06.169 "num_base_bdevs_discovered": 4, 00:14:06.169 "num_base_bdevs_operational": 4, 00:14:06.169 "process": { 00:14:06.169 "type": "rebuild", 00:14:06.169 "target": "spare", 00:14:06.169 "progress": { 00:14:06.169 "blocks": 65280, 00:14:06.169 "percent": 33 00:14:06.169 } 00:14:06.169 }, 00:14:06.169 "base_bdevs_list": [ 00:14:06.169 { 00:14:06.169 "name": "spare", 00:14:06.169 "uuid": "83b2c087-3dc3-57fd-901b-49219a82421f", 00:14:06.169 "is_configured": true, 00:14:06.169 "data_offset": 0, 00:14:06.169 "data_size": 65536 00:14:06.169 }, 00:14:06.169 { 00:14:06.169 "name": "BaseBdev2", 00:14:06.169 "uuid": "eb03108c-ecda-51e0-9f0e-c54f21f1353d", 00:14:06.169 "is_configured": true, 00:14:06.169 "data_offset": 0, 00:14:06.169 "data_size": 65536 00:14:06.169 }, 00:14:06.169 { 00:14:06.169 "name": "BaseBdev3", 00:14:06.169 "uuid": "0f1041c7-edcf-5738-854d-b694da9829b6", 00:14:06.169 "is_configured": true, 00:14:06.169 "data_offset": 0, 00:14:06.169 "data_size": 65536 00:14:06.169 }, 00:14:06.169 { 00:14:06.169 "name": "BaseBdev4", 00:14:06.169 "uuid": "948dc191-0fba-551a-a044-ef9c4d004931", 00:14:06.169 "is_configured": true, 00:14:06.169 "data_offset": 0, 00:14:06.169 "data_size": 65536 00:14:06.169 } 00:14:06.169 ] 00:14:06.169 }' 00:14:06.169 23:47:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:06.429 23:47:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:06.429 23:47:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:06.429 23:47:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:06.429 23:47:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:14:07.369 23:47:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:07.369 23:47:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:07.369 23:47:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:07.369 23:47:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:07.369 23:47:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:07.370 23:47:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:07.370 23:47:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:07.370 23:47:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:07.370 23:47:55 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:07.370 23:47:55 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:07.370 23:47:55 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:07.370 23:47:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:07.370 "name": "raid_bdev1", 00:14:07.370 "uuid": "fb5fe94c-b292-46be-acce-5eed89b870f4", 00:14:07.370 "strip_size_kb": 64, 00:14:07.370 "state": "online", 00:14:07.370 "raid_level": "raid5f", 00:14:07.370 "superblock": false, 00:14:07.370 "num_base_bdevs": 4, 00:14:07.370 "num_base_bdevs_discovered": 4, 00:14:07.370 "num_base_bdevs_operational": 4, 00:14:07.370 "process": { 00:14:07.370 "type": "rebuild", 00:14:07.370 "target": "spare", 00:14:07.370 "progress": { 00:14:07.370 "blocks": 86400, 00:14:07.370 "percent": 43 00:14:07.370 } 00:14:07.370 }, 00:14:07.370 "base_bdevs_list": [ 00:14:07.370 { 00:14:07.370 "name": "spare", 00:14:07.370 "uuid": "83b2c087-3dc3-57fd-901b-49219a82421f", 00:14:07.370 "is_configured": true, 00:14:07.370 "data_offset": 0, 00:14:07.370 "data_size": 65536 00:14:07.370 }, 00:14:07.370 { 00:14:07.370 "name": "BaseBdev2", 00:14:07.370 "uuid": "eb03108c-ecda-51e0-9f0e-c54f21f1353d", 00:14:07.370 "is_configured": true, 00:14:07.370 "data_offset": 0, 00:14:07.370 "data_size": 65536 00:14:07.370 }, 00:14:07.370 { 00:14:07.370 "name": "BaseBdev3", 00:14:07.370 "uuid": "0f1041c7-edcf-5738-854d-b694da9829b6", 00:14:07.370 "is_configured": true, 00:14:07.370 "data_offset": 0, 00:14:07.370 "data_size": 65536 00:14:07.370 }, 00:14:07.370 { 00:14:07.370 "name": "BaseBdev4", 00:14:07.370 "uuid": "948dc191-0fba-551a-a044-ef9c4d004931", 00:14:07.370 "is_configured": true, 00:14:07.370 "data_offset": 0, 00:14:07.370 "data_size": 65536 00:14:07.370 } 00:14:07.370 ] 00:14:07.370 }' 00:14:07.370 23:47:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:07.370 23:47:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:07.370 23:47:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:07.370 23:47:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:07.370 23:47:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:14:08.749 23:47:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:08.749 23:47:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:08.749 23:47:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:08.749 23:47:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:08.749 23:47:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:08.749 23:47:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:08.749 23:47:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:08.749 23:47:56 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:08.749 23:47:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:08.749 23:47:56 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:08.749 23:47:56 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:08.749 23:47:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:08.749 "name": "raid_bdev1", 00:14:08.749 "uuid": "fb5fe94c-b292-46be-acce-5eed89b870f4", 00:14:08.749 "strip_size_kb": 64, 00:14:08.749 "state": "online", 00:14:08.749 "raid_level": "raid5f", 00:14:08.749 "superblock": false, 00:14:08.749 "num_base_bdevs": 4, 00:14:08.749 "num_base_bdevs_discovered": 4, 00:14:08.749 "num_base_bdevs_operational": 4, 00:14:08.749 "process": { 00:14:08.749 "type": "rebuild", 00:14:08.749 "target": "spare", 00:14:08.749 "progress": { 00:14:08.749 "blocks": 107520, 00:14:08.749 "percent": 54 00:14:08.749 } 00:14:08.749 }, 00:14:08.749 "base_bdevs_list": [ 00:14:08.749 { 00:14:08.749 "name": "spare", 00:14:08.749 "uuid": "83b2c087-3dc3-57fd-901b-49219a82421f", 00:14:08.749 "is_configured": true, 00:14:08.749 "data_offset": 0, 00:14:08.749 "data_size": 65536 00:14:08.749 }, 00:14:08.749 { 00:14:08.749 "name": "BaseBdev2", 00:14:08.749 "uuid": "eb03108c-ecda-51e0-9f0e-c54f21f1353d", 00:14:08.749 "is_configured": true, 00:14:08.749 "data_offset": 0, 00:14:08.749 "data_size": 65536 00:14:08.749 }, 00:14:08.749 { 00:14:08.749 "name": "BaseBdev3", 00:14:08.749 "uuid": "0f1041c7-edcf-5738-854d-b694da9829b6", 00:14:08.749 "is_configured": true, 00:14:08.749 "data_offset": 0, 00:14:08.749 "data_size": 65536 00:14:08.749 }, 00:14:08.749 { 00:14:08.749 "name": "BaseBdev4", 00:14:08.749 "uuid": "948dc191-0fba-551a-a044-ef9c4d004931", 00:14:08.749 "is_configured": true, 00:14:08.749 "data_offset": 0, 00:14:08.749 "data_size": 65536 00:14:08.749 } 00:14:08.749 ] 00:14:08.749 }' 00:14:08.749 23:47:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:08.749 23:47:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:08.749 23:47:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:08.749 23:47:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:08.749 23:47:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:14:09.686 23:47:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:09.686 23:47:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:09.686 23:47:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:09.686 23:47:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:09.686 23:47:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:09.686 23:47:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:09.686 23:47:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:09.686 23:47:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:09.686 23:47:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:09.686 23:47:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:09.686 23:47:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:09.686 23:47:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:09.686 "name": "raid_bdev1", 00:14:09.686 "uuid": "fb5fe94c-b292-46be-acce-5eed89b870f4", 00:14:09.686 "strip_size_kb": 64, 00:14:09.686 "state": "online", 00:14:09.686 "raid_level": "raid5f", 00:14:09.686 "superblock": false, 00:14:09.686 "num_base_bdevs": 4, 00:14:09.686 "num_base_bdevs_discovered": 4, 00:14:09.686 "num_base_bdevs_operational": 4, 00:14:09.686 "process": { 00:14:09.686 "type": "rebuild", 00:14:09.686 "target": "spare", 00:14:09.686 "progress": { 00:14:09.686 "blocks": 130560, 00:14:09.686 "percent": 66 00:14:09.686 } 00:14:09.686 }, 00:14:09.686 "base_bdevs_list": [ 00:14:09.686 { 00:14:09.686 "name": "spare", 00:14:09.686 "uuid": "83b2c087-3dc3-57fd-901b-49219a82421f", 00:14:09.686 "is_configured": true, 00:14:09.686 "data_offset": 0, 00:14:09.686 "data_size": 65536 00:14:09.686 }, 00:14:09.686 { 00:14:09.686 "name": "BaseBdev2", 00:14:09.686 "uuid": "eb03108c-ecda-51e0-9f0e-c54f21f1353d", 00:14:09.686 "is_configured": true, 00:14:09.686 "data_offset": 0, 00:14:09.686 "data_size": 65536 00:14:09.686 }, 00:14:09.686 { 00:14:09.686 "name": "BaseBdev3", 00:14:09.686 "uuid": "0f1041c7-edcf-5738-854d-b694da9829b6", 00:14:09.686 "is_configured": true, 00:14:09.686 "data_offset": 0, 00:14:09.686 "data_size": 65536 00:14:09.686 }, 00:14:09.686 { 00:14:09.686 "name": "BaseBdev4", 00:14:09.686 "uuid": "948dc191-0fba-551a-a044-ef9c4d004931", 00:14:09.686 "is_configured": true, 00:14:09.686 "data_offset": 0, 00:14:09.686 "data_size": 65536 00:14:09.686 } 00:14:09.686 ] 00:14:09.686 }' 00:14:09.686 23:47:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:09.687 23:47:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:09.687 23:47:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:09.946 23:47:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:09.946 23:47:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:14:10.885 23:47:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:10.885 23:47:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:10.885 23:47:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:10.885 23:47:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:10.885 23:47:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:10.885 23:47:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:10.885 23:47:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:10.885 23:47:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:10.885 23:47:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:10.885 23:47:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:10.885 23:47:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:10.885 23:47:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:10.885 "name": "raid_bdev1", 00:14:10.885 "uuid": "fb5fe94c-b292-46be-acce-5eed89b870f4", 00:14:10.885 "strip_size_kb": 64, 00:14:10.885 "state": "online", 00:14:10.885 "raid_level": "raid5f", 00:14:10.885 "superblock": false, 00:14:10.885 "num_base_bdevs": 4, 00:14:10.885 "num_base_bdevs_discovered": 4, 00:14:10.885 "num_base_bdevs_operational": 4, 00:14:10.885 "process": { 00:14:10.885 "type": "rebuild", 00:14:10.885 "target": "spare", 00:14:10.885 "progress": { 00:14:10.885 "blocks": 151680, 00:14:10.885 "percent": 77 00:14:10.885 } 00:14:10.885 }, 00:14:10.885 "base_bdevs_list": [ 00:14:10.885 { 00:14:10.885 "name": "spare", 00:14:10.885 "uuid": "83b2c087-3dc3-57fd-901b-49219a82421f", 00:14:10.885 "is_configured": true, 00:14:10.885 "data_offset": 0, 00:14:10.885 "data_size": 65536 00:14:10.885 }, 00:14:10.885 { 00:14:10.885 "name": "BaseBdev2", 00:14:10.885 "uuid": "eb03108c-ecda-51e0-9f0e-c54f21f1353d", 00:14:10.885 "is_configured": true, 00:14:10.885 "data_offset": 0, 00:14:10.885 "data_size": 65536 00:14:10.885 }, 00:14:10.885 { 00:14:10.885 "name": "BaseBdev3", 00:14:10.885 "uuid": "0f1041c7-edcf-5738-854d-b694da9829b6", 00:14:10.885 "is_configured": true, 00:14:10.885 "data_offset": 0, 00:14:10.885 "data_size": 65536 00:14:10.885 }, 00:14:10.885 { 00:14:10.885 "name": "BaseBdev4", 00:14:10.885 "uuid": "948dc191-0fba-551a-a044-ef9c4d004931", 00:14:10.885 "is_configured": true, 00:14:10.885 "data_offset": 0, 00:14:10.885 "data_size": 65536 00:14:10.885 } 00:14:10.885 ] 00:14:10.885 }' 00:14:10.885 23:47:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:10.885 23:47:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:10.885 23:47:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:10.885 23:47:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:10.885 23:47:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:14:12.265 23:47:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:12.265 23:47:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:12.265 23:47:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:12.265 23:47:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:12.265 23:47:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:12.265 23:47:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:12.265 23:47:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:12.265 23:47:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:12.265 23:47:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:12.265 23:47:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:12.265 23:47:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:12.265 23:48:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:12.265 "name": "raid_bdev1", 00:14:12.265 "uuid": "fb5fe94c-b292-46be-acce-5eed89b870f4", 00:14:12.265 "strip_size_kb": 64, 00:14:12.265 "state": "online", 00:14:12.265 "raid_level": "raid5f", 00:14:12.265 "superblock": false, 00:14:12.265 "num_base_bdevs": 4, 00:14:12.265 "num_base_bdevs_discovered": 4, 00:14:12.265 "num_base_bdevs_operational": 4, 00:14:12.265 "process": { 00:14:12.265 "type": "rebuild", 00:14:12.265 "target": "spare", 00:14:12.265 "progress": { 00:14:12.265 "blocks": 174720, 00:14:12.265 "percent": 88 00:14:12.265 } 00:14:12.265 }, 00:14:12.265 "base_bdevs_list": [ 00:14:12.265 { 00:14:12.265 "name": "spare", 00:14:12.265 "uuid": "83b2c087-3dc3-57fd-901b-49219a82421f", 00:14:12.265 "is_configured": true, 00:14:12.265 "data_offset": 0, 00:14:12.265 "data_size": 65536 00:14:12.265 }, 00:14:12.265 { 00:14:12.265 "name": "BaseBdev2", 00:14:12.265 "uuid": "eb03108c-ecda-51e0-9f0e-c54f21f1353d", 00:14:12.265 "is_configured": true, 00:14:12.265 "data_offset": 0, 00:14:12.265 "data_size": 65536 00:14:12.265 }, 00:14:12.265 { 00:14:12.265 "name": "BaseBdev3", 00:14:12.265 "uuid": "0f1041c7-edcf-5738-854d-b694da9829b6", 00:14:12.265 "is_configured": true, 00:14:12.265 "data_offset": 0, 00:14:12.265 "data_size": 65536 00:14:12.265 }, 00:14:12.265 { 00:14:12.265 "name": "BaseBdev4", 00:14:12.265 "uuid": "948dc191-0fba-551a-a044-ef9c4d004931", 00:14:12.265 "is_configured": true, 00:14:12.265 "data_offset": 0, 00:14:12.265 "data_size": 65536 00:14:12.265 } 00:14:12.265 ] 00:14:12.265 }' 00:14:12.265 23:48:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:12.265 23:48:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:12.265 23:48:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:12.265 23:48:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:12.265 23:48:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:14:13.203 23:48:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:13.203 23:48:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:13.203 23:48:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:13.203 23:48:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:13.203 23:48:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:13.203 23:48:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:13.203 23:48:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:13.203 23:48:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:13.203 23:48:01 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:13.203 23:48:01 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:13.203 [2024-11-26 23:48:01.156574] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:14:13.203 23:48:01 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:13.203 [2024-11-26 23:48:01.156681] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:14:13.203 [2024-11-26 23:48:01.156734] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:13.203 23:48:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:13.203 "name": "raid_bdev1", 00:14:13.203 "uuid": "fb5fe94c-b292-46be-acce-5eed89b870f4", 00:14:13.203 "strip_size_kb": 64, 00:14:13.203 "state": "online", 00:14:13.203 "raid_level": "raid5f", 00:14:13.203 "superblock": false, 00:14:13.203 "num_base_bdevs": 4, 00:14:13.203 "num_base_bdevs_discovered": 4, 00:14:13.203 "num_base_bdevs_operational": 4, 00:14:13.203 "process": { 00:14:13.203 "type": "rebuild", 00:14:13.203 "target": "spare", 00:14:13.203 "progress": { 00:14:13.203 "blocks": 195840, 00:14:13.203 "percent": 99 00:14:13.203 } 00:14:13.203 }, 00:14:13.203 "base_bdevs_list": [ 00:14:13.203 { 00:14:13.203 "name": "spare", 00:14:13.203 "uuid": "83b2c087-3dc3-57fd-901b-49219a82421f", 00:14:13.203 "is_configured": true, 00:14:13.203 "data_offset": 0, 00:14:13.203 "data_size": 65536 00:14:13.203 }, 00:14:13.203 { 00:14:13.203 "name": "BaseBdev2", 00:14:13.203 "uuid": "eb03108c-ecda-51e0-9f0e-c54f21f1353d", 00:14:13.203 "is_configured": true, 00:14:13.203 "data_offset": 0, 00:14:13.203 "data_size": 65536 00:14:13.203 }, 00:14:13.203 { 00:14:13.203 "name": "BaseBdev3", 00:14:13.203 "uuid": "0f1041c7-edcf-5738-854d-b694da9829b6", 00:14:13.203 "is_configured": true, 00:14:13.203 "data_offset": 0, 00:14:13.203 "data_size": 65536 00:14:13.203 }, 00:14:13.203 { 00:14:13.203 "name": "BaseBdev4", 00:14:13.203 "uuid": "948dc191-0fba-551a-a044-ef9c4d004931", 00:14:13.203 "is_configured": true, 00:14:13.203 "data_offset": 0, 00:14:13.203 "data_size": 65536 00:14:13.203 } 00:14:13.203 ] 00:14:13.203 }' 00:14:13.203 23:48:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:13.203 23:48:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:13.203 23:48:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:13.203 23:48:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:13.203 23:48:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:14:14.584 23:48:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:14.584 23:48:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:14.584 23:48:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:14.584 23:48:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:14.584 23:48:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:14.584 23:48:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:14.584 23:48:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:14.584 23:48:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:14.584 23:48:02 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:14.584 23:48:02 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:14.584 23:48:02 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:14.584 23:48:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:14.584 "name": "raid_bdev1", 00:14:14.584 "uuid": "fb5fe94c-b292-46be-acce-5eed89b870f4", 00:14:14.584 "strip_size_kb": 64, 00:14:14.584 "state": "online", 00:14:14.584 "raid_level": "raid5f", 00:14:14.584 "superblock": false, 00:14:14.584 "num_base_bdevs": 4, 00:14:14.584 "num_base_bdevs_discovered": 4, 00:14:14.584 "num_base_bdevs_operational": 4, 00:14:14.584 "base_bdevs_list": [ 00:14:14.584 { 00:14:14.584 "name": "spare", 00:14:14.584 "uuid": "83b2c087-3dc3-57fd-901b-49219a82421f", 00:14:14.584 "is_configured": true, 00:14:14.584 "data_offset": 0, 00:14:14.584 "data_size": 65536 00:14:14.584 }, 00:14:14.584 { 00:14:14.584 "name": "BaseBdev2", 00:14:14.584 "uuid": "eb03108c-ecda-51e0-9f0e-c54f21f1353d", 00:14:14.584 "is_configured": true, 00:14:14.584 "data_offset": 0, 00:14:14.584 "data_size": 65536 00:14:14.584 }, 00:14:14.584 { 00:14:14.584 "name": "BaseBdev3", 00:14:14.584 "uuid": "0f1041c7-edcf-5738-854d-b694da9829b6", 00:14:14.584 "is_configured": true, 00:14:14.584 "data_offset": 0, 00:14:14.584 "data_size": 65536 00:14:14.584 }, 00:14:14.584 { 00:14:14.584 "name": "BaseBdev4", 00:14:14.584 "uuid": "948dc191-0fba-551a-a044-ef9c4d004931", 00:14:14.584 "is_configured": true, 00:14:14.584 "data_offset": 0, 00:14:14.584 "data_size": 65536 00:14:14.584 } 00:14:14.584 ] 00:14:14.584 }' 00:14:14.584 23:48:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:14.584 23:48:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:14:14.584 23:48:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:14.584 23:48:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:14:14.584 23:48:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@709 -- # break 00:14:14.584 23:48:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:14:14.584 23:48:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:14.584 23:48:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:14:14.584 23:48:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=none 00:14:14.584 23:48:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:14.584 23:48:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:14.584 23:48:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:14.584 23:48:02 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:14.584 23:48:02 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:14.584 23:48:02 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:14.584 23:48:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:14.584 "name": "raid_bdev1", 00:14:14.584 "uuid": "fb5fe94c-b292-46be-acce-5eed89b870f4", 00:14:14.584 "strip_size_kb": 64, 00:14:14.584 "state": "online", 00:14:14.584 "raid_level": "raid5f", 00:14:14.584 "superblock": false, 00:14:14.584 "num_base_bdevs": 4, 00:14:14.584 "num_base_bdevs_discovered": 4, 00:14:14.584 "num_base_bdevs_operational": 4, 00:14:14.584 "base_bdevs_list": [ 00:14:14.584 { 00:14:14.584 "name": "spare", 00:14:14.584 "uuid": "83b2c087-3dc3-57fd-901b-49219a82421f", 00:14:14.584 "is_configured": true, 00:14:14.584 "data_offset": 0, 00:14:14.584 "data_size": 65536 00:14:14.584 }, 00:14:14.584 { 00:14:14.584 "name": "BaseBdev2", 00:14:14.584 "uuid": "eb03108c-ecda-51e0-9f0e-c54f21f1353d", 00:14:14.584 "is_configured": true, 00:14:14.584 "data_offset": 0, 00:14:14.584 "data_size": 65536 00:14:14.584 }, 00:14:14.584 { 00:14:14.584 "name": "BaseBdev3", 00:14:14.584 "uuid": "0f1041c7-edcf-5738-854d-b694da9829b6", 00:14:14.584 "is_configured": true, 00:14:14.584 "data_offset": 0, 00:14:14.584 "data_size": 65536 00:14:14.584 }, 00:14:14.584 { 00:14:14.584 "name": "BaseBdev4", 00:14:14.584 "uuid": "948dc191-0fba-551a-a044-ef9c4d004931", 00:14:14.584 "is_configured": true, 00:14:14.584 "data_offset": 0, 00:14:14.584 "data_size": 65536 00:14:14.584 } 00:14:14.584 ] 00:14:14.584 }' 00:14:14.584 23:48:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:14.584 23:48:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:14:14.584 23:48:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:14.584 23:48:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:14:14.584 23:48:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 4 00:14:14.584 23:48:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:14.584 23:48:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:14.584 23:48:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:14.584 23:48:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:14.585 23:48:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:14.585 23:48:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:14.585 23:48:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:14.585 23:48:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:14.585 23:48:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:14.585 23:48:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:14.585 23:48:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:14.585 23:48:02 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:14.585 23:48:02 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:14.585 23:48:02 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:14.585 23:48:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:14.585 "name": "raid_bdev1", 00:14:14.585 "uuid": "fb5fe94c-b292-46be-acce-5eed89b870f4", 00:14:14.585 "strip_size_kb": 64, 00:14:14.585 "state": "online", 00:14:14.585 "raid_level": "raid5f", 00:14:14.585 "superblock": false, 00:14:14.585 "num_base_bdevs": 4, 00:14:14.585 "num_base_bdevs_discovered": 4, 00:14:14.585 "num_base_bdevs_operational": 4, 00:14:14.585 "base_bdevs_list": [ 00:14:14.585 { 00:14:14.585 "name": "spare", 00:14:14.585 "uuid": "83b2c087-3dc3-57fd-901b-49219a82421f", 00:14:14.585 "is_configured": true, 00:14:14.585 "data_offset": 0, 00:14:14.585 "data_size": 65536 00:14:14.585 }, 00:14:14.585 { 00:14:14.585 "name": "BaseBdev2", 00:14:14.585 "uuid": "eb03108c-ecda-51e0-9f0e-c54f21f1353d", 00:14:14.585 "is_configured": true, 00:14:14.585 "data_offset": 0, 00:14:14.585 "data_size": 65536 00:14:14.585 }, 00:14:14.585 { 00:14:14.585 "name": "BaseBdev3", 00:14:14.585 "uuid": "0f1041c7-edcf-5738-854d-b694da9829b6", 00:14:14.585 "is_configured": true, 00:14:14.585 "data_offset": 0, 00:14:14.585 "data_size": 65536 00:14:14.585 }, 00:14:14.585 { 00:14:14.585 "name": "BaseBdev4", 00:14:14.585 "uuid": "948dc191-0fba-551a-a044-ef9c4d004931", 00:14:14.585 "is_configured": true, 00:14:14.585 "data_offset": 0, 00:14:14.585 "data_size": 65536 00:14:14.585 } 00:14:14.585 ] 00:14:14.585 }' 00:14:14.585 23:48:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:14.585 23:48:02 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:15.154 23:48:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:14:15.154 23:48:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:15.154 23:48:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:15.154 [2024-11-26 23:48:03.027029] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:14:15.154 [2024-11-26 23:48:03.027063] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:14:15.154 [2024-11-26 23:48:03.027151] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:14:15.154 [2024-11-26 23:48:03.027239] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:14:15.154 [2024-11-26 23:48:03.027250] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:14:15.154 23:48:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:15.155 23:48:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:15.155 23:48:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@720 -- # jq length 00:14:15.155 23:48:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:15.155 23:48:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:15.155 23:48:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:15.155 23:48:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:14:15.155 23:48:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:14:15.155 23:48:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:14:15.155 23:48:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:14:15.155 23:48:03 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:14:15.155 23:48:03 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:14:15.155 23:48:03 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@10 -- # local bdev_list 00:14:15.155 23:48:03 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:14:15.155 23:48:03 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@11 -- # local nbd_list 00:14:15.155 23:48:03 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@12 -- # local i 00:14:15.155 23:48:03 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:14:15.155 23:48:03 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:14:15.155 23:48:03 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:14:15.155 /dev/nbd0 00:14:15.414 23:48:03 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:14:15.414 23:48:03 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:14:15.414 23:48:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:14:15.414 23:48:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@873 -- # local i 00:14:15.414 23:48:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:14:15.414 23:48:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:14:15.414 23:48:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:14:15.414 23:48:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@877 -- # break 00:14:15.414 23:48:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:14:15.414 23:48:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:14:15.414 23:48:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:14:15.414 1+0 records in 00:14:15.414 1+0 records out 00:14:15.414 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000501271 s, 8.2 MB/s 00:14:15.414 23:48:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:15.414 23:48:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@890 -- # size=4096 00:14:15.414 23:48:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:15.414 23:48:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:14:15.414 23:48:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@893 -- # return 0 00:14:15.414 23:48:03 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:14:15.414 23:48:03 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:14:15.414 23:48:03 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:14:15.414 /dev/nbd1 00:14:15.414 23:48:03 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:14:15.673 23:48:03 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:14:15.673 23:48:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:14:15.673 23:48:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@873 -- # local i 00:14:15.673 23:48:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:14:15.673 23:48:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:14:15.673 23:48:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:14:15.673 23:48:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@877 -- # break 00:14:15.673 23:48:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:14:15.673 23:48:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:14:15.673 23:48:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:14:15.673 1+0 records in 00:14:15.673 1+0 records out 00:14:15.673 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000277193 s, 14.8 MB/s 00:14:15.673 23:48:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:15.673 23:48:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@890 -- # size=4096 00:14:15.673 23:48:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:15.673 23:48:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:14:15.673 23:48:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@893 -- # return 0 00:14:15.673 23:48:03 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:14:15.673 23:48:03 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:14:15.673 23:48:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@738 -- # cmp -i 0 /dev/nbd0 /dev/nbd1 00:14:15.673 23:48:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:14:15.673 23:48:03 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:14:15.673 23:48:03 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:14:15.673 23:48:03 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@50 -- # local nbd_list 00:14:15.673 23:48:03 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@51 -- # local i 00:14:15.673 23:48:03 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:14:15.673 23:48:03 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:14:15.673 23:48:03 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:14:15.674 23:48:03 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:14:15.674 23:48:03 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:14:15.674 23:48:03 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:14:15.674 23:48:03 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:14:15.674 23:48:03 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:14:15.936 23:48:03 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:14:15.936 23:48:03 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:14:15.936 23:48:03 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:14:15.936 23:48:03 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:14:15.936 23:48:03 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:14:15.936 23:48:03 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:14:15.936 23:48:03 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:14:15.936 23:48:03 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:14:15.936 23:48:03 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:14:15.936 23:48:03 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:14:15.936 23:48:03 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:14:15.936 23:48:03 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:14:15.936 23:48:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@743 -- # '[' false = true ']' 00:14:15.936 23:48:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@784 -- # killprocess 94682 00:14:15.936 23:48:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@954 -- # '[' -z 94682 ']' 00:14:15.936 23:48:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@958 -- # kill -0 94682 00:14:15.936 23:48:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@959 -- # uname 00:14:15.936 23:48:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:14:15.936 23:48:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 94682 00:14:15.936 killing process with pid 94682 00:14:15.936 Received shutdown signal, test time was about 60.000000 seconds 00:14:15.936 00:14:15.936 Latency(us) 00:14:15.936 [2024-11-26T23:48:04.068Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:14:15.936 [2024-11-26T23:48:04.068Z] =================================================================================================================== 00:14:15.936 [2024-11-26T23:48:04.068Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:14:15.936 23:48:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:14:15.936 23:48:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:14:15.936 23:48:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 94682' 00:14:15.936 23:48:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@973 -- # kill 94682 00:14:15.936 [2024-11-26 23:48:04.034133] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:14:15.936 23:48:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@978 -- # wait 94682 00:14:16.206 [2024-11-26 23:48:04.082683] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:14:16.207 23:48:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@786 -- # return 0 00:14:16.207 00:14:16.207 real 0m18.123s 00:14:16.207 user 0m21.952s 00:14:16.207 sys 0m2.120s 00:14:16.207 ************************************ 00:14:16.207 END TEST raid5f_rebuild_test 00:14:16.207 ************************************ 00:14:16.207 23:48:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:14:16.207 23:48:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:16.482 23:48:04 bdev_raid -- bdev/bdev_raid.sh@991 -- # run_test raid5f_rebuild_test_sb raid_rebuild_test raid5f 4 true false true 00:14:16.482 23:48:04 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:14:16.482 23:48:04 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:14:16.482 23:48:04 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:14:16.482 ************************************ 00:14:16.482 START TEST raid5f_rebuild_test_sb 00:14:16.482 ************************************ 00:14:16.482 23:48:04 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@1129 -- # raid_rebuild_test raid5f 4 true false true 00:14:16.482 23:48:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@569 -- # local raid_level=raid5f 00:14:16.482 23:48:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=4 00:14:16.482 23:48:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@571 -- # local superblock=true 00:14:16.482 23:48:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:14:16.482 23:48:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@573 -- # local verify=true 00:14:16.482 23:48:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:14:16.482 23:48:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:14:16.482 23:48:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:14:16.482 23:48:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:14:16.482 23:48:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:14:16.482 23:48:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:14:16.482 23:48:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:14:16.482 23:48:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:14:16.482 23:48:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev3 00:14:16.482 23:48:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:14:16.482 23:48:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:14:16.482 23:48:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev4 00:14:16.482 23:48:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:14:16.482 23:48:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:14:16.482 23:48:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:14:16.482 23:48:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:14:16.482 23:48:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:14:16.482 23:48:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # local strip_size 00:14:16.482 23:48:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@577 -- # local create_arg 00:14:16.482 23:48:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:14:16.482 23:48:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@579 -- # local data_offset 00:14:16.482 23:48:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@581 -- # '[' raid5f '!=' raid1 ']' 00:14:16.482 23:48:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@582 -- # '[' false = true ']' 00:14:16.482 23:48:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@586 -- # strip_size=64 00:14:16.483 23:48:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@587 -- # create_arg+=' -z 64' 00:14:16.483 23:48:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@592 -- # '[' true = true ']' 00:14:16.483 23:48:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@593 -- # create_arg+=' -s' 00:14:16.483 23:48:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@597 -- # raid_pid=95180 00:14:16.483 23:48:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@598 -- # waitforlisten 95180 00:14:16.483 23:48:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:14:16.483 23:48:04 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@835 -- # '[' -z 95180 ']' 00:14:16.483 23:48:04 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:14:16.483 23:48:04 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@840 -- # local max_retries=100 00:14:16.483 23:48:04 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:14:16.483 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:14:16.483 23:48:04 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@844 -- # xtrace_disable 00:14:16.483 23:48:04 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:16.483 [2024-11-26 23:48:04.445911] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:14:16.483 [2024-11-26 23:48:04.446117] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.ealI/O size of 3145728 is greater than zero copy threshold (65536). 00:14:16.483 Zero copy mechanism will not be used. 00:14:16.483 :6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid95180 ] 00:14:16.483 [2024-11-26 23:48:04.599152] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:14:16.743 [2024-11-26 23:48:04.624286] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:14:16.743 [2024-11-26 23:48:04.667111] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:14:16.743 [2024-11-26 23:48:04.667143] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:14:17.312 23:48:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:14:17.313 23:48:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@868 -- # return 0 00:14:17.313 23:48:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:14:17.313 23:48:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:14:17.313 23:48:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:17.313 23:48:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:17.313 BaseBdev1_malloc 00:14:17.313 23:48:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:17.313 23:48:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:14:17.313 23:48:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:17.313 23:48:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:17.313 [2024-11-26 23:48:05.302406] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:14:17.313 [2024-11-26 23:48:05.302462] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:17.313 [2024-11-26 23:48:05.302486] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:14:17.313 [2024-11-26 23:48:05.302497] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:17.313 [2024-11-26 23:48:05.304705] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:17.313 [2024-11-26 23:48:05.304739] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:14:17.313 BaseBdev1 00:14:17.313 23:48:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:17.313 23:48:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:14:17.313 23:48:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:14:17.313 23:48:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:17.313 23:48:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:17.313 BaseBdev2_malloc 00:14:17.313 23:48:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:17.313 23:48:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:14:17.313 23:48:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:17.313 23:48:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:17.313 [2024-11-26 23:48:05.331045] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:14:17.313 [2024-11-26 23:48:05.331098] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:17.313 [2024-11-26 23:48:05.331137] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:14:17.313 [2024-11-26 23:48:05.331145] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:17.313 [2024-11-26 23:48:05.333187] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:17.313 [2024-11-26 23:48:05.333225] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:14:17.313 BaseBdev2 00:14:17.313 23:48:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:17.313 23:48:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:14:17.313 23:48:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:14:17.313 23:48:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:17.313 23:48:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:17.313 BaseBdev3_malloc 00:14:17.313 23:48:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:17.313 23:48:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev3_malloc -p BaseBdev3 00:14:17.313 23:48:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:17.313 23:48:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:17.313 [2024-11-26 23:48:05.359553] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev3_malloc 00:14:17.313 [2024-11-26 23:48:05.359601] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:17.313 [2024-11-26 23:48:05.359623] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:14:17.313 [2024-11-26 23:48:05.359631] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:17.313 [2024-11-26 23:48:05.361581] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:17.313 [2024-11-26 23:48:05.361664] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:14:17.313 BaseBdev3 00:14:17.313 23:48:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:17.313 23:48:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:14:17.313 23:48:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:14:17.313 23:48:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:17.313 23:48:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:17.313 BaseBdev4_malloc 00:14:17.313 23:48:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:17.313 23:48:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev4_malloc -p BaseBdev4 00:14:17.313 23:48:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:17.313 23:48:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:17.313 [2024-11-26 23:48:05.411553] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev4_malloc 00:14:17.313 [2024-11-26 23:48:05.411650] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:17.313 [2024-11-26 23:48:05.411699] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:14:17.313 [2024-11-26 23:48:05.411720] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:17.313 [2024-11-26 23:48:05.415309] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:17.313 [2024-11-26 23:48:05.415380] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:14:17.313 BaseBdev4 00:14:17.313 23:48:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:17.313 23:48:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:14:17.313 23:48:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:17.313 23:48:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:17.313 spare_malloc 00:14:17.313 23:48:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:17.313 23:48:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:14:17.313 23:48:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:17.313 23:48:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:17.573 spare_delay 00:14:17.573 23:48:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:17.573 23:48:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:14:17.573 23:48:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:17.573 23:48:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:17.573 [2024-11-26 23:48:05.452636] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:14:17.573 [2024-11-26 23:48:05.452679] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:17.573 [2024-11-26 23:48:05.452698] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009c80 00:14:17.573 [2024-11-26 23:48:05.452706] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:17.573 [2024-11-26 23:48:05.454792] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:17.573 [2024-11-26 23:48:05.454862] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:14:17.573 spare 00:14:17.573 23:48:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:17.573 23:48:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 00:14:17.573 23:48:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:17.573 23:48:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:17.573 [2024-11-26 23:48:05.464693] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:14:17.573 [2024-11-26 23:48:05.466470] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:14:17.573 [2024-11-26 23:48:05.466530] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:14:17.573 [2024-11-26 23:48:05.466602] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:14:17.573 [2024-11-26 23:48:05.466770] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:14:17.573 [2024-11-26 23:48:05.466781] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:14:17.573 [2024-11-26 23:48:05.467013] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:14:17.573 [2024-11-26 23:48:05.467501] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:14:17.573 [2024-11-26 23:48:05.467515] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:14:17.573 [2024-11-26 23:48:05.467629] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:17.573 23:48:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:17.573 23:48:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 4 00:14:17.573 23:48:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:17.573 23:48:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:17.574 23:48:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:17.574 23:48:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:17.574 23:48:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:17.574 23:48:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:17.574 23:48:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:17.574 23:48:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:17.574 23:48:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:17.574 23:48:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:17.574 23:48:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:17.574 23:48:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:17.574 23:48:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:17.574 23:48:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:17.574 23:48:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:17.574 "name": "raid_bdev1", 00:14:17.574 "uuid": "d8dd1718-7857-4e41-889d-86a3ec73047a", 00:14:17.574 "strip_size_kb": 64, 00:14:17.574 "state": "online", 00:14:17.574 "raid_level": "raid5f", 00:14:17.574 "superblock": true, 00:14:17.574 "num_base_bdevs": 4, 00:14:17.574 "num_base_bdevs_discovered": 4, 00:14:17.574 "num_base_bdevs_operational": 4, 00:14:17.574 "base_bdevs_list": [ 00:14:17.574 { 00:14:17.574 "name": "BaseBdev1", 00:14:17.574 "uuid": "6c5fc277-b15f-5313-9b21-89e4e67a70d5", 00:14:17.574 "is_configured": true, 00:14:17.574 "data_offset": 2048, 00:14:17.574 "data_size": 63488 00:14:17.574 }, 00:14:17.574 { 00:14:17.574 "name": "BaseBdev2", 00:14:17.574 "uuid": "9a0be541-d5b7-5b3a-a946-e13a4325db1a", 00:14:17.574 "is_configured": true, 00:14:17.574 "data_offset": 2048, 00:14:17.574 "data_size": 63488 00:14:17.574 }, 00:14:17.574 { 00:14:17.574 "name": "BaseBdev3", 00:14:17.574 "uuid": "01335c00-2adc-5bc4-8d68-965b0b20c407", 00:14:17.574 "is_configured": true, 00:14:17.574 "data_offset": 2048, 00:14:17.574 "data_size": 63488 00:14:17.574 }, 00:14:17.574 { 00:14:17.574 "name": "BaseBdev4", 00:14:17.574 "uuid": "23b26d34-25c3-57d6-96d0-d4cd067bae4e", 00:14:17.574 "is_configured": true, 00:14:17.574 "data_offset": 2048, 00:14:17.574 "data_size": 63488 00:14:17.574 } 00:14:17.574 ] 00:14:17.574 }' 00:14:17.574 23:48:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:17.574 23:48:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:17.834 23:48:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:14:17.834 23:48:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:17.834 23:48:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:17.834 23:48:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:14:17.834 [2024-11-26 23:48:05.916721] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:14:17.834 23:48:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:17.834 23:48:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=190464 00:14:18.094 23:48:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:14:18.094 23:48:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:18.094 23:48:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:18.094 23:48:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:18.094 23:48:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:18.094 23:48:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # data_offset=2048 00:14:18.094 23:48:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:14:18.094 23:48:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:14:18.094 23:48:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:14:18.094 23:48:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:14:18.094 23:48:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:14:18.094 23:48:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:14:18.094 23:48:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # local bdev_list 00:14:18.094 23:48:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:14:18.094 23:48:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # local nbd_list 00:14:18.094 23:48:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@12 -- # local i 00:14:18.094 23:48:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:14:18.094 23:48:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:14:18.094 23:48:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:14:18.094 [2024-11-26 23:48:06.188151] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000027a0 00:14:18.094 /dev/nbd0 00:14:18.355 23:48:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:14:18.355 23:48:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:14:18.355 23:48:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:14:18.355 23:48:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@873 -- # local i 00:14:18.355 23:48:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:14:18.355 23:48:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:14:18.355 23:48:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:14:18.355 23:48:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@877 -- # break 00:14:18.355 23:48:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:14:18.355 23:48:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:14:18.355 23:48:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:14:18.355 1+0 records in 00:14:18.355 1+0 records out 00:14:18.355 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000298596 s, 13.7 MB/s 00:14:18.355 23:48:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:18.355 23:48:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@890 -- # size=4096 00:14:18.355 23:48:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:18.355 23:48:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:14:18.355 23:48:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@893 -- # return 0 00:14:18.355 23:48:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:14:18.355 23:48:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:14:18.355 23:48:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@629 -- # '[' raid5f = raid5f ']' 00:14:18.355 23:48:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@630 -- # write_unit_size=384 00:14:18.355 23:48:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@631 -- # echo 192 00:14:18.355 23:48:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=196608 count=496 oflag=direct 00:14:18.615 496+0 records in 00:14:18.615 496+0 records out 00:14:18.615 97517568 bytes (98 MB, 93 MiB) copied, 0.381222 s, 256 MB/s 00:14:18.615 23:48:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:14:18.615 23:48:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:14:18.615 23:48:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:14:18.615 23:48:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # local nbd_list 00:14:18.615 23:48:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@51 -- # local i 00:14:18.615 23:48:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:14:18.615 23:48:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:14:18.874 23:48:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:14:18.874 [2024-11-26 23:48:06.852420] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:18.874 23:48:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:14:18.874 23:48:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:14:18.874 23:48:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:14:18.874 23:48:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:14:18.874 23:48:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:14:18.874 23:48:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:14:18.874 23:48:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:14:18.874 23:48:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:14:18.874 23:48:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:18.874 23:48:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:18.874 [2024-11-26 23:48:06.868465] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:14:18.874 23:48:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:18.875 23:48:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:14:18.875 23:48:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:18.875 23:48:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:18.875 23:48:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:18.875 23:48:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:18.875 23:48:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:18.875 23:48:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:18.875 23:48:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:18.875 23:48:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:18.875 23:48:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:18.875 23:48:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:18.875 23:48:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:18.875 23:48:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:18.875 23:48:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:18.875 23:48:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:18.875 23:48:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:18.875 "name": "raid_bdev1", 00:14:18.875 "uuid": "d8dd1718-7857-4e41-889d-86a3ec73047a", 00:14:18.875 "strip_size_kb": 64, 00:14:18.875 "state": "online", 00:14:18.875 "raid_level": "raid5f", 00:14:18.875 "superblock": true, 00:14:18.875 "num_base_bdevs": 4, 00:14:18.875 "num_base_bdevs_discovered": 3, 00:14:18.875 "num_base_bdevs_operational": 3, 00:14:18.875 "base_bdevs_list": [ 00:14:18.875 { 00:14:18.875 "name": null, 00:14:18.875 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:18.875 "is_configured": false, 00:14:18.875 "data_offset": 0, 00:14:18.875 "data_size": 63488 00:14:18.875 }, 00:14:18.875 { 00:14:18.875 "name": "BaseBdev2", 00:14:18.875 "uuid": "9a0be541-d5b7-5b3a-a946-e13a4325db1a", 00:14:18.875 "is_configured": true, 00:14:18.875 "data_offset": 2048, 00:14:18.875 "data_size": 63488 00:14:18.875 }, 00:14:18.875 { 00:14:18.875 "name": "BaseBdev3", 00:14:18.875 "uuid": "01335c00-2adc-5bc4-8d68-965b0b20c407", 00:14:18.875 "is_configured": true, 00:14:18.875 "data_offset": 2048, 00:14:18.875 "data_size": 63488 00:14:18.875 }, 00:14:18.875 { 00:14:18.875 "name": "BaseBdev4", 00:14:18.875 "uuid": "23b26d34-25c3-57d6-96d0-d4cd067bae4e", 00:14:18.875 "is_configured": true, 00:14:18.875 "data_offset": 2048, 00:14:18.875 "data_size": 63488 00:14:18.875 } 00:14:18.875 ] 00:14:18.875 }' 00:14:18.875 23:48:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:18.875 23:48:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:19.444 23:48:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:14:19.444 23:48:07 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:19.444 23:48:07 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:19.444 [2024-11-26 23:48:07.335687] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:14:19.444 [2024-11-26 23:48:07.339862] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000270a0 00:14:19.444 23:48:07 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:19.444 23:48:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@647 -- # sleep 1 00:14:19.444 [2024-11-26 23:48:07.341999] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:14:20.383 23:48:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:20.383 23:48:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:20.383 23:48:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:20.383 23:48:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:20.383 23:48:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:20.383 23:48:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:20.383 23:48:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:20.383 23:48:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:20.383 23:48:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:20.383 23:48:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:20.383 23:48:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:20.383 "name": "raid_bdev1", 00:14:20.383 "uuid": "d8dd1718-7857-4e41-889d-86a3ec73047a", 00:14:20.383 "strip_size_kb": 64, 00:14:20.383 "state": "online", 00:14:20.383 "raid_level": "raid5f", 00:14:20.383 "superblock": true, 00:14:20.383 "num_base_bdevs": 4, 00:14:20.383 "num_base_bdevs_discovered": 4, 00:14:20.383 "num_base_bdevs_operational": 4, 00:14:20.383 "process": { 00:14:20.383 "type": "rebuild", 00:14:20.383 "target": "spare", 00:14:20.383 "progress": { 00:14:20.383 "blocks": 19200, 00:14:20.383 "percent": 10 00:14:20.383 } 00:14:20.383 }, 00:14:20.383 "base_bdevs_list": [ 00:14:20.383 { 00:14:20.383 "name": "spare", 00:14:20.383 "uuid": "a88541d9-452e-5793-a3ed-3bb8392daf8f", 00:14:20.383 "is_configured": true, 00:14:20.383 "data_offset": 2048, 00:14:20.383 "data_size": 63488 00:14:20.383 }, 00:14:20.383 { 00:14:20.383 "name": "BaseBdev2", 00:14:20.383 "uuid": "9a0be541-d5b7-5b3a-a946-e13a4325db1a", 00:14:20.383 "is_configured": true, 00:14:20.383 "data_offset": 2048, 00:14:20.383 "data_size": 63488 00:14:20.383 }, 00:14:20.383 { 00:14:20.383 "name": "BaseBdev3", 00:14:20.383 "uuid": "01335c00-2adc-5bc4-8d68-965b0b20c407", 00:14:20.383 "is_configured": true, 00:14:20.383 "data_offset": 2048, 00:14:20.383 "data_size": 63488 00:14:20.383 }, 00:14:20.383 { 00:14:20.383 "name": "BaseBdev4", 00:14:20.383 "uuid": "23b26d34-25c3-57d6-96d0-d4cd067bae4e", 00:14:20.383 "is_configured": true, 00:14:20.383 "data_offset": 2048, 00:14:20.383 "data_size": 63488 00:14:20.383 } 00:14:20.383 ] 00:14:20.383 }' 00:14:20.383 23:48:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:20.383 23:48:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:20.383 23:48:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:20.383 23:48:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:20.383 23:48:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:14:20.383 23:48:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:20.383 23:48:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:20.383 [2024-11-26 23:48:08.502801] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:14:20.643 [2024-11-26 23:48:08.547433] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:14:20.643 [2024-11-26 23:48:08.547495] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:20.643 [2024-11-26 23:48:08.547514] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:14:20.643 [2024-11-26 23:48:08.547521] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:14:20.643 23:48:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:20.643 23:48:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:14:20.643 23:48:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:20.643 23:48:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:20.643 23:48:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:20.643 23:48:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:20.643 23:48:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:20.643 23:48:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:20.643 23:48:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:20.643 23:48:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:20.643 23:48:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:20.644 23:48:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:20.644 23:48:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:20.644 23:48:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:20.644 23:48:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:20.644 23:48:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:20.644 23:48:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:20.644 "name": "raid_bdev1", 00:14:20.644 "uuid": "d8dd1718-7857-4e41-889d-86a3ec73047a", 00:14:20.644 "strip_size_kb": 64, 00:14:20.644 "state": "online", 00:14:20.644 "raid_level": "raid5f", 00:14:20.644 "superblock": true, 00:14:20.644 "num_base_bdevs": 4, 00:14:20.644 "num_base_bdevs_discovered": 3, 00:14:20.644 "num_base_bdevs_operational": 3, 00:14:20.644 "base_bdevs_list": [ 00:14:20.644 { 00:14:20.644 "name": null, 00:14:20.644 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:20.644 "is_configured": false, 00:14:20.644 "data_offset": 0, 00:14:20.644 "data_size": 63488 00:14:20.644 }, 00:14:20.644 { 00:14:20.644 "name": "BaseBdev2", 00:14:20.644 "uuid": "9a0be541-d5b7-5b3a-a946-e13a4325db1a", 00:14:20.644 "is_configured": true, 00:14:20.644 "data_offset": 2048, 00:14:20.644 "data_size": 63488 00:14:20.644 }, 00:14:20.644 { 00:14:20.644 "name": "BaseBdev3", 00:14:20.644 "uuid": "01335c00-2adc-5bc4-8d68-965b0b20c407", 00:14:20.644 "is_configured": true, 00:14:20.644 "data_offset": 2048, 00:14:20.644 "data_size": 63488 00:14:20.644 }, 00:14:20.644 { 00:14:20.644 "name": "BaseBdev4", 00:14:20.644 "uuid": "23b26d34-25c3-57d6-96d0-d4cd067bae4e", 00:14:20.644 "is_configured": true, 00:14:20.644 "data_offset": 2048, 00:14:20.644 "data_size": 63488 00:14:20.644 } 00:14:20.644 ] 00:14:20.644 }' 00:14:20.644 23:48:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:20.644 23:48:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:20.904 23:48:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:14:20.904 23:48:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:20.904 23:48:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:14:20.904 23:48:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:14:20.904 23:48:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:20.904 23:48:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:20.904 23:48:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:20.904 23:48:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:20.904 23:48:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:20.904 23:48:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:20.904 23:48:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:20.904 "name": "raid_bdev1", 00:14:20.904 "uuid": "d8dd1718-7857-4e41-889d-86a3ec73047a", 00:14:20.904 "strip_size_kb": 64, 00:14:20.904 "state": "online", 00:14:20.904 "raid_level": "raid5f", 00:14:20.904 "superblock": true, 00:14:20.904 "num_base_bdevs": 4, 00:14:20.904 "num_base_bdevs_discovered": 3, 00:14:20.904 "num_base_bdevs_operational": 3, 00:14:20.904 "base_bdevs_list": [ 00:14:20.904 { 00:14:20.904 "name": null, 00:14:20.904 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:20.904 "is_configured": false, 00:14:20.904 "data_offset": 0, 00:14:20.904 "data_size": 63488 00:14:20.904 }, 00:14:20.904 { 00:14:20.904 "name": "BaseBdev2", 00:14:20.904 "uuid": "9a0be541-d5b7-5b3a-a946-e13a4325db1a", 00:14:20.904 "is_configured": true, 00:14:20.904 "data_offset": 2048, 00:14:20.904 "data_size": 63488 00:14:20.904 }, 00:14:20.904 { 00:14:20.904 "name": "BaseBdev3", 00:14:20.904 "uuid": "01335c00-2adc-5bc4-8d68-965b0b20c407", 00:14:20.904 "is_configured": true, 00:14:20.904 "data_offset": 2048, 00:14:20.904 "data_size": 63488 00:14:20.904 }, 00:14:20.904 { 00:14:20.904 "name": "BaseBdev4", 00:14:20.904 "uuid": "23b26d34-25c3-57d6-96d0-d4cd067bae4e", 00:14:20.904 "is_configured": true, 00:14:20.904 "data_offset": 2048, 00:14:20.904 "data_size": 63488 00:14:20.904 } 00:14:20.904 ] 00:14:20.904 }' 00:14:20.904 23:48:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:21.175 23:48:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:14:21.175 23:48:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:21.175 23:48:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:14:21.175 23:48:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:14:21.175 23:48:09 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:21.175 23:48:09 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:21.175 [2024-11-26 23:48:09.108195] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:14:21.175 [2024-11-26 23:48:09.112314] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000027170 00:14:21.175 23:48:09 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:21.175 23:48:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@663 -- # sleep 1 00:14:21.175 [2024-11-26 23:48:09.114405] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:14:22.120 23:48:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:22.121 23:48:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:22.121 23:48:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:22.121 23:48:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:22.121 23:48:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:22.121 23:48:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:22.121 23:48:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:22.121 23:48:10 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:22.121 23:48:10 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:22.121 23:48:10 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:22.121 23:48:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:22.121 "name": "raid_bdev1", 00:14:22.121 "uuid": "d8dd1718-7857-4e41-889d-86a3ec73047a", 00:14:22.121 "strip_size_kb": 64, 00:14:22.121 "state": "online", 00:14:22.121 "raid_level": "raid5f", 00:14:22.121 "superblock": true, 00:14:22.121 "num_base_bdevs": 4, 00:14:22.121 "num_base_bdevs_discovered": 4, 00:14:22.121 "num_base_bdevs_operational": 4, 00:14:22.121 "process": { 00:14:22.121 "type": "rebuild", 00:14:22.121 "target": "spare", 00:14:22.121 "progress": { 00:14:22.121 "blocks": 19200, 00:14:22.121 "percent": 10 00:14:22.121 } 00:14:22.121 }, 00:14:22.121 "base_bdevs_list": [ 00:14:22.121 { 00:14:22.121 "name": "spare", 00:14:22.121 "uuid": "a88541d9-452e-5793-a3ed-3bb8392daf8f", 00:14:22.121 "is_configured": true, 00:14:22.121 "data_offset": 2048, 00:14:22.121 "data_size": 63488 00:14:22.121 }, 00:14:22.121 { 00:14:22.121 "name": "BaseBdev2", 00:14:22.121 "uuid": "9a0be541-d5b7-5b3a-a946-e13a4325db1a", 00:14:22.121 "is_configured": true, 00:14:22.121 "data_offset": 2048, 00:14:22.121 "data_size": 63488 00:14:22.121 }, 00:14:22.121 { 00:14:22.121 "name": "BaseBdev3", 00:14:22.121 "uuid": "01335c00-2adc-5bc4-8d68-965b0b20c407", 00:14:22.121 "is_configured": true, 00:14:22.121 "data_offset": 2048, 00:14:22.121 "data_size": 63488 00:14:22.121 }, 00:14:22.121 { 00:14:22.121 "name": "BaseBdev4", 00:14:22.121 "uuid": "23b26d34-25c3-57d6-96d0-d4cd067bae4e", 00:14:22.121 "is_configured": true, 00:14:22.121 "data_offset": 2048, 00:14:22.121 "data_size": 63488 00:14:22.121 } 00:14:22.121 ] 00:14:22.121 }' 00:14:22.121 23:48:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:22.121 23:48:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:22.121 23:48:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:22.381 23:48:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:22.381 23:48:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@666 -- # '[' true = true ']' 00:14:22.381 23:48:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@666 -- # '[' = false ']' 00:14:22.381 /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh: line 666: [: =: unary operator expected 00:14:22.381 23:48:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=4 00:14:22.381 23:48:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@693 -- # '[' raid5f = raid1 ']' 00:14:22.381 23:48:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@706 -- # local timeout=522 00:14:22.381 23:48:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:22.381 23:48:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:22.381 23:48:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:22.381 23:48:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:22.381 23:48:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:22.381 23:48:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:22.381 23:48:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:22.381 23:48:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:22.381 23:48:10 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:22.381 23:48:10 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:22.382 23:48:10 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:22.382 23:48:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:22.382 "name": "raid_bdev1", 00:14:22.382 "uuid": "d8dd1718-7857-4e41-889d-86a3ec73047a", 00:14:22.382 "strip_size_kb": 64, 00:14:22.382 "state": "online", 00:14:22.382 "raid_level": "raid5f", 00:14:22.382 "superblock": true, 00:14:22.382 "num_base_bdevs": 4, 00:14:22.382 "num_base_bdevs_discovered": 4, 00:14:22.382 "num_base_bdevs_operational": 4, 00:14:22.382 "process": { 00:14:22.382 "type": "rebuild", 00:14:22.382 "target": "spare", 00:14:22.382 "progress": { 00:14:22.382 "blocks": 21120, 00:14:22.382 "percent": 11 00:14:22.382 } 00:14:22.382 }, 00:14:22.382 "base_bdevs_list": [ 00:14:22.382 { 00:14:22.382 "name": "spare", 00:14:22.382 "uuid": "a88541d9-452e-5793-a3ed-3bb8392daf8f", 00:14:22.382 "is_configured": true, 00:14:22.382 "data_offset": 2048, 00:14:22.382 "data_size": 63488 00:14:22.382 }, 00:14:22.382 { 00:14:22.382 "name": "BaseBdev2", 00:14:22.382 "uuid": "9a0be541-d5b7-5b3a-a946-e13a4325db1a", 00:14:22.382 "is_configured": true, 00:14:22.382 "data_offset": 2048, 00:14:22.382 "data_size": 63488 00:14:22.382 }, 00:14:22.382 { 00:14:22.382 "name": "BaseBdev3", 00:14:22.382 "uuid": "01335c00-2adc-5bc4-8d68-965b0b20c407", 00:14:22.382 "is_configured": true, 00:14:22.382 "data_offset": 2048, 00:14:22.382 "data_size": 63488 00:14:22.382 }, 00:14:22.382 { 00:14:22.382 "name": "BaseBdev4", 00:14:22.382 "uuid": "23b26d34-25c3-57d6-96d0-d4cd067bae4e", 00:14:22.382 "is_configured": true, 00:14:22.382 "data_offset": 2048, 00:14:22.382 "data_size": 63488 00:14:22.382 } 00:14:22.382 ] 00:14:22.382 }' 00:14:22.382 23:48:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:22.382 23:48:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:22.382 23:48:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:22.382 23:48:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:22.382 23:48:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:14:23.321 23:48:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:23.321 23:48:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:23.321 23:48:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:23.321 23:48:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:23.321 23:48:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:23.321 23:48:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:23.321 23:48:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:23.321 23:48:11 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:23.321 23:48:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:23.321 23:48:11 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:23.321 23:48:11 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:23.321 23:48:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:23.321 "name": "raid_bdev1", 00:14:23.321 "uuid": "d8dd1718-7857-4e41-889d-86a3ec73047a", 00:14:23.321 "strip_size_kb": 64, 00:14:23.321 "state": "online", 00:14:23.321 "raid_level": "raid5f", 00:14:23.321 "superblock": true, 00:14:23.321 "num_base_bdevs": 4, 00:14:23.321 "num_base_bdevs_discovered": 4, 00:14:23.321 "num_base_bdevs_operational": 4, 00:14:23.321 "process": { 00:14:23.321 "type": "rebuild", 00:14:23.321 "target": "spare", 00:14:23.321 "progress": { 00:14:23.321 "blocks": 42240, 00:14:23.321 "percent": 22 00:14:23.321 } 00:14:23.321 }, 00:14:23.321 "base_bdevs_list": [ 00:14:23.321 { 00:14:23.321 "name": "spare", 00:14:23.321 "uuid": "a88541d9-452e-5793-a3ed-3bb8392daf8f", 00:14:23.321 "is_configured": true, 00:14:23.321 "data_offset": 2048, 00:14:23.321 "data_size": 63488 00:14:23.321 }, 00:14:23.321 { 00:14:23.321 "name": "BaseBdev2", 00:14:23.321 "uuid": "9a0be541-d5b7-5b3a-a946-e13a4325db1a", 00:14:23.321 "is_configured": true, 00:14:23.321 "data_offset": 2048, 00:14:23.321 "data_size": 63488 00:14:23.321 }, 00:14:23.321 { 00:14:23.321 "name": "BaseBdev3", 00:14:23.321 "uuid": "01335c00-2adc-5bc4-8d68-965b0b20c407", 00:14:23.321 "is_configured": true, 00:14:23.321 "data_offset": 2048, 00:14:23.321 "data_size": 63488 00:14:23.321 }, 00:14:23.321 { 00:14:23.321 "name": "BaseBdev4", 00:14:23.321 "uuid": "23b26d34-25c3-57d6-96d0-d4cd067bae4e", 00:14:23.321 "is_configured": true, 00:14:23.321 "data_offset": 2048, 00:14:23.321 "data_size": 63488 00:14:23.321 } 00:14:23.321 ] 00:14:23.321 }' 00:14:23.321 23:48:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:23.580 23:48:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:23.580 23:48:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:23.580 23:48:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:23.580 23:48:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:14:24.520 23:48:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:24.520 23:48:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:24.520 23:48:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:24.520 23:48:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:24.520 23:48:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:24.520 23:48:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:24.520 23:48:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:24.520 23:48:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:24.520 23:48:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:24.520 23:48:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:24.520 23:48:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:24.520 23:48:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:24.520 "name": "raid_bdev1", 00:14:24.520 "uuid": "d8dd1718-7857-4e41-889d-86a3ec73047a", 00:14:24.520 "strip_size_kb": 64, 00:14:24.520 "state": "online", 00:14:24.520 "raid_level": "raid5f", 00:14:24.520 "superblock": true, 00:14:24.520 "num_base_bdevs": 4, 00:14:24.520 "num_base_bdevs_discovered": 4, 00:14:24.520 "num_base_bdevs_operational": 4, 00:14:24.520 "process": { 00:14:24.520 "type": "rebuild", 00:14:24.520 "target": "spare", 00:14:24.520 "progress": { 00:14:24.520 "blocks": 63360, 00:14:24.520 "percent": 33 00:14:24.520 } 00:14:24.520 }, 00:14:24.520 "base_bdevs_list": [ 00:14:24.520 { 00:14:24.520 "name": "spare", 00:14:24.520 "uuid": "a88541d9-452e-5793-a3ed-3bb8392daf8f", 00:14:24.520 "is_configured": true, 00:14:24.520 "data_offset": 2048, 00:14:24.520 "data_size": 63488 00:14:24.520 }, 00:14:24.520 { 00:14:24.520 "name": "BaseBdev2", 00:14:24.520 "uuid": "9a0be541-d5b7-5b3a-a946-e13a4325db1a", 00:14:24.520 "is_configured": true, 00:14:24.520 "data_offset": 2048, 00:14:24.520 "data_size": 63488 00:14:24.520 }, 00:14:24.520 { 00:14:24.520 "name": "BaseBdev3", 00:14:24.520 "uuid": "01335c00-2adc-5bc4-8d68-965b0b20c407", 00:14:24.520 "is_configured": true, 00:14:24.520 "data_offset": 2048, 00:14:24.520 "data_size": 63488 00:14:24.520 }, 00:14:24.520 { 00:14:24.520 "name": "BaseBdev4", 00:14:24.520 "uuid": "23b26d34-25c3-57d6-96d0-d4cd067bae4e", 00:14:24.520 "is_configured": true, 00:14:24.520 "data_offset": 2048, 00:14:24.520 "data_size": 63488 00:14:24.520 } 00:14:24.520 ] 00:14:24.520 }' 00:14:24.520 23:48:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:24.520 23:48:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:24.520 23:48:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:24.521 23:48:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:24.521 23:48:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:14:25.917 23:48:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:25.917 23:48:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:25.917 23:48:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:25.917 23:48:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:25.917 23:48:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:25.917 23:48:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:25.917 23:48:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:25.917 23:48:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:25.917 23:48:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:25.917 23:48:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:25.917 23:48:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:25.917 23:48:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:25.917 "name": "raid_bdev1", 00:14:25.917 "uuid": "d8dd1718-7857-4e41-889d-86a3ec73047a", 00:14:25.917 "strip_size_kb": 64, 00:14:25.917 "state": "online", 00:14:25.917 "raid_level": "raid5f", 00:14:25.917 "superblock": true, 00:14:25.917 "num_base_bdevs": 4, 00:14:25.917 "num_base_bdevs_discovered": 4, 00:14:25.917 "num_base_bdevs_operational": 4, 00:14:25.917 "process": { 00:14:25.917 "type": "rebuild", 00:14:25.917 "target": "spare", 00:14:25.917 "progress": { 00:14:25.917 "blocks": 86400, 00:14:25.917 "percent": 45 00:14:25.917 } 00:14:25.917 }, 00:14:25.917 "base_bdevs_list": [ 00:14:25.917 { 00:14:25.917 "name": "spare", 00:14:25.917 "uuid": "a88541d9-452e-5793-a3ed-3bb8392daf8f", 00:14:25.917 "is_configured": true, 00:14:25.917 "data_offset": 2048, 00:14:25.917 "data_size": 63488 00:14:25.917 }, 00:14:25.917 { 00:14:25.917 "name": "BaseBdev2", 00:14:25.917 "uuid": "9a0be541-d5b7-5b3a-a946-e13a4325db1a", 00:14:25.917 "is_configured": true, 00:14:25.917 "data_offset": 2048, 00:14:25.917 "data_size": 63488 00:14:25.917 }, 00:14:25.917 { 00:14:25.917 "name": "BaseBdev3", 00:14:25.917 "uuid": "01335c00-2adc-5bc4-8d68-965b0b20c407", 00:14:25.917 "is_configured": true, 00:14:25.917 "data_offset": 2048, 00:14:25.917 "data_size": 63488 00:14:25.917 }, 00:14:25.917 { 00:14:25.917 "name": "BaseBdev4", 00:14:25.917 "uuid": "23b26d34-25c3-57d6-96d0-d4cd067bae4e", 00:14:25.917 "is_configured": true, 00:14:25.917 "data_offset": 2048, 00:14:25.917 "data_size": 63488 00:14:25.917 } 00:14:25.917 ] 00:14:25.917 }' 00:14:25.917 23:48:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:25.917 23:48:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:25.917 23:48:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:25.917 23:48:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:25.917 23:48:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:14:26.857 23:48:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:26.857 23:48:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:26.857 23:48:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:26.857 23:48:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:26.857 23:48:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:26.857 23:48:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:26.857 23:48:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:26.857 23:48:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:26.857 23:48:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:26.857 23:48:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:26.857 23:48:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:26.857 23:48:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:26.857 "name": "raid_bdev1", 00:14:26.857 "uuid": "d8dd1718-7857-4e41-889d-86a3ec73047a", 00:14:26.857 "strip_size_kb": 64, 00:14:26.857 "state": "online", 00:14:26.857 "raid_level": "raid5f", 00:14:26.857 "superblock": true, 00:14:26.857 "num_base_bdevs": 4, 00:14:26.857 "num_base_bdevs_discovered": 4, 00:14:26.857 "num_base_bdevs_operational": 4, 00:14:26.857 "process": { 00:14:26.857 "type": "rebuild", 00:14:26.857 "target": "spare", 00:14:26.857 "progress": { 00:14:26.857 "blocks": 107520, 00:14:26.857 "percent": 56 00:14:26.857 } 00:14:26.857 }, 00:14:26.857 "base_bdevs_list": [ 00:14:26.857 { 00:14:26.857 "name": "spare", 00:14:26.857 "uuid": "a88541d9-452e-5793-a3ed-3bb8392daf8f", 00:14:26.857 "is_configured": true, 00:14:26.857 "data_offset": 2048, 00:14:26.857 "data_size": 63488 00:14:26.857 }, 00:14:26.857 { 00:14:26.857 "name": "BaseBdev2", 00:14:26.857 "uuid": "9a0be541-d5b7-5b3a-a946-e13a4325db1a", 00:14:26.857 "is_configured": true, 00:14:26.857 "data_offset": 2048, 00:14:26.857 "data_size": 63488 00:14:26.857 }, 00:14:26.857 { 00:14:26.857 "name": "BaseBdev3", 00:14:26.857 "uuid": "01335c00-2adc-5bc4-8d68-965b0b20c407", 00:14:26.857 "is_configured": true, 00:14:26.857 "data_offset": 2048, 00:14:26.857 "data_size": 63488 00:14:26.857 }, 00:14:26.857 { 00:14:26.857 "name": "BaseBdev4", 00:14:26.857 "uuid": "23b26d34-25c3-57d6-96d0-d4cd067bae4e", 00:14:26.857 "is_configured": true, 00:14:26.857 "data_offset": 2048, 00:14:26.857 "data_size": 63488 00:14:26.857 } 00:14:26.857 ] 00:14:26.857 }' 00:14:26.857 23:48:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:26.857 23:48:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:26.857 23:48:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:26.857 23:48:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:26.857 23:48:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:14:27.796 23:48:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:27.796 23:48:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:27.796 23:48:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:27.796 23:48:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:27.796 23:48:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:27.796 23:48:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:27.796 23:48:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:27.796 23:48:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:27.796 23:48:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:27.796 23:48:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:27.796 23:48:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:28.055 23:48:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:28.055 "name": "raid_bdev1", 00:14:28.055 "uuid": "d8dd1718-7857-4e41-889d-86a3ec73047a", 00:14:28.055 "strip_size_kb": 64, 00:14:28.055 "state": "online", 00:14:28.055 "raid_level": "raid5f", 00:14:28.055 "superblock": true, 00:14:28.055 "num_base_bdevs": 4, 00:14:28.055 "num_base_bdevs_discovered": 4, 00:14:28.055 "num_base_bdevs_operational": 4, 00:14:28.055 "process": { 00:14:28.055 "type": "rebuild", 00:14:28.055 "target": "spare", 00:14:28.055 "progress": { 00:14:28.055 "blocks": 128640, 00:14:28.055 "percent": 67 00:14:28.055 } 00:14:28.055 }, 00:14:28.056 "base_bdevs_list": [ 00:14:28.056 { 00:14:28.056 "name": "spare", 00:14:28.056 "uuid": "a88541d9-452e-5793-a3ed-3bb8392daf8f", 00:14:28.056 "is_configured": true, 00:14:28.056 "data_offset": 2048, 00:14:28.056 "data_size": 63488 00:14:28.056 }, 00:14:28.056 { 00:14:28.056 "name": "BaseBdev2", 00:14:28.056 "uuid": "9a0be541-d5b7-5b3a-a946-e13a4325db1a", 00:14:28.056 "is_configured": true, 00:14:28.056 "data_offset": 2048, 00:14:28.056 "data_size": 63488 00:14:28.056 }, 00:14:28.056 { 00:14:28.056 "name": "BaseBdev3", 00:14:28.056 "uuid": "01335c00-2adc-5bc4-8d68-965b0b20c407", 00:14:28.056 "is_configured": true, 00:14:28.056 "data_offset": 2048, 00:14:28.056 "data_size": 63488 00:14:28.056 }, 00:14:28.056 { 00:14:28.056 "name": "BaseBdev4", 00:14:28.056 "uuid": "23b26d34-25c3-57d6-96d0-d4cd067bae4e", 00:14:28.056 "is_configured": true, 00:14:28.056 "data_offset": 2048, 00:14:28.056 "data_size": 63488 00:14:28.056 } 00:14:28.056 ] 00:14:28.056 }' 00:14:28.056 23:48:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:28.056 23:48:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:28.056 23:48:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:28.056 23:48:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:28.056 23:48:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:14:28.994 23:48:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:28.994 23:48:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:28.994 23:48:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:28.994 23:48:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:28.994 23:48:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:28.994 23:48:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:28.994 23:48:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:28.994 23:48:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:28.994 23:48:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:28.994 23:48:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:28.994 23:48:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:28.994 23:48:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:28.994 "name": "raid_bdev1", 00:14:28.994 "uuid": "d8dd1718-7857-4e41-889d-86a3ec73047a", 00:14:28.994 "strip_size_kb": 64, 00:14:28.994 "state": "online", 00:14:28.994 "raid_level": "raid5f", 00:14:28.994 "superblock": true, 00:14:28.994 "num_base_bdevs": 4, 00:14:28.994 "num_base_bdevs_discovered": 4, 00:14:28.994 "num_base_bdevs_operational": 4, 00:14:28.994 "process": { 00:14:28.994 "type": "rebuild", 00:14:28.994 "target": "spare", 00:14:28.994 "progress": { 00:14:28.994 "blocks": 151680, 00:14:28.994 "percent": 79 00:14:28.994 } 00:14:28.994 }, 00:14:28.994 "base_bdevs_list": [ 00:14:28.994 { 00:14:28.994 "name": "spare", 00:14:28.994 "uuid": "a88541d9-452e-5793-a3ed-3bb8392daf8f", 00:14:28.994 "is_configured": true, 00:14:28.994 "data_offset": 2048, 00:14:28.994 "data_size": 63488 00:14:28.994 }, 00:14:28.994 { 00:14:28.994 "name": "BaseBdev2", 00:14:28.994 "uuid": "9a0be541-d5b7-5b3a-a946-e13a4325db1a", 00:14:28.994 "is_configured": true, 00:14:28.994 "data_offset": 2048, 00:14:28.994 "data_size": 63488 00:14:28.994 }, 00:14:28.994 { 00:14:28.994 "name": "BaseBdev3", 00:14:28.994 "uuid": "01335c00-2adc-5bc4-8d68-965b0b20c407", 00:14:28.994 "is_configured": true, 00:14:28.994 "data_offset": 2048, 00:14:28.994 "data_size": 63488 00:14:28.994 }, 00:14:28.994 { 00:14:28.994 "name": "BaseBdev4", 00:14:28.994 "uuid": "23b26d34-25c3-57d6-96d0-d4cd067bae4e", 00:14:28.994 "is_configured": true, 00:14:28.994 "data_offset": 2048, 00:14:28.994 "data_size": 63488 00:14:28.994 } 00:14:28.994 ] 00:14:28.994 }' 00:14:28.994 23:48:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:29.255 23:48:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:29.255 23:48:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:29.255 23:48:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:29.255 23:48:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:14:30.193 23:48:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:30.193 23:48:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:30.193 23:48:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:30.193 23:48:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:30.193 23:48:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:30.193 23:48:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:30.193 23:48:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:30.193 23:48:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:30.193 23:48:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:30.193 23:48:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:30.193 23:48:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:30.193 23:48:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:30.193 "name": "raid_bdev1", 00:14:30.193 "uuid": "d8dd1718-7857-4e41-889d-86a3ec73047a", 00:14:30.193 "strip_size_kb": 64, 00:14:30.193 "state": "online", 00:14:30.193 "raid_level": "raid5f", 00:14:30.193 "superblock": true, 00:14:30.193 "num_base_bdevs": 4, 00:14:30.193 "num_base_bdevs_discovered": 4, 00:14:30.193 "num_base_bdevs_operational": 4, 00:14:30.193 "process": { 00:14:30.193 "type": "rebuild", 00:14:30.193 "target": "spare", 00:14:30.193 "progress": { 00:14:30.193 "blocks": 172800, 00:14:30.193 "percent": 90 00:14:30.193 } 00:14:30.193 }, 00:14:30.193 "base_bdevs_list": [ 00:14:30.193 { 00:14:30.193 "name": "spare", 00:14:30.193 "uuid": "a88541d9-452e-5793-a3ed-3bb8392daf8f", 00:14:30.193 "is_configured": true, 00:14:30.193 "data_offset": 2048, 00:14:30.193 "data_size": 63488 00:14:30.193 }, 00:14:30.193 { 00:14:30.193 "name": "BaseBdev2", 00:14:30.193 "uuid": "9a0be541-d5b7-5b3a-a946-e13a4325db1a", 00:14:30.193 "is_configured": true, 00:14:30.193 "data_offset": 2048, 00:14:30.193 "data_size": 63488 00:14:30.193 }, 00:14:30.193 { 00:14:30.193 "name": "BaseBdev3", 00:14:30.193 "uuid": "01335c00-2adc-5bc4-8d68-965b0b20c407", 00:14:30.193 "is_configured": true, 00:14:30.193 "data_offset": 2048, 00:14:30.193 "data_size": 63488 00:14:30.193 }, 00:14:30.193 { 00:14:30.193 "name": "BaseBdev4", 00:14:30.193 "uuid": "23b26d34-25c3-57d6-96d0-d4cd067bae4e", 00:14:30.193 "is_configured": true, 00:14:30.193 "data_offset": 2048, 00:14:30.193 "data_size": 63488 00:14:30.193 } 00:14:30.193 ] 00:14:30.193 }' 00:14:30.193 23:48:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:30.193 23:48:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:30.193 23:48:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:30.453 23:48:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:30.453 23:48:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:14:31.393 [2024-11-26 23:48:19.155140] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:14:31.393 [2024-11-26 23:48:19.155219] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:14:31.393 [2024-11-26 23:48:19.155318] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:31.393 23:48:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:31.393 23:48:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:31.393 23:48:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:31.393 23:48:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:31.393 23:48:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:31.393 23:48:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:31.393 23:48:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:31.393 23:48:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:31.393 23:48:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:31.393 23:48:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:31.393 23:48:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:31.393 23:48:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:31.393 "name": "raid_bdev1", 00:14:31.393 "uuid": "d8dd1718-7857-4e41-889d-86a3ec73047a", 00:14:31.393 "strip_size_kb": 64, 00:14:31.393 "state": "online", 00:14:31.393 "raid_level": "raid5f", 00:14:31.393 "superblock": true, 00:14:31.393 "num_base_bdevs": 4, 00:14:31.393 "num_base_bdevs_discovered": 4, 00:14:31.393 "num_base_bdevs_operational": 4, 00:14:31.393 "base_bdevs_list": [ 00:14:31.393 { 00:14:31.394 "name": "spare", 00:14:31.394 "uuid": "a88541d9-452e-5793-a3ed-3bb8392daf8f", 00:14:31.394 "is_configured": true, 00:14:31.394 "data_offset": 2048, 00:14:31.394 "data_size": 63488 00:14:31.394 }, 00:14:31.394 { 00:14:31.394 "name": "BaseBdev2", 00:14:31.394 "uuid": "9a0be541-d5b7-5b3a-a946-e13a4325db1a", 00:14:31.394 "is_configured": true, 00:14:31.394 "data_offset": 2048, 00:14:31.394 "data_size": 63488 00:14:31.394 }, 00:14:31.394 { 00:14:31.394 "name": "BaseBdev3", 00:14:31.394 "uuid": "01335c00-2adc-5bc4-8d68-965b0b20c407", 00:14:31.394 "is_configured": true, 00:14:31.394 "data_offset": 2048, 00:14:31.394 "data_size": 63488 00:14:31.394 }, 00:14:31.394 { 00:14:31.394 "name": "BaseBdev4", 00:14:31.394 "uuid": "23b26d34-25c3-57d6-96d0-d4cd067bae4e", 00:14:31.394 "is_configured": true, 00:14:31.394 "data_offset": 2048, 00:14:31.394 "data_size": 63488 00:14:31.394 } 00:14:31.394 ] 00:14:31.394 }' 00:14:31.394 23:48:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:31.394 23:48:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:14:31.394 23:48:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:31.394 23:48:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:14:31.394 23:48:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@709 -- # break 00:14:31.394 23:48:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:14:31.394 23:48:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:31.394 23:48:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:14:31.394 23:48:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:14:31.394 23:48:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:31.394 23:48:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:31.394 23:48:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:31.394 23:48:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:31.394 23:48:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:31.394 23:48:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:31.653 23:48:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:31.653 "name": "raid_bdev1", 00:14:31.653 "uuid": "d8dd1718-7857-4e41-889d-86a3ec73047a", 00:14:31.653 "strip_size_kb": 64, 00:14:31.653 "state": "online", 00:14:31.653 "raid_level": "raid5f", 00:14:31.653 "superblock": true, 00:14:31.653 "num_base_bdevs": 4, 00:14:31.653 "num_base_bdevs_discovered": 4, 00:14:31.653 "num_base_bdevs_operational": 4, 00:14:31.653 "base_bdevs_list": [ 00:14:31.653 { 00:14:31.653 "name": "spare", 00:14:31.653 "uuid": "a88541d9-452e-5793-a3ed-3bb8392daf8f", 00:14:31.653 "is_configured": true, 00:14:31.653 "data_offset": 2048, 00:14:31.653 "data_size": 63488 00:14:31.653 }, 00:14:31.653 { 00:14:31.653 "name": "BaseBdev2", 00:14:31.653 "uuid": "9a0be541-d5b7-5b3a-a946-e13a4325db1a", 00:14:31.653 "is_configured": true, 00:14:31.653 "data_offset": 2048, 00:14:31.653 "data_size": 63488 00:14:31.653 }, 00:14:31.653 { 00:14:31.653 "name": "BaseBdev3", 00:14:31.653 "uuid": "01335c00-2adc-5bc4-8d68-965b0b20c407", 00:14:31.653 "is_configured": true, 00:14:31.653 "data_offset": 2048, 00:14:31.653 "data_size": 63488 00:14:31.653 }, 00:14:31.653 { 00:14:31.653 "name": "BaseBdev4", 00:14:31.653 "uuid": "23b26d34-25c3-57d6-96d0-d4cd067bae4e", 00:14:31.653 "is_configured": true, 00:14:31.653 "data_offset": 2048, 00:14:31.653 "data_size": 63488 00:14:31.653 } 00:14:31.653 ] 00:14:31.653 }' 00:14:31.653 23:48:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:31.653 23:48:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:14:31.653 23:48:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:31.653 23:48:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:14:31.653 23:48:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 4 00:14:31.653 23:48:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:31.653 23:48:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:31.653 23:48:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:31.653 23:48:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:31.653 23:48:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:31.653 23:48:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:31.653 23:48:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:31.653 23:48:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:31.653 23:48:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:31.653 23:48:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:31.653 23:48:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:31.653 23:48:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:31.653 23:48:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:31.653 23:48:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:31.653 23:48:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:31.653 "name": "raid_bdev1", 00:14:31.653 "uuid": "d8dd1718-7857-4e41-889d-86a3ec73047a", 00:14:31.653 "strip_size_kb": 64, 00:14:31.653 "state": "online", 00:14:31.653 "raid_level": "raid5f", 00:14:31.653 "superblock": true, 00:14:31.653 "num_base_bdevs": 4, 00:14:31.653 "num_base_bdevs_discovered": 4, 00:14:31.653 "num_base_bdevs_operational": 4, 00:14:31.653 "base_bdevs_list": [ 00:14:31.653 { 00:14:31.653 "name": "spare", 00:14:31.653 "uuid": "a88541d9-452e-5793-a3ed-3bb8392daf8f", 00:14:31.653 "is_configured": true, 00:14:31.653 "data_offset": 2048, 00:14:31.653 "data_size": 63488 00:14:31.653 }, 00:14:31.653 { 00:14:31.653 "name": "BaseBdev2", 00:14:31.653 "uuid": "9a0be541-d5b7-5b3a-a946-e13a4325db1a", 00:14:31.653 "is_configured": true, 00:14:31.653 "data_offset": 2048, 00:14:31.653 "data_size": 63488 00:14:31.653 }, 00:14:31.653 { 00:14:31.653 "name": "BaseBdev3", 00:14:31.653 "uuid": "01335c00-2adc-5bc4-8d68-965b0b20c407", 00:14:31.653 "is_configured": true, 00:14:31.653 "data_offset": 2048, 00:14:31.653 "data_size": 63488 00:14:31.653 }, 00:14:31.653 { 00:14:31.653 "name": "BaseBdev4", 00:14:31.653 "uuid": "23b26d34-25c3-57d6-96d0-d4cd067bae4e", 00:14:31.653 "is_configured": true, 00:14:31.653 "data_offset": 2048, 00:14:31.653 "data_size": 63488 00:14:31.653 } 00:14:31.653 ] 00:14:31.653 }' 00:14:31.653 23:48:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:31.653 23:48:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:32.224 23:48:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:14:32.224 23:48:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:32.224 23:48:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:32.224 [2024-11-26 23:48:20.075106] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:14:32.224 [2024-11-26 23:48:20.075138] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:14:32.224 [2024-11-26 23:48:20.075226] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:14:32.224 [2024-11-26 23:48:20.075320] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:14:32.224 [2024-11-26 23:48:20.075353] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:14:32.224 23:48:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:32.224 23:48:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:32.224 23:48:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:32.224 23:48:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # jq length 00:14:32.224 23:48:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:32.224 23:48:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:32.224 23:48:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:14:32.224 23:48:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:14:32.224 23:48:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:14:32.224 23:48:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:14:32.224 23:48:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:14:32.224 23:48:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:14:32.224 23:48:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # local bdev_list 00:14:32.224 23:48:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:14:32.224 23:48:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # local nbd_list 00:14:32.224 23:48:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@12 -- # local i 00:14:32.224 23:48:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:14:32.224 23:48:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:14:32.224 23:48:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:14:32.224 /dev/nbd0 00:14:32.485 23:48:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:14:32.485 23:48:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:14:32.485 23:48:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:14:32.485 23:48:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@873 -- # local i 00:14:32.485 23:48:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:14:32.486 23:48:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:14:32.486 23:48:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:14:32.486 23:48:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@877 -- # break 00:14:32.486 23:48:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:14:32.486 23:48:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:14:32.486 23:48:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:14:32.486 1+0 records in 00:14:32.486 1+0 records out 00:14:32.486 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000338703 s, 12.1 MB/s 00:14:32.486 23:48:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:32.486 23:48:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@890 -- # size=4096 00:14:32.486 23:48:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:32.486 23:48:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:14:32.486 23:48:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@893 -- # return 0 00:14:32.486 23:48:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:14:32.486 23:48:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:14:32.486 23:48:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:14:32.486 /dev/nbd1 00:14:32.486 23:48:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:14:32.486 23:48:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:14:32.486 23:48:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:14:32.486 23:48:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@873 -- # local i 00:14:32.486 23:48:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:14:32.486 23:48:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:14:32.486 23:48:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:14:32.486 23:48:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@877 -- # break 00:14:32.486 23:48:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:14:32.486 23:48:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:14:32.486 23:48:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:14:32.486 1+0 records in 00:14:32.486 1+0 records out 00:14:32.486 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000406642 s, 10.1 MB/s 00:14:32.486 23:48:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:32.746 23:48:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@890 -- # size=4096 00:14:32.746 23:48:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:32.746 23:48:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:14:32.746 23:48:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@893 -- # return 0 00:14:32.746 23:48:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:14:32.746 23:48:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:14:32.746 23:48:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@738 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:14:32.746 23:48:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:14:32.746 23:48:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:14:32.746 23:48:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:14:32.746 23:48:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # local nbd_list 00:14:32.746 23:48:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@51 -- # local i 00:14:32.747 23:48:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:14:32.747 23:48:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:14:32.747 23:48:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:14:32.747 23:48:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:14:32.747 23:48:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:14:32.747 23:48:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:14:32.747 23:48:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:14:32.747 23:48:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:14:32.747 23:48:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:14:32.747 23:48:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:14:32.747 23:48:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:14:32.747 23:48:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:14:33.007 23:48:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:14:33.007 23:48:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:14:33.007 23:48:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:14:33.007 23:48:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:14:33.007 23:48:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:14:33.007 23:48:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:14:33.007 23:48:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:14:33.007 23:48:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:14:33.007 23:48:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@743 -- # '[' true = true ']' 00:14:33.007 23:48:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@745 -- # rpc_cmd bdev_passthru_delete spare 00:14:33.007 23:48:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:33.007 23:48:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:33.007 23:48:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:33.007 23:48:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@746 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:14:33.007 23:48:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:33.007 23:48:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:33.007 [2024-11-26 23:48:21.079442] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:14:33.007 [2024-11-26 23:48:21.079510] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:33.007 [2024-11-26 23:48:21.079532] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ae80 00:14:33.007 [2024-11-26 23:48:21.079543] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:33.007 [2024-11-26 23:48:21.081652] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:33.007 [2024-11-26 23:48:21.081692] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:14:33.007 [2024-11-26 23:48:21.081785] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:14:33.007 [2024-11-26 23:48:21.081834] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:14:33.007 [2024-11-26 23:48:21.081952] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:14:33.007 [2024-11-26 23:48:21.082050] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:14:33.007 [2024-11-26 23:48:21.082123] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:14:33.007 spare 00:14:33.007 23:48:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:33.007 23:48:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@747 -- # rpc_cmd bdev_wait_for_examine 00:14:33.007 23:48:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:33.007 23:48:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:33.268 [2024-11-26 23:48:21.182015] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001580 00:14:33.268 [2024-11-26 23:48:21.182052] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:14:33.268 [2024-11-26 23:48:21.182282] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000045820 00:14:33.268 [2024-11-26 23:48:21.182765] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001580 00:14:33.268 [2024-11-26 23:48:21.182789] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001580 00:14:33.268 [2024-11-26 23:48:21.182935] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:33.268 23:48:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:33.268 23:48:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@749 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 4 00:14:33.268 23:48:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:33.268 23:48:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:33.268 23:48:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:33.268 23:48:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:33.268 23:48:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:33.268 23:48:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:33.268 23:48:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:33.268 23:48:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:33.268 23:48:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:33.268 23:48:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:33.268 23:48:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:33.268 23:48:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:33.268 23:48:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:33.268 23:48:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:33.268 23:48:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:33.268 "name": "raid_bdev1", 00:14:33.268 "uuid": "d8dd1718-7857-4e41-889d-86a3ec73047a", 00:14:33.268 "strip_size_kb": 64, 00:14:33.268 "state": "online", 00:14:33.268 "raid_level": "raid5f", 00:14:33.268 "superblock": true, 00:14:33.268 "num_base_bdevs": 4, 00:14:33.268 "num_base_bdevs_discovered": 4, 00:14:33.268 "num_base_bdevs_operational": 4, 00:14:33.268 "base_bdevs_list": [ 00:14:33.268 { 00:14:33.268 "name": "spare", 00:14:33.268 "uuid": "a88541d9-452e-5793-a3ed-3bb8392daf8f", 00:14:33.268 "is_configured": true, 00:14:33.268 "data_offset": 2048, 00:14:33.268 "data_size": 63488 00:14:33.268 }, 00:14:33.268 { 00:14:33.268 "name": "BaseBdev2", 00:14:33.268 "uuid": "9a0be541-d5b7-5b3a-a946-e13a4325db1a", 00:14:33.268 "is_configured": true, 00:14:33.268 "data_offset": 2048, 00:14:33.268 "data_size": 63488 00:14:33.268 }, 00:14:33.268 { 00:14:33.268 "name": "BaseBdev3", 00:14:33.268 "uuid": "01335c00-2adc-5bc4-8d68-965b0b20c407", 00:14:33.268 "is_configured": true, 00:14:33.268 "data_offset": 2048, 00:14:33.268 "data_size": 63488 00:14:33.268 }, 00:14:33.268 { 00:14:33.268 "name": "BaseBdev4", 00:14:33.268 "uuid": "23b26d34-25c3-57d6-96d0-d4cd067bae4e", 00:14:33.268 "is_configured": true, 00:14:33.268 "data_offset": 2048, 00:14:33.268 "data_size": 63488 00:14:33.268 } 00:14:33.268 ] 00:14:33.268 }' 00:14:33.268 23:48:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:33.268 23:48:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:33.539 23:48:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@750 -- # verify_raid_bdev_process raid_bdev1 none none 00:14:33.539 23:48:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:33.539 23:48:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:14:33.539 23:48:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:14:33.539 23:48:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:33.539 23:48:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:33.539 23:48:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:33.539 23:48:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:33.539 23:48:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:33.539 23:48:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:33.539 23:48:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:33.539 "name": "raid_bdev1", 00:14:33.539 "uuid": "d8dd1718-7857-4e41-889d-86a3ec73047a", 00:14:33.539 "strip_size_kb": 64, 00:14:33.539 "state": "online", 00:14:33.539 "raid_level": "raid5f", 00:14:33.539 "superblock": true, 00:14:33.539 "num_base_bdevs": 4, 00:14:33.539 "num_base_bdevs_discovered": 4, 00:14:33.539 "num_base_bdevs_operational": 4, 00:14:33.539 "base_bdevs_list": [ 00:14:33.539 { 00:14:33.539 "name": "spare", 00:14:33.539 "uuid": "a88541d9-452e-5793-a3ed-3bb8392daf8f", 00:14:33.539 "is_configured": true, 00:14:33.539 "data_offset": 2048, 00:14:33.539 "data_size": 63488 00:14:33.539 }, 00:14:33.539 { 00:14:33.539 "name": "BaseBdev2", 00:14:33.539 "uuid": "9a0be541-d5b7-5b3a-a946-e13a4325db1a", 00:14:33.539 "is_configured": true, 00:14:33.539 "data_offset": 2048, 00:14:33.539 "data_size": 63488 00:14:33.539 }, 00:14:33.539 { 00:14:33.539 "name": "BaseBdev3", 00:14:33.539 "uuid": "01335c00-2adc-5bc4-8d68-965b0b20c407", 00:14:33.539 "is_configured": true, 00:14:33.539 "data_offset": 2048, 00:14:33.539 "data_size": 63488 00:14:33.539 }, 00:14:33.539 { 00:14:33.539 "name": "BaseBdev4", 00:14:33.539 "uuid": "23b26d34-25c3-57d6-96d0-d4cd067bae4e", 00:14:33.539 "is_configured": true, 00:14:33.539 "data_offset": 2048, 00:14:33.539 "data_size": 63488 00:14:33.539 } 00:14:33.539 ] 00:14:33.539 }' 00:14:33.539 23:48:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:33.799 23:48:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:14:33.799 23:48:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:33.799 23:48:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:14:33.799 23:48:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:33.799 23:48:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:33.799 23:48:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:33.799 23:48:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # jq -r '.[].base_bdevs_list[0].name' 00:14:33.799 23:48:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:33.799 23:48:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # [[ spare == \s\p\a\r\e ]] 00:14:33.799 23:48:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@754 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:14:33.799 23:48:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:33.799 23:48:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:33.799 [2024-11-26 23:48:21.815343] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:14:33.799 23:48:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:33.799 23:48:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@755 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:14:33.799 23:48:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:33.799 23:48:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:33.799 23:48:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:33.799 23:48:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:33.799 23:48:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:33.799 23:48:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:33.799 23:48:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:33.799 23:48:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:33.799 23:48:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:33.799 23:48:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:33.799 23:48:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:33.799 23:48:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:33.799 23:48:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:33.799 23:48:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:33.799 23:48:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:33.799 "name": "raid_bdev1", 00:14:33.799 "uuid": "d8dd1718-7857-4e41-889d-86a3ec73047a", 00:14:33.799 "strip_size_kb": 64, 00:14:33.799 "state": "online", 00:14:33.799 "raid_level": "raid5f", 00:14:33.799 "superblock": true, 00:14:33.799 "num_base_bdevs": 4, 00:14:33.799 "num_base_bdevs_discovered": 3, 00:14:33.799 "num_base_bdevs_operational": 3, 00:14:33.799 "base_bdevs_list": [ 00:14:33.799 { 00:14:33.799 "name": null, 00:14:33.799 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:33.799 "is_configured": false, 00:14:33.799 "data_offset": 0, 00:14:33.799 "data_size": 63488 00:14:33.799 }, 00:14:33.799 { 00:14:33.799 "name": "BaseBdev2", 00:14:33.799 "uuid": "9a0be541-d5b7-5b3a-a946-e13a4325db1a", 00:14:33.799 "is_configured": true, 00:14:33.799 "data_offset": 2048, 00:14:33.799 "data_size": 63488 00:14:33.799 }, 00:14:33.799 { 00:14:33.799 "name": "BaseBdev3", 00:14:33.799 "uuid": "01335c00-2adc-5bc4-8d68-965b0b20c407", 00:14:33.799 "is_configured": true, 00:14:33.799 "data_offset": 2048, 00:14:33.799 "data_size": 63488 00:14:33.799 }, 00:14:33.799 { 00:14:33.799 "name": "BaseBdev4", 00:14:33.799 "uuid": "23b26d34-25c3-57d6-96d0-d4cd067bae4e", 00:14:33.799 "is_configured": true, 00:14:33.799 "data_offset": 2048, 00:14:33.799 "data_size": 63488 00:14:33.799 } 00:14:33.799 ] 00:14:33.799 }' 00:14:33.799 23:48:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:33.799 23:48:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:34.369 23:48:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@756 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:14:34.369 23:48:22 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:34.369 23:48:22 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:34.369 [2024-11-26 23:48:22.230711] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:14:34.369 [2024-11-26 23:48:22.230895] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:14:34.369 [2024-11-26 23:48:22.230915] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:14:34.369 [2024-11-26 23:48:22.230955] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:14:34.369 [2024-11-26 23:48:22.234890] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000458f0 00:14:34.369 23:48:22 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:34.369 23:48:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@757 -- # sleep 1 00:14:34.369 [2024-11-26 23:48:22.236990] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:14:35.309 23:48:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@758 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:35.309 23:48:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:35.309 23:48:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:35.309 23:48:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:35.309 23:48:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:35.309 23:48:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:35.309 23:48:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:35.309 23:48:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:35.309 23:48:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:35.309 23:48:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:35.309 23:48:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:35.309 "name": "raid_bdev1", 00:14:35.309 "uuid": "d8dd1718-7857-4e41-889d-86a3ec73047a", 00:14:35.309 "strip_size_kb": 64, 00:14:35.309 "state": "online", 00:14:35.309 "raid_level": "raid5f", 00:14:35.309 "superblock": true, 00:14:35.309 "num_base_bdevs": 4, 00:14:35.309 "num_base_bdevs_discovered": 4, 00:14:35.309 "num_base_bdevs_operational": 4, 00:14:35.309 "process": { 00:14:35.309 "type": "rebuild", 00:14:35.309 "target": "spare", 00:14:35.309 "progress": { 00:14:35.309 "blocks": 19200, 00:14:35.309 "percent": 10 00:14:35.309 } 00:14:35.309 }, 00:14:35.309 "base_bdevs_list": [ 00:14:35.309 { 00:14:35.309 "name": "spare", 00:14:35.309 "uuid": "a88541d9-452e-5793-a3ed-3bb8392daf8f", 00:14:35.309 "is_configured": true, 00:14:35.309 "data_offset": 2048, 00:14:35.309 "data_size": 63488 00:14:35.309 }, 00:14:35.309 { 00:14:35.309 "name": "BaseBdev2", 00:14:35.309 "uuid": "9a0be541-d5b7-5b3a-a946-e13a4325db1a", 00:14:35.309 "is_configured": true, 00:14:35.309 "data_offset": 2048, 00:14:35.309 "data_size": 63488 00:14:35.309 }, 00:14:35.309 { 00:14:35.309 "name": "BaseBdev3", 00:14:35.309 "uuid": "01335c00-2adc-5bc4-8d68-965b0b20c407", 00:14:35.309 "is_configured": true, 00:14:35.309 "data_offset": 2048, 00:14:35.309 "data_size": 63488 00:14:35.309 }, 00:14:35.309 { 00:14:35.309 "name": "BaseBdev4", 00:14:35.309 "uuid": "23b26d34-25c3-57d6-96d0-d4cd067bae4e", 00:14:35.309 "is_configured": true, 00:14:35.309 "data_offset": 2048, 00:14:35.309 "data_size": 63488 00:14:35.309 } 00:14:35.309 ] 00:14:35.309 }' 00:14:35.309 23:48:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:35.309 23:48:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:35.309 23:48:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:35.309 23:48:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:35.309 23:48:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@761 -- # rpc_cmd bdev_passthru_delete spare 00:14:35.309 23:48:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:35.309 23:48:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:35.309 [2024-11-26 23:48:23.394130] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:14:35.571 [2024-11-26 23:48:23.442107] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:14:35.571 [2024-11-26 23:48:23.442179] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:35.571 [2024-11-26 23:48:23.442197] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:14:35.571 [2024-11-26 23:48:23.442205] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:14:35.571 23:48:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:35.571 23:48:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@762 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:14:35.571 23:48:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:35.571 23:48:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:35.571 23:48:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:35.571 23:48:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:35.571 23:48:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:35.571 23:48:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:35.571 23:48:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:35.571 23:48:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:35.571 23:48:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:35.571 23:48:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:35.571 23:48:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:35.571 23:48:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:35.571 23:48:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:35.571 23:48:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:35.571 23:48:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:35.571 "name": "raid_bdev1", 00:14:35.571 "uuid": "d8dd1718-7857-4e41-889d-86a3ec73047a", 00:14:35.571 "strip_size_kb": 64, 00:14:35.571 "state": "online", 00:14:35.571 "raid_level": "raid5f", 00:14:35.571 "superblock": true, 00:14:35.571 "num_base_bdevs": 4, 00:14:35.571 "num_base_bdevs_discovered": 3, 00:14:35.571 "num_base_bdevs_operational": 3, 00:14:35.571 "base_bdevs_list": [ 00:14:35.571 { 00:14:35.571 "name": null, 00:14:35.571 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:35.571 "is_configured": false, 00:14:35.571 "data_offset": 0, 00:14:35.571 "data_size": 63488 00:14:35.571 }, 00:14:35.571 { 00:14:35.571 "name": "BaseBdev2", 00:14:35.571 "uuid": "9a0be541-d5b7-5b3a-a946-e13a4325db1a", 00:14:35.571 "is_configured": true, 00:14:35.571 "data_offset": 2048, 00:14:35.571 "data_size": 63488 00:14:35.571 }, 00:14:35.571 { 00:14:35.571 "name": "BaseBdev3", 00:14:35.571 "uuid": "01335c00-2adc-5bc4-8d68-965b0b20c407", 00:14:35.571 "is_configured": true, 00:14:35.571 "data_offset": 2048, 00:14:35.571 "data_size": 63488 00:14:35.571 }, 00:14:35.571 { 00:14:35.571 "name": "BaseBdev4", 00:14:35.571 "uuid": "23b26d34-25c3-57d6-96d0-d4cd067bae4e", 00:14:35.571 "is_configured": true, 00:14:35.571 "data_offset": 2048, 00:14:35.571 "data_size": 63488 00:14:35.571 } 00:14:35.571 ] 00:14:35.571 }' 00:14:35.571 23:48:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:35.571 23:48:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:35.833 23:48:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@763 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:14:35.833 23:48:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:35.833 23:48:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:35.833 [2024-11-26 23:48:23.866611] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:14:35.833 [2024-11-26 23:48:23.866660] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:35.833 [2024-11-26 23:48:23.866694] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b780 00:14:35.833 [2024-11-26 23:48:23.866709] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:35.833 [2024-11-26 23:48:23.867128] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:35.833 [2024-11-26 23:48:23.867152] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:14:35.833 [2024-11-26 23:48:23.867233] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:14:35.833 [2024-11-26 23:48:23.867250] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:14:35.833 [2024-11-26 23:48:23.867262] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:14:35.833 [2024-11-26 23:48:23.867296] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:14:35.833 [2024-11-26 23:48:23.870573] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000459c0 00:14:35.833 spare 00:14:35.833 23:48:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:35.833 23:48:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@764 -- # sleep 1 00:14:35.833 [2024-11-26 23:48:23.872676] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:14:36.772 23:48:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@765 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:36.772 23:48:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:36.772 23:48:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:36.772 23:48:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:36.772 23:48:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:36.772 23:48:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:36.772 23:48:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:36.772 23:48:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:36.772 23:48:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:37.033 23:48:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:37.033 23:48:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:37.033 "name": "raid_bdev1", 00:14:37.033 "uuid": "d8dd1718-7857-4e41-889d-86a3ec73047a", 00:14:37.033 "strip_size_kb": 64, 00:14:37.033 "state": "online", 00:14:37.033 "raid_level": "raid5f", 00:14:37.033 "superblock": true, 00:14:37.033 "num_base_bdevs": 4, 00:14:37.033 "num_base_bdevs_discovered": 4, 00:14:37.033 "num_base_bdevs_operational": 4, 00:14:37.033 "process": { 00:14:37.033 "type": "rebuild", 00:14:37.033 "target": "spare", 00:14:37.033 "progress": { 00:14:37.033 "blocks": 19200, 00:14:37.033 "percent": 10 00:14:37.033 } 00:14:37.033 }, 00:14:37.033 "base_bdevs_list": [ 00:14:37.033 { 00:14:37.033 "name": "spare", 00:14:37.033 "uuid": "a88541d9-452e-5793-a3ed-3bb8392daf8f", 00:14:37.033 "is_configured": true, 00:14:37.033 "data_offset": 2048, 00:14:37.033 "data_size": 63488 00:14:37.033 }, 00:14:37.033 { 00:14:37.033 "name": "BaseBdev2", 00:14:37.033 "uuid": "9a0be541-d5b7-5b3a-a946-e13a4325db1a", 00:14:37.033 "is_configured": true, 00:14:37.033 "data_offset": 2048, 00:14:37.033 "data_size": 63488 00:14:37.033 }, 00:14:37.033 { 00:14:37.033 "name": "BaseBdev3", 00:14:37.033 "uuid": "01335c00-2adc-5bc4-8d68-965b0b20c407", 00:14:37.033 "is_configured": true, 00:14:37.033 "data_offset": 2048, 00:14:37.033 "data_size": 63488 00:14:37.033 }, 00:14:37.033 { 00:14:37.033 "name": "BaseBdev4", 00:14:37.033 "uuid": "23b26d34-25c3-57d6-96d0-d4cd067bae4e", 00:14:37.033 "is_configured": true, 00:14:37.033 "data_offset": 2048, 00:14:37.033 "data_size": 63488 00:14:37.033 } 00:14:37.033 ] 00:14:37.033 }' 00:14:37.033 23:48:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:37.033 23:48:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:37.033 23:48:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:37.033 23:48:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:37.033 23:48:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@768 -- # rpc_cmd bdev_passthru_delete spare 00:14:37.033 23:48:25 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:37.033 23:48:25 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:37.033 [2024-11-26 23:48:25.025826] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:14:37.033 [2024-11-26 23:48:25.077813] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:14:37.033 [2024-11-26 23:48:25.077867] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:37.033 [2024-11-26 23:48:25.077898] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:14:37.033 [2024-11-26 23:48:25.077906] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:14:37.033 23:48:25 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:37.033 23:48:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@769 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:14:37.033 23:48:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:37.033 23:48:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:37.033 23:48:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:37.033 23:48:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:37.033 23:48:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:37.033 23:48:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:37.033 23:48:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:37.033 23:48:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:37.033 23:48:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:37.033 23:48:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:37.033 23:48:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:37.033 23:48:25 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:37.033 23:48:25 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:37.033 23:48:25 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:37.033 23:48:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:37.033 "name": "raid_bdev1", 00:14:37.033 "uuid": "d8dd1718-7857-4e41-889d-86a3ec73047a", 00:14:37.033 "strip_size_kb": 64, 00:14:37.033 "state": "online", 00:14:37.033 "raid_level": "raid5f", 00:14:37.033 "superblock": true, 00:14:37.033 "num_base_bdevs": 4, 00:14:37.033 "num_base_bdevs_discovered": 3, 00:14:37.033 "num_base_bdevs_operational": 3, 00:14:37.033 "base_bdevs_list": [ 00:14:37.033 { 00:14:37.033 "name": null, 00:14:37.033 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:37.033 "is_configured": false, 00:14:37.033 "data_offset": 0, 00:14:37.033 "data_size": 63488 00:14:37.033 }, 00:14:37.033 { 00:14:37.033 "name": "BaseBdev2", 00:14:37.033 "uuid": "9a0be541-d5b7-5b3a-a946-e13a4325db1a", 00:14:37.033 "is_configured": true, 00:14:37.033 "data_offset": 2048, 00:14:37.033 "data_size": 63488 00:14:37.033 }, 00:14:37.033 { 00:14:37.033 "name": "BaseBdev3", 00:14:37.033 "uuid": "01335c00-2adc-5bc4-8d68-965b0b20c407", 00:14:37.033 "is_configured": true, 00:14:37.033 "data_offset": 2048, 00:14:37.033 "data_size": 63488 00:14:37.033 }, 00:14:37.033 { 00:14:37.033 "name": "BaseBdev4", 00:14:37.033 "uuid": "23b26d34-25c3-57d6-96d0-d4cd067bae4e", 00:14:37.033 "is_configured": true, 00:14:37.033 "data_offset": 2048, 00:14:37.033 "data_size": 63488 00:14:37.033 } 00:14:37.033 ] 00:14:37.033 }' 00:14:37.033 23:48:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:37.033 23:48:25 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:37.602 23:48:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@770 -- # verify_raid_bdev_process raid_bdev1 none none 00:14:37.603 23:48:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:37.603 23:48:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:14:37.603 23:48:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:14:37.603 23:48:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:37.603 23:48:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:37.603 23:48:25 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:37.603 23:48:25 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:37.603 23:48:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:37.603 23:48:25 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:37.603 23:48:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:37.603 "name": "raid_bdev1", 00:14:37.603 "uuid": "d8dd1718-7857-4e41-889d-86a3ec73047a", 00:14:37.603 "strip_size_kb": 64, 00:14:37.603 "state": "online", 00:14:37.603 "raid_level": "raid5f", 00:14:37.603 "superblock": true, 00:14:37.603 "num_base_bdevs": 4, 00:14:37.603 "num_base_bdevs_discovered": 3, 00:14:37.603 "num_base_bdevs_operational": 3, 00:14:37.603 "base_bdevs_list": [ 00:14:37.603 { 00:14:37.603 "name": null, 00:14:37.603 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:37.603 "is_configured": false, 00:14:37.603 "data_offset": 0, 00:14:37.603 "data_size": 63488 00:14:37.603 }, 00:14:37.603 { 00:14:37.603 "name": "BaseBdev2", 00:14:37.603 "uuid": "9a0be541-d5b7-5b3a-a946-e13a4325db1a", 00:14:37.603 "is_configured": true, 00:14:37.603 "data_offset": 2048, 00:14:37.603 "data_size": 63488 00:14:37.603 }, 00:14:37.603 { 00:14:37.603 "name": "BaseBdev3", 00:14:37.603 "uuid": "01335c00-2adc-5bc4-8d68-965b0b20c407", 00:14:37.603 "is_configured": true, 00:14:37.603 "data_offset": 2048, 00:14:37.603 "data_size": 63488 00:14:37.603 }, 00:14:37.603 { 00:14:37.603 "name": "BaseBdev4", 00:14:37.603 "uuid": "23b26d34-25c3-57d6-96d0-d4cd067bae4e", 00:14:37.603 "is_configured": true, 00:14:37.603 "data_offset": 2048, 00:14:37.603 "data_size": 63488 00:14:37.603 } 00:14:37.603 ] 00:14:37.603 }' 00:14:37.603 23:48:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:37.603 23:48:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:14:37.603 23:48:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:37.603 23:48:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:14:37.603 23:48:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@773 -- # rpc_cmd bdev_passthru_delete BaseBdev1 00:14:37.603 23:48:25 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:37.603 23:48:25 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:37.603 23:48:25 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:37.603 23:48:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@774 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:14:37.603 23:48:25 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:37.603 23:48:25 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:37.603 [2024-11-26 23:48:25.690035] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:14:37.603 [2024-11-26 23:48:25.690100] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:37.603 [2024-11-26 23:48:25.690118] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000bd80 00:14:37.603 [2024-11-26 23:48:25.690128] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:37.603 [2024-11-26 23:48:25.690543] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:37.603 [2024-11-26 23:48:25.690570] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:14:37.603 [2024-11-26 23:48:25.690645] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:14:37.603 [2024-11-26 23:48:25.690664] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:14:37.603 [2024-11-26 23:48:25.690671] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:14:37.603 [2024-11-26 23:48:25.690681] bdev_raid.c:3894:raid_bdev_examine_done: *ERROR*: Failed to examine bdev BaseBdev1: Invalid argument 00:14:37.603 BaseBdev1 00:14:37.603 23:48:25 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:37.603 23:48:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@775 -- # sleep 1 00:14:38.985 23:48:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@776 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:14:38.985 23:48:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:38.985 23:48:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:38.985 23:48:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:38.985 23:48:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:38.985 23:48:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:38.985 23:48:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:38.985 23:48:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:38.985 23:48:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:38.985 23:48:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:38.985 23:48:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:38.985 23:48:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:38.985 23:48:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:38.985 23:48:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:38.985 23:48:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:38.985 23:48:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:38.985 "name": "raid_bdev1", 00:14:38.985 "uuid": "d8dd1718-7857-4e41-889d-86a3ec73047a", 00:14:38.985 "strip_size_kb": 64, 00:14:38.985 "state": "online", 00:14:38.985 "raid_level": "raid5f", 00:14:38.985 "superblock": true, 00:14:38.985 "num_base_bdevs": 4, 00:14:38.985 "num_base_bdevs_discovered": 3, 00:14:38.985 "num_base_bdevs_operational": 3, 00:14:38.985 "base_bdevs_list": [ 00:14:38.985 { 00:14:38.985 "name": null, 00:14:38.985 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:38.985 "is_configured": false, 00:14:38.985 "data_offset": 0, 00:14:38.985 "data_size": 63488 00:14:38.985 }, 00:14:38.985 { 00:14:38.985 "name": "BaseBdev2", 00:14:38.985 "uuid": "9a0be541-d5b7-5b3a-a946-e13a4325db1a", 00:14:38.985 "is_configured": true, 00:14:38.985 "data_offset": 2048, 00:14:38.985 "data_size": 63488 00:14:38.985 }, 00:14:38.985 { 00:14:38.985 "name": "BaseBdev3", 00:14:38.985 "uuid": "01335c00-2adc-5bc4-8d68-965b0b20c407", 00:14:38.985 "is_configured": true, 00:14:38.985 "data_offset": 2048, 00:14:38.985 "data_size": 63488 00:14:38.985 }, 00:14:38.985 { 00:14:38.985 "name": "BaseBdev4", 00:14:38.985 "uuid": "23b26d34-25c3-57d6-96d0-d4cd067bae4e", 00:14:38.985 "is_configured": true, 00:14:38.985 "data_offset": 2048, 00:14:38.985 "data_size": 63488 00:14:38.985 } 00:14:38.985 ] 00:14:38.985 }' 00:14:38.985 23:48:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:38.985 23:48:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:39.245 23:48:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@777 -- # verify_raid_bdev_process raid_bdev1 none none 00:14:39.245 23:48:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:39.245 23:48:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:14:39.245 23:48:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:14:39.245 23:48:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:39.245 23:48:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:39.245 23:48:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:39.245 23:48:27 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:39.246 23:48:27 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:39.246 23:48:27 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:39.246 23:48:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:39.246 "name": "raid_bdev1", 00:14:39.246 "uuid": "d8dd1718-7857-4e41-889d-86a3ec73047a", 00:14:39.246 "strip_size_kb": 64, 00:14:39.246 "state": "online", 00:14:39.246 "raid_level": "raid5f", 00:14:39.246 "superblock": true, 00:14:39.246 "num_base_bdevs": 4, 00:14:39.246 "num_base_bdevs_discovered": 3, 00:14:39.246 "num_base_bdevs_operational": 3, 00:14:39.246 "base_bdevs_list": [ 00:14:39.246 { 00:14:39.246 "name": null, 00:14:39.246 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:39.246 "is_configured": false, 00:14:39.246 "data_offset": 0, 00:14:39.246 "data_size": 63488 00:14:39.246 }, 00:14:39.246 { 00:14:39.246 "name": "BaseBdev2", 00:14:39.246 "uuid": "9a0be541-d5b7-5b3a-a946-e13a4325db1a", 00:14:39.246 "is_configured": true, 00:14:39.246 "data_offset": 2048, 00:14:39.246 "data_size": 63488 00:14:39.246 }, 00:14:39.246 { 00:14:39.246 "name": "BaseBdev3", 00:14:39.246 "uuid": "01335c00-2adc-5bc4-8d68-965b0b20c407", 00:14:39.246 "is_configured": true, 00:14:39.246 "data_offset": 2048, 00:14:39.246 "data_size": 63488 00:14:39.246 }, 00:14:39.246 { 00:14:39.246 "name": "BaseBdev4", 00:14:39.246 "uuid": "23b26d34-25c3-57d6-96d0-d4cd067bae4e", 00:14:39.246 "is_configured": true, 00:14:39.246 "data_offset": 2048, 00:14:39.246 "data_size": 63488 00:14:39.246 } 00:14:39.246 ] 00:14:39.246 }' 00:14:39.246 23:48:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:39.246 23:48:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:14:39.246 23:48:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:39.246 23:48:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:14:39.246 23:48:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@778 -- # NOT rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:14:39.246 23:48:27 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@652 -- # local es=0 00:14:39.246 23:48:27 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:14:39.246 23:48:27 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:14:39.246 23:48:27 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:14:39.246 23:48:27 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:14:39.246 23:48:27 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:14:39.246 23:48:27 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:14:39.246 23:48:27 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:39.246 23:48:27 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:39.246 [2024-11-26 23:48:27.283334] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:14:39.246 [2024-11-26 23:48:27.283488] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:14:39.246 [2024-11-26 23:48:27.283503] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:14:39.246 request: 00:14:39.246 { 00:14:39.246 "base_bdev": "BaseBdev1", 00:14:39.246 "raid_bdev": "raid_bdev1", 00:14:39.246 "method": "bdev_raid_add_base_bdev", 00:14:39.246 "req_id": 1 00:14:39.246 } 00:14:39.246 Got JSON-RPC error response 00:14:39.246 response: 00:14:39.246 { 00:14:39.246 "code": -22, 00:14:39.246 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:14:39.246 } 00:14:39.246 23:48:27 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:14:39.246 23:48:27 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@655 -- # es=1 00:14:39.246 23:48:27 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:14:39.246 23:48:27 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:14:39.246 23:48:27 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:14:39.246 23:48:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@779 -- # sleep 1 00:14:40.186 23:48:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@780 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:14:40.186 23:48:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:40.186 23:48:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:40.186 23:48:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:40.186 23:48:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:40.186 23:48:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:40.186 23:48:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:40.186 23:48:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:40.186 23:48:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:40.186 23:48:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:40.186 23:48:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:40.186 23:48:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:40.186 23:48:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:40.186 23:48:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:40.447 23:48:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:40.447 23:48:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:40.447 "name": "raid_bdev1", 00:14:40.447 "uuid": "d8dd1718-7857-4e41-889d-86a3ec73047a", 00:14:40.447 "strip_size_kb": 64, 00:14:40.447 "state": "online", 00:14:40.447 "raid_level": "raid5f", 00:14:40.447 "superblock": true, 00:14:40.447 "num_base_bdevs": 4, 00:14:40.447 "num_base_bdevs_discovered": 3, 00:14:40.447 "num_base_bdevs_operational": 3, 00:14:40.447 "base_bdevs_list": [ 00:14:40.447 { 00:14:40.447 "name": null, 00:14:40.447 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:40.447 "is_configured": false, 00:14:40.447 "data_offset": 0, 00:14:40.447 "data_size": 63488 00:14:40.447 }, 00:14:40.447 { 00:14:40.447 "name": "BaseBdev2", 00:14:40.447 "uuid": "9a0be541-d5b7-5b3a-a946-e13a4325db1a", 00:14:40.447 "is_configured": true, 00:14:40.447 "data_offset": 2048, 00:14:40.447 "data_size": 63488 00:14:40.447 }, 00:14:40.447 { 00:14:40.447 "name": "BaseBdev3", 00:14:40.447 "uuid": "01335c00-2adc-5bc4-8d68-965b0b20c407", 00:14:40.447 "is_configured": true, 00:14:40.447 "data_offset": 2048, 00:14:40.447 "data_size": 63488 00:14:40.447 }, 00:14:40.447 { 00:14:40.447 "name": "BaseBdev4", 00:14:40.447 "uuid": "23b26d34-25c3-57d6-96d0-d4cd067bae4e", 00:14:40.447 "is_configured": true, 00:14:40.447 "data_offset": 2048, 00:14:40.447 "data_size": 63488 00:14:40.447 } 00:14:40.447 ] 00:14:40.447 }' 00:14:40.447 23:48:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:40.447 23:48:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:40.707 23:48:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@781 -- # verify_raid_bdev_process raid_bdev1 none none 00:14:40.707 23:48:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:40.708 23:48:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:14:40.708 23:48:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:14:40.708 23:48:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:40.708 23:48:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:40.708 23:48:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:40.708 23:48:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:40.708 23:48:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:40.708 23:48:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:40.708 23:48:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:40.708 "name": "raid_bdev1", 00:14:40.708 "uuid": "d8dd1718-7857-4e41-889d-86a3ec73047a", 00:14:40.708 "strip_size_kb": 64, 00:14:40.708 "state": "online", 00:14:40.708 "raid_level": "raid5f", 00:14:40.708 "superblock": true, 00:14:40.708 "num_base_bdevs": 4, 00:14:40.708 "num_base_bdevs_discovered": 3, 00:14:40.708 "num_base_bdevs_operational": 3, 00:14:40.708 "base_bdevs_list": [ 00:14:40.708 { 00:14:40.708 "name": null, 00:14:40.708 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:40.708 "is_configured": false, 00:14:40.708 "data_offset": 0, 00:14:40.708 "data_size": 63488 00:14:40.708 }, 00:14:40.708 { 00:14:40.708 "name": "BaseBdev2", 00:14:40.708 "uuid": "9a0be541-d5b7-5b3a-a946-e13a4325db1a", 00:14:40.708 "is_configured": true, 00:14:40.708 "data_offset": 2048, 00:14:40.708 "data_size": 63488 00:14:40.708 }, 00:14:40.708 { 00:14:40.708 "name": "BaseBdev3", 00:14:40.708 "uuid": "01335c00-2adc-5bc4-8d68-965b0b20c407", 00:14:40.708 "is_configured": true, 00:14:40.708 "data_offset": 2048, 00:14:40.708 "data_size": 63488 00:14:40.708 }, 00:14:40.708 { 00:14:40.708 "name": "BaseBdev4", 00:14:40.708 "uuid": "23b26d34-25c3-57d6-96d0-d4cd067bae4e", 00:14:40.708 "is_configured": true, 00:14:40.708 "data_offset": 2048, 00:14:40.708 "data_size": 63488 00:14:40.708 } 00:14:40.708 ] 00:14:40.708 }' 00:14:40.708 23:48:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:40.708 23:48:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:14:40.708 23:48:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:40.983 23:48:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:14:40.983 23:48:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@784 -- # killprocess 95180 00:14:40.983 23:48:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@954 -- # '[' -z 95180 ']' 00:14:40.983 23:48:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@958 -- # kill -0 95180 00:14:40.983 23:48:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@959 -- # uname 00:14:40.983 23:48:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:14:40.983 23:48:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 95180 00:14:40.983 23:48:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:14:40.983 23:48:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:14:40.983 killing process with pid 95180 00:14:40.983 23:48:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@972 -- # echo 'killing process with pid 95180' 00:14:40.983 23:48:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@973 -- # kill 95180 00:14:40.983 Received shutdown signal, test time was about 60.000000 seconds 00:14:40.983 00:14:40.983 Latency(us) 00:14:40.984 [2024-11-26T23:48:29.116Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:14:40.984 [2024-11-26T23:48:29.116Z] =================================================================================================================== 00:14:40.984 [2024-11-26T23:48:29.116Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:14:40.984 [2024-11-26 23:48:28.907856] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:14:40.984 [2024-11-26 23:48:28.907963] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:14:40.984 23:48:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@978 -- # wait 95180 00:14:40.984 [2024-11-26 23:48:28.908043] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:14:40.984 [2024-11-26 23:48:28.908052] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state offline 00:14:40.984 [2024-11-26 23:48:28.956741] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:14:41.249 23:48:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@786 -- # return 0 00:14:41.249 00:14:41.249 real 0m24.798s 00:14:41.249 user 0m31.512s 00:14:41.249 sys 0m2.852s 00:14:41.249 23:48:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@1130 -- # xtrace_disable 00:14:41.249 23:48:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:41.249 ************************************ 00:14:41.249 END TEST raid5f_rebuild_test_sb 00:14:41.249 ************************************ 00:14:41.249 23:48:29 bdev_raid -- bdev/bdev_raid.sh@995 -- # base_blocklen=4096 00:14:41.249 23:48:29 bdev_raid -- bdev/bdev_raid.sh@997 -- # run_test raid_state_function_test_sb_4k raid_state_function_test raid1 2 true 00:14:41.249 23:48:29 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:14:41.249 23:48:29 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:14:41.249 23:48:29 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:14:41.249 ************************************ 00:14:41.249 START TEST raid_state_function_test_sb_4k 00:14:41.249 ************************************ 00:14:41.249 23:48:29 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@1129 -- # raid_state_function_test raid1 2 true 00:14:41.249 23:48:29 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@205 -- # local raid_level=raid1 00:14:41.249 23:48:29 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=2 00:14:41.249 23:48:29 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:14:41.249 23:48:29 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:14:41.249 23:48:29 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:14:41.249 23:48:29 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:14:41.249 23:48:29 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:14:41.249 23:48:29 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:14:41.249 23:48:29 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:14:41.249 23:48:29 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:14:41.249 23:48:29 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:14:41.249 23:48:29 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:14:41.249 23:48:29 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:14:41.249 23:48:29 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:14:41.249 23:48:29 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:14:41.249 23:48:29 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@211 -- # local strip_size 00:14:41.249 23:48:29 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:14:41.249 23:48:29 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:14:41.249 23:48:29 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@215 -- # '[' raid1 '!=' raid1 ']' 00:14:41.249 23:48:29 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@219 -- # strip_size=0 00:14:41.249 23:48:29 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:14:41.249 23:48:29 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:14:41.249 23:48:29 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@229 -- # raid_pid=95968 00:14:41.249 23:48:29 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:14:41.249 Process raid pid: 95968 00:14:41.249 23:48:29 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 95968' 00:14:41.249 23:48:29 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@231 -- # waitforlisten 95968 00:14:41.249 23:48:29 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@835 -- # '[' -z 95968 ']' 00:14:41.249 23:48:29 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:14:41.249 23:48:29 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@840 -- # local max_retries=100 00:14:41.249 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:14:41.249 23:48:29 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:14:41.249 23:48:29 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@844 -- # xtrace_disable 00:14:41.249 23:48:29 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:14:41.249 [2024-11-26 23:48:29.304595] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:14:41.249 [2024-11-26 23:48:29.304713] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:14:41.508 [2024-11-26 23:48:29.457393] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:14:41.508 [2024-11-26 23:48:29.481326] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:14:41.508 [2024-11-26 23:48:29.522070] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:14:41.508 [2024-11-26 23:48:29.522107] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:14:42.079 23:48:30 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:14:42.079 23:48:30 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@868 -- # return 0 00:14:42.079 23:48:30 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:14:42.079 23:48:30 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:42.079 23:48:30 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:14:42.079 [2024-11-26 23:48:30.119810] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:14:42.079 [2024-11-26 23:48:30.119866] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:14:42.079 [2024-11-26 23:48:30.119876] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:14:42.079 [2024-11-26 23:48:30.119886] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:14:42.079 23:48:30 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:42.079 23:48:30 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:14:42.079 23:48:30 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:42.079 23:48:30 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:42.079 23:48:30 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:14:42.079 23:48:30 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:14:42.079 23:48:30 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:14:42.079 23:48:30 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:42.079 23:48:30 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:42.079 23:48:30 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:42.079 23:48:30 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:42.079 23:48:30 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:42.079 23:48:30 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:42.079 23:48:30 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:42.079 23:48:30 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:14:42.079 23:48:30 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:42.079 23:48:30 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:42.079 "name": "Existed_Raid", 00:14:42.079 "uuid": "f021da05-429c-4927-afb6-3874415c98b0", 00:14:42.079 "strip_size_kb": 0, 00:14:42.079 "state": "configuring", 00:14:42.079 "raid_level": "raid1", 00:14:42.079 "superblock": true, 00:14:42.079 "num_base_bdevs": 2, 00:14:42.079 "num_base_bdevs_discovered": 0, 00:14:42.079 "num_base_bdevs_operational": 2, 00:14:42.079 "base_bdevs_list": [ 00:14:42.079 { 00:14:42.079 "name": "BaseBdev1", 00:14:42.079 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:42.079 "is_configured": false, 00:14:42.079 "data_offset": 0, 00:14:42.079 "data_size": 0 00:14:42.079 }, 00:14:42.079 { 00:14:42.079 "name": "BaseBdev2", 00:14:42.079 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:42.079 "is_configured": false, 00:14:42.079 "data_offset": 0, 00:14:42.079 "data_size": 0 00:14:42.079 } 00:14:42.079 ] 00:14:42.079 }' 00:14:42.079 23:48:30 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:42.079 23:48:30 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:14:42.650 23:48:30 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:14:42.650 23:48:30 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:42.650 23:48:30 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:14:42.650 [2024-11-26 23:48:30.590907] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:14:42.650 [2024-11-26 23:48:30.590949] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:14:42.650 23:48:30 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:42.650 23:48:30 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:14:42.650 23:48:30 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:42.650 23:48:30 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:14:42.650 [2024-11-26 23:48:30.602893] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:14:42.650 [2024-11-26 23:48:30.602932] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:14:42.650 [2024-11-26 23:48:30.602955] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:14:42.650 [2024-11-26 23:48:30.602973] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:14:42.650 23:48:30 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:42.650 23:48:30 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 4096 -b BaseBdev1 00:14:42.650 23:48:30 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:42.650 23:48:30 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:14:42.650 [2024-11-26 23:48:30.623455] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:14:42.650 BaseBdev1 00:14:42.650 23:48:30 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:42.650 23:48:30 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:14:42.650 23:48:30 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:14:42.650 23:48:30 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:14:42.650 23:48:30 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@905 -- # local i 00:14:42.650 23:48:30 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:14:42.650 23:48:30 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:14:42.650 23:48:30 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:14:42.650 23:48:30 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:42.650 23:48:30 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:14:42.650 23:48:30 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:42.650 23:48:30 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:14:42.650 23:48:30 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:42.650 23:48:30 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:14:42.650 [ 00:14:42.650 { 00:14:42.650 "name": "BaseBdev1", 00:14:42.650 "aliases": [ 00:14:42.650 "6ae67878-967c-4c18-ba6c-a0793b78c9c8" 00:14:42.650 ], 00:14:42.650 "product_name": "Malloc disk", 00:14:42.650 "block_size": 4096, 00:14:42.650 "num_blocks": 8192, 00:14:42.650 "uuid": "6ae67878-967c-4c18-ba6c-a0793b78c9c8", 00:14:42.650 "assigned_rate_limits": { 00:14:42.650 "rw_ios_per_sec": 0, 00:14:42.650 "rw_mbytes_per_sec": 0, 00:14:42.650 "r_mbytes_per_sec": 0, 00:14:42.650 "w_mbytes_per_sec": 0 00:14:42.650 }, 00:14:42.650 "claimed": true, 00:14:42.650 "claim_type": "exclusive_write", 00:14:42.650 "zoned": false, 00:14:42.650 "supported_io_types": { 00:14:42.650 "read": true, 00:14:42.650 "write": true, 00:14:42.650 "unmap": true, 00:14:42.650 "flush": true, 00:14:42.650 "reset": true, 00:14:42.650 "nvme_admin": false, 00:14:42.650 "nvme_io": false, 00:14:42.650 "nvme_io_md": false, 00:14:42.650 "write_zeroes": true, 00:14:42.650 "zcopy": true, 00:14:42.650 "get_zone_info": false, 00:14:42.650 "zone_management": false, 00:14:42.650 "zone_append": false, 00:14:42.650 "compare": false, 00:14:42.650 "compare_and_write": false, 00:14:42.650 "abort": true, 00:14:42.650 "seek_hole": false, 00:14:42.650 "seek_data": false, 00:14:42.650 "copy": true, 00:14:42.650 "nvme_iov_md": false 00:14:42.650 }, 00:14:42.650 "memory_domains": [ 00:14:42.650 { 00:14:42.650 "dma_device_id": "system", 00:14:42.650 "dma_device_type": 1 00:14:42.650 }, 00:14:42.650 { 00:14:42.650 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:42.650 "dma_device_type": 2 00:14:42.650 } 00:14:42.650 ], 00:14:42.650 "driver_specific": {} 00:14:42.650 } 00:14:42.650 ] 00:14:42.650 23:48:30 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:42.650 23:48:30 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@911 -- # return 0 00:14:42.650 23:48:30 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:14:42.650 23:48:30 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:42.650 23:48:30 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:42.650 23:48:30 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:14:42.650 23:48:30 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:14:42.650 23:48:30 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:14:42.650 23:48:30 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:42.650 23:48:30 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:42.650 23:48:30 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:42.650 23:48:30 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:42.650 23:48:30 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:42.650 23:48:30 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:42.650 23:48:30 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:42.650 23:48:30 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:14:42.650 23:48:30 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:42.650 23:48:30 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:42.650 "name": "Existed_Raid", 00:14:42.650 "uuid": "40c22859-82a1-409d-a095-167dc92124eb", 00:14:42.650 "strip_size_kb": 0, 00:14:42.650 "state": "configuring", 00:14:42.650 "raid_level": "raid1", 00:14:42.650 "superblock": true, 00:14:42.650 "num_base_bdevs": 2, 00:14:42.650 "num_base_bdevs_discovered": 1, 00:14:42.650 "num_base_bdevs_operational": 2, 00:14:42.650 "base_bdevs_list": [ 00:14:42.650 { 00:14:42.650 "name": "BaseBdev1", 00:14:42.650 "uuid": "6ae67878-967c-4c18-ba6c-a0793b78c9c8", 00:14:42.650 "is_configured": true, 00:14:42.651 "data_offset": 256, 00:14:42.651 "data_size": 7936 00:14:42.651 }, 00:14:42.651 { 00:14:42.651 "name": "BaseBdev2", 00:14:42.651 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:42.651 "is_configured": false, 00:14:42.651 "data_offset": 0, 00:14:42.651 "data_size": 0 00:14:42.651 } 00:14:42.651 ] 00:14:42.651 }' 00:14:42.651 23:48:30 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:42.651 23:48:30 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:14:43.222 23:48:31 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:14:43.222 23:48:31 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:43.222 23:48:31 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:14:43.222 [2024-11-26 23:48:31.102653] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:14:43.222 [2024-11-26 23:48:31.102711] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:14:43.222 23:48:31 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:43.222 23:48:31 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:14:43.222 23:48:31 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:43.222 23:48:31 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:14:43.222 [2024-11-26 23:48:31.114655] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:14:43.222 [2024-11-26 23:48:31.116529] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:14:43.222 [2024-11-26 23:48:31.116565] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:14:43.222 23:48:31 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:43.222 23:48:31 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:14:43.222 23:48:31 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:14:43.222 23:48:31 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:14:43.222 23:48:31 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:43.222 23:48:31 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:43.222 23:48:31 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:14:43.222 23:48:31 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:14:43.222 23:48:31 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:14:43.222 23:48:31 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:43.222 23:48:31 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:43.222 23:48:31 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:43.222 23:48:31 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:43.222 23:48:31 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:43.222 23:48:31 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:43.222 23:48:31 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:43.222 23:48:31 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:14:43.222 23:48:31 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:43.222 23:48:31 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:43.222 "name": "Existed_Raid", 00:14:43.222 "uuid": "07ee2102-3cbe-4069-b2b6-05f879f21453", 00:14:43.222 "strip_size_kb": 0, 00:14:43.222 "state": "configuring", 00:14:43.222 "raid_level": "raid1", 00:14:43.222 "superblock": true, 00:14:43.222 "num_base_bdevs": 2, 00:14:43.222 "num_base_bdevs_discovered": 1, 00:14:43.222 "num_base_bdevs_operational": 2, 00:14:43.222 "base_bdevs_list": [ 00:14:43.222 { 00:14:43.222 "name": "BaseBdev1", 00:14:43.222 "uuid": "6ae67878-967c-4c18-ba6c-a0793b78c9c8", 00:14:43.222 "is_configured": true, 00:14:43.222 "data_offset": 256, 00:14:43.222 "data_size": 7936 00:14:43.222 }, 00:14:43.222 { 00:14:43.222 "name": "BaseBdev2", 00:14:43.222 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:43.222 "is_configured": false, 00:14:43.222 "data_offset": 0, 00:14:43.222 "data_size": 0 00:14:43.222 } 00:14:43.222 ] 00:14:43.222 }' 00:14:43.222 23:48:31 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:43.222 23:48:31 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:14:43.481 23:48:31 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 4096 -b BaseBdev2 00:14:43.481 23:48:31 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:43.481 23:48:31 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:14:43.481 [2024-11-26 23:48:31.588580] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:14:43.481 [2024-11-26 23:48:31.588762] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:14:43.481 [2024-11-26 23:48:31.588777] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:14:43.481 [2024-11-26 23:48:31.589059] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:14:43.481 BaseBdev2 00:14:43.481 [2024-11-26 23:48:31.589233] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:14:43.481 [2024-11-26 23:48:31.589247] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:14:43.481 [2024-11-26 23:48:31.589382] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:43.481 23:48:31 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:43.481 23:48:31 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:14:43.481 23:48:31 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:14:43.481 23:48:31 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:14:43.481 23:48:31 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@905 -- # local i 00:14:43.481 23:48:31 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:14:43.481 23:48:31 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:14:43.481 23:48:31 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:14:43.481 23:48:31 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:43.481 23:48:31 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:14:43.481 23:48:31 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:43.481 23:48:31 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:14:43.481 23:48:31 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:43.481 23:48:31 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:14:43.740 [ 00:14:43.740 { 00:14:43.740 "name": "BaseBdev2", 00:14:43.740 "aliases": [ 00:14:43.740 "7c931e43-a1d8-40ef-9c7c-97b75c1e542d" 00:14:43.740 ], 00:14:43.740 "product_name": "Malloc disk", 00:14:43.740 "block_size": 4096, 00:14:43.740 "num_blocks": 8192, 00:14:43.740 "uuid": "7c931e43-a1d8-40ef-9c7c-97b75c1e542d", 00:14:43.740 "assigned_rate_limits": { 00:14:43.740 "rw_ios_per_sec": 0, 00:14:43.740 "rw_mbytes_per_sec": 0, 00:14:43.740 "r_mbytes_per_sec": 0, 00:14:43.740 "w_mbytes_per_sec": 0 00:14:43.740 }, 00:14:43.740 "claimed": true, 00:14:43.740 "claim_type": "exclusive_write", 00:14:43.740 "zoned": false, 00:14:43.740 "supported_io_types": { 00:14:43.740 "read": true, 00:14:43.740 "write": true, 00:14:43.740 "unmap": true, 00:14:43.740 "flush": true, 00:14:43.740 "reset": true, 00:14:43.740 "nvme_admin": false, 00:14:43.740 "nvme_io": false, 00:14:43.740 "nvme_io_md": false, 00:14:43.740 "write_zeroes": true, 00:14:43.740 "zcopy": true, 00:14:43.740 "get_zone_info": false, 00:14:43.740 "zone_management": false, 00:14:43.740 "zone_append": false, 00:14:43.740 "compare": false, 00:14:43.740 "compare_and_write": false, 00:14:43.740 "abort": true, 00:14:43.740 "seek_hole": false, 00:14:43.740 "seek_data": false, 00:14:43.740 "copy": true, 00:14:43.740 "nvme_iov_md": false 00:14:43.740 }, 00:14:43.740 "memory_domains": [ 00:14:43.740 { 00:14:43.740 "dma_device_id": "system", 00:14:43.740 "dma_device_type": 1 00:14:43.740 }, 00:14:43.740 { 00:14:43.740 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:43.740 "dma_device_type": 2 00:14:43.740 } 00:14:43.740 ], 00:14:43.740 "driver_specific": {} 00:14:43.740 } 00:14:43.740 ] 00:14:43.740 23:48:31 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:43.740 23:48:31 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@911 -- # return 0 00:14:43.740 23:48:31 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:14:43.740 23:48:31 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:14:43.740 23:48:31 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid1 0 2 00:14:43.740 23:48:31 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:43.740 23:48:31 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:43.740 23:48:31 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:14:43.740 23:48:31 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:14:43.740 23:48:31 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:14:43.740 23:48:31 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:43.740 23:48:31 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:43.740 23:48:31 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:43.740 23:48:31 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:43.740 23:48:31 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:43.740 23:48:31 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:43.740 23:48:31 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:14:43.740 23:48:31 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:43.740 23:48:31 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:43.740 23:48:31 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:43.740 "name": "Existed_Raid", 00:14:43.740 "uuid": "07ee2102-3cbe-4069-b2b6-05f879f21453", 00:14:43.740 "strip_size_kb": 0, 00:14:43.740 "state": "online", 00:14:43.740 "raid_level": "raid1", 00:14:43.740 "superblock": true, 00:14:43.740 "num_base_bdevs": 2, 00:14:43.740 "num_base_bdevs_discovered": 2, 00:14:43.740 "num_base_bdevs_operational": 2, 00:14:43.740 "base_bdevs_list": [ 00:14:43.740 { 00:14:43.740 "name": "BaseBdev1", 00:14:43.740 "uuid": "6ae67878-967c-4c18-ba6c-a0793b78c9c8", 00:14:43.740 "is_configured": true, 00:14:43.740 "data_offset": 256, 00:14:43.740 "data_size": 7936 00:14:43.740 }, 00:14:43.740 { 00:14:43.740 "name": "BaseBdev2", 00:14:43.740 "uuid": "7c931e43-a1d8-40ef-9c7c-97b75c1e542d", 00:14:43.740 "is_configured": true, 00:14:43.740 "data_offset": 256, 00:14:43.740 "data_size": 7936 00:14:43.740 } 00:14:43.740 ] 00:14:43.740 }' 00:14:43.740 23:48:31 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:43.740 23:48:31 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:14:44.000 23:48:32 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:14:44.000 23:48:32 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:14:44.000 23:48:32 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:14:44.000 23:48:32 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:14:44.000 23:48:32 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@184 -- # local name 00:14:44.000 23:48:32 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:14:44.000 23:48:32 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:14:44.000 23:48:32 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:44.000 23:48:32 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:14:44.000 23:48:32 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:14:44.000 [2024-11-26 23:48:32.100044] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:14:44.000 23:48:32 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:44.262 23:48:32 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:14:44.262 "name": "Existed_Raid", 00:14:44.262 "aliases": [ 00:14:44.262 "07ee2102-3cbe-4069-b2b6-05f879f21453" 00:14:44.262 ], 00:14:44.262 "product_name": "Raid Volume", 00:14:44.262 "block_size": 4096, 00:14:44.262 "num_blocks": 7936, 00:14:44.262 "uuid": "07ee2102-3cbe-4069-b2b6-05f879f21453", 00:14:44.262 "assigned_rate_limits": { 00:14:44.262 "rw_ios_per_sec": 0, 00:14:44.262 "rw_mbytes_per_sec": 0, 00:14:44.262 "r_mbytes_per_sec": 0, 00:14:44.262 "w_mbytes_per_sec": 0 00:14:44.262 }, 00:14:44.262 "claimed": false, 00:14:44.262 "zoned": false, 00:14:44.262 "supported_io_types": { 00:14:44.262 "read": true, 00:14:44.262 "write": true, 00:14:44.262 "unmap": false, 00:14:44.262 "flush": false, 00:14:44.262 "reset": true, 00:14:44.262 "nvme_admin": false, 00:14:44.262 "nvme_io": false, 00:14:44.262 "nvme_io_md": false, 00:14:44.262 "write_zeroes": true, 00:14:44.262 "zcopy": false, 00:14:44.262 "get_zone_info": false, 00:14:44.262 "zone_management": false, 00:14:44.262 "zone_append": false, 00:14:44.262 "compare": false, 00:14:44.262 "compare_and_write": false, 00:14:44.262 "abort": false, 00:14:44.262 "seek_hole": false, 00:14:44.262 "seek_data": false, 00:14:44.262 "copy": false, 00:14:44.262 "nvme_iov_md": false 00:14:44.262 }, 00:14:44.262 "memory_domains": [ 00:14:44.262 { 00:14:44.262 "dma_device_id": "system", 00:14:44.262 "dma_device_type": 1 00:14:44.262 }, 00:14:44.262 { 00:14:44.262 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:44.262 "dma_device_type": 2 00:14:44.262 }, 00:14:44.262 { 00:14:44.262 "dma_device_id": "system", 00:14:44.262 "dma_device_type": 1 00:14:44.262 }, 00:14:44.262 { 00:14:44.262 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:44.262 "dma_device_type": 2 00:14:44.262 } 00:14:44.262 ], 00:14:44.262 "driver_specific": { 00:14:44.262 "raid": { 00:14:44.262 "uuid": "07ee2102-3cbe-4069-b2b6-05f879f21453", 00:14:44.262 "strip_size_kb": 0, 00:14:44.262 "state": "online", 00:14:44.263 "raid_level": "raid1", 00:14:44.263 "superblock": true, 00:14:44.263 "num_base_bdevs": 2, 00:14:44.263 "num_base_bdevs_discovered": 2, 00:14:44.263 "num_base_bdevs_operational": 2, 00:14:44.263 "base_bdevs_list": [ 00:14:44.263 { 00:14:44.263 "name": "BaseBdev1", 00:14:44.263 "uuid": "6ae67878-967c-4c18-ba6c-a0793b78c9c8", 00:14:44.263 "is_configured": true, 00:14:44.263 "data_offset": 256, 00:14:44.263 "data_size": 7936 00:14:44.263 }, 00:14:44.263 { 00:14:44.263 "name": "BaseBdev2", 00:14:44.263 "uuid": "7c931e43-a1d8-40ef-9c7c-97b75c1e542d", 00:14:44.263 "is_configured": true, 00:14:44.263 "data_offset": 256, 00:14:44.263 "data_size": 7936 00:14:44.263 } 00:14:44.263 ] 00:14:44.263 } 00:14:44.263 } 00:14:44.263 }' 00:14:44.263 23:48:32 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:14:44.263 23:48:32 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:14:44.263 BaseBdev2' 00:14:44.263 23:48:32 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:44.263 23:48:32 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='4096 ' 00:14:44.263 23:48:32 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:44.263 23:48:32 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:44.263 23:48:32 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:14:44.263 23:48:32 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:44.263 23:48:32 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:14:44.263 23:48:32 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:44.263 23:48:32 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 ' 00:14:44.263 23:48:32 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@193 -- # [[ 4096 == \4\0\9\6\ \ \ ]] 00:14:44.263 23:48:32 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:44.263 23:48:32 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:14:44.263 23:48:32 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:44.263 23:48:32 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:44.263 23:48:32 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:14:44.263 23:48:32 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:44.263 23:48:32 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 ' 00:14:44.263 23:48:32 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@193 -- # [[ 4096 == \4\0\9\6\ \ \ ]] 00:14:44.263 23:48:32 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:14:44.263 23:48:32 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:44.263 23:48:32 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:14:44.263 [2024-11-26 23:48:32.299483] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:14:44.263 23:48:32 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:44.263 23:48:32 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@260 -- # local expected_state 00:14:44.263 23:48:32 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@261 -- # has_redundancy raid1 00:14:44.263 23:48:32 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@198 -- # case $1 in 00:14:44.263 23:48:32 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@199 -- # return 0 00:14:44.263 23:48:32 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:14:44.263 23:48:32 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid1 0 1 00:14:44.263 23:48:32 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:44.263 23:48:32 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:44.263 23:48:32 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:14:44.263 23:48:32 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:14:44.263 23:48:32 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:14:44.263 23:48:32 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:44.263 23:48:32 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:44.263 23:48:32 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:44.263 23:48:32 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:44.263 23:48:32 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:44.263 23:48:32 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:44.263 23:48:32 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:44.263 23:48:32 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:14:44.263 23:48:32 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:44.263 23:48:32 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:44.263 "name": "Existed_Raid", 00:14:44.263 "uuid": "07ee2102-3cbe-4069-b2b6-05f879f21453", 00:14:44.263 "strip_size_kb": 0, 00:14:44.263 "state": "online", 00:14:44.263 "raid_level": "raid1", 00:14:44.263 "superblock": true, 00:14:44.263 "num_base_bdevs": 2, 00:14:44.263 "num_base_bdevs_discovered": 1, 00:14:44.263 "num_base_bdevs_operational": 1, 00:14:44.263 "base_bdevs_list": [ 00:14:44.263 { 00:14:44.263 "name": null, 00:14:44.263 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:44.263 "is_configured": false, 00:14:44.263 "data_offset": 0, 00:14:44.263 "data_size": 7936 00:14:44.263 }, 00:14:44.263 { 00:14:44.263 "name": "BaseBdev2", 00:14:44.263 "uuid": "7c931e43-a1d8-40ef-9c7c-97b75c1e542d", 00:14:44.263 "is_configured": true, 00:14:44.263 "data_offset": 256, 00:14:44.263 "data_size": 7936 00:14:44.263 } 00:14:44.263 ] 00:14:44.263 }' 00:14:44.263 23:48:32 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:44.263 23:48:32 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:14:44.833 23:48:32 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:14:44.833 23:48:32 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:14:44.833 23:48:32 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:44.833 23:48:32 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:14:44.833 23:48:32 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:44.833 23:48:32 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:14:44.833 23:48:32 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:44.833 23:48:32 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:14:44.833 23:48:32 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:14:44.833 23:48:32 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:14:44.833 23:48:32 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:44.833 23:48:32 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:14:44.833 [2024-11-26 23:48:32.825732] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:14:44.833 [2024-11-26 23:48:32.825871] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:14:44.833 [2024-11-26 23:48:32.837419] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:14:44.833 [2024-11-26 23:48:32.837469] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:14:44.833 [2024-11-26 23:48:32.837480] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:14:44.833 23:48:32 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:44.833 23:48:32 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:14:44.833 23:48:32 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:14:44.833 23:48:32 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:44.833 23:48:32 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:14:44.833 23:48:32 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:44.833 23:48:32 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:14:44.833 23:48:32 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:44.833 23:48:32 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:14:44.833 23:48:32 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:14:44.833 23:48:32 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@284 -- # '[' 2 -gt 2 ']' 00:14:44.833 23:48:32 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@326 -- # killprocess 95968 00:14:44.833 23:48:32 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@954 -- # '[' -z 95968 ']' 00:14:44.833 23:48:32 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@958 -- # kill -0 95968 00:14:44.833 23:48:32 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@959 -- # uname 00:14:44.833 23:48:32 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:14:44.833 23:48:32 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 95968 00:14:44.833 killing process with pid 95968 00:14:44.833 23:48:32 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:14:44.833 23:48:32 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:14:44.833 23:48:32 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@972 -- # echo 'killing process with pid 95968' 00:14:44.833 23:48:32 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@973 -- # kill 95968 00:14:44.833 [2024-11-26 23:48:32.935672] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:14:44.833 23:48:32 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@978 -- # wait 95968 00:14:44.833 [2024-11-26 23:48:32.936655] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:14:45.094 23:48:33 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@328 -- # return 0 00:14:45.094 00:14:45.094 real 0m3.936s 00:14:45.094 user 0m6.275s 00:14:45.094 sys 0m0.772s 00:14:45.094 23:48:33 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@1130 -- # xtrace_disable 00:14:45.094 ************************************ 00:14:45.094 END TEST raid_state_function_test_sb_4k 00:14:45.094 ************************************ 00:14:45.094 23:48:33 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:14:45.094 23:48:33 bdev_raid -- bdev/bdev_raid.sh@998 -- # run_test raid_superblock_test_4k raid_superblock_test raid1 2 00:14:45.094 23:48:33 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:14:45.094 23:48:33 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:14:45.094 23:48:33 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:14:45.354 ************************************ 00:14:45.354 START TEST raid_superblock_test_4k 00:14:45.354 ************************************ 00:14:45.354 23:48:33 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@1129 -- # raid_superblock_test raid1 2 00:14:45.354 23:48:33 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@393 -- # local raid_level=raid1 00:14:45.354 23:48:33 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=2 00:14:45.354 23:48:33 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:14:45.354 23:48:33 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:14:45.354 23:48:33 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:14:45.354 23:48:33 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:14:45.354 23:48:33 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:14:45.354 23:48:33 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:14:45.354 23:48:33 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:14:45.354 23:48:33 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@399 -- # local strip_size 00:14:45.354 23:48:33 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:14:45.354 23:48:33 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:14:45.354 23:48:33 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:14:45.354 23:48:33 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@404 -- # '[' raid1 '!=' raid1 ']' 00:14:45.354 23:48:33 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@408 -- # strip_size=0 00:14:45.354 23:48:33 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@412 -- # raid_pid=96208 00:14:45.354 23:48:33 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:14:45.354 23:48:33 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@413 -- # waitforlisten 96208 00:14:45.354 23:48:33 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@835 -- # '[' -z 96208 ']' 00:14:45.354 23:48:33 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:14:45.354 23:48:33 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@840 -- # local max_retries=100 00:14:45.354 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:14:45.354 23:48:33 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:14:45.354 23:48:33 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@844 -- # xtrace_disable 00:14:45.354 23:48:33 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:14:45.354 [2024-11-26 23:48:33.318289] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:14:45.354 [2024-11-26 23:48:33.318436] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid96208 ] 00:14:45.354 [2024-11-26 23:48:33.472896] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:14:45.613 [2024-11-26 23:48:33.497179] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:14:45.613 [2024-11-26 23:48:33.538049] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:14:45.613 [2024-11-26 23:48:33.538090] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:14:46.209 23:48:34 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:14:46.209 23:48:34 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@868 -- # return 0 00:14:46.209 23:48:34 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:14:46.209 23:48:34 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:14:46.209 23:48:34 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:14:46.209 23:48:34 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:14:46.209 23:48:34 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:14:46.209 23:48:34 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:14:46.209 23:48:34 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:14:46.209 23:48:34 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:14:46.209 23:48:34 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 4096 -b malloc1 00:14:46.209 23:48:34 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:46.209 23:48:34 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:14:46.209 malloc1 00:14:46.209 23:48:34 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:46.209 23:48:34 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:14:46.209 23:48:34 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:46.209 23:48:34 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:14:46.209 [2024-11-26 23:48:34.148810] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:14:46.209 [2024-11-26 23:48:34.148928] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:46.209 [2024-11-26 23:48:34.148966] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:14:46.209 [2024-11-26 23:48:34.148999] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:46.209 [2024-11-26 23:48:34.151072] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:46.209 [2024-11-26 23:48:34.151149] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:14:46.209 pt1 00:14:46.209 23:48:34 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:46.209 23:48:34 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:14:46.209 23:48:34 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:14:46.209 23:48:34 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:14:46.209 23:48:34 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:14:46.209 23:48:34 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:14:46.209 23:48:34 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:14:46.209 23:48:34 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:14:46.209 23:48:34 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:14:46.210 23:48:34 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 4096 -b malloc2 00:14:46.210 23:48:34 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:46.210 23:48:34 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:14:46.210 malloc2 00:14:46.210 23:48:34 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:46.210 23:48:34 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:14:46.210 23:48:34 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:46.210 23:48:34 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:14:46.210 [2024-11-26 23:48:34.180985] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:14:46.210 [2024-11-26 23:48:34.181051] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:46.210 [2024-11-26 23:48:34.181068] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:14:46.210 [2024-11-26 23:48:34.181078] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:46.210 [2024-11-26 23:48:34.183107] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:46.210 [2024-11-26 23:48:34.183145] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:14:46.210 pt2 00:14:46.210 23:48:34 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:46.210 23:48:34 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:14:46.210 23:48:34 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:14:46.210 23:48:34 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''pt1 pt2'\''' -n raid_bdev1 -s 00:14:46.210 23:48:34 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:46.210 23:48:34 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:14:46.210 [2024-11-26 23:48:34.192999] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:14:46.210 [2024-11-26 23:48:34.194848] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:14:46.210 [2024-11-26 23:48:34.194985] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:14:46.210 [2024-11-26 23:48:34.194999] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:14:46.210 [2024-11-26 23:48:34.195237] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:14:46.210 [2024-11-26 23:48:34.195399] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:14:46.210 [2024-11-26 23:48:34.195410] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:14:46.210 [2024-11-26 23:48:34.195544] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:46.210 23:48:34 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:46.210 23:48:34 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:14:46.210 23:48:34 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:46.210 23:48:34 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:46.210 23:48:34 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:14:46.210 23:48:34 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:14:46.210 23:48:34 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:14:46.210 23:48:34 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:46.210 23:48:34 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:46.210 23:48:34 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:46.210 23:48:34 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:46.210 23:48:34 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:46.210 23:48:34 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:46.210 23:48:34 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:14:46.210 23:48:34 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:46.210 23:48:34 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:46.210 23:48:34 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:46.210 "name": "raid_bdev1", 00:14:46.210 "uuid": "d8455956-bc2b-4f8f-91e4-f39d8f6d3738", 00:14:46.210 "strip_size_kb": 0, 00:14:46.210 "state": "online", 00:14:46.210 "raid_level": "raid1", 00:14:46.210 "superblock": true, 00:14:46.210 "num_base_bdevs": 2, 00:14:46.210 "num_base_bdevs_discovered": 2, 00:14:46.210 "num_base_bdevs_operational": 2, 00:14:46.210 "base_bdevs_list": [ 00:14:46.210 { 00:14:46.210 "name": "pt1", 00:14:46.210 "uuid": "00000000-0000-0000-0000-000000000001", 00:14:46.210 "is_configured": true, 00:14:46.210 "data_offset": 256, 00:14:46.210 "data_size": 7936 00:14:46.210 }, 00:14:46.210 { 00:14:46.210 "name": "pt2", 00:14:46.210 "uuid": "00000000-0000-0000-0000-000000000002", 00:14:46.210 "is_configured": true, 00:14:46.210 "data_offset": 256, 00:14:46.210 "data_size": 7936 00:14:46.210 } 00:14:46.210 ] 00:14:46.210 }' 00:14:46.210 23:48:34 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:46.210 23:48:34 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:14:46.781 23:48:34 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:14:46.781 23:48:34 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:14:46.781 23:48:34 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:14:46.781 23:48:34 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:14:46.781 23:48:34 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@184 -- # local name 00:14:46.781 23:48:34 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:14:46.781 23:48:34 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:14:46.781 23:48:34 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:14:46.781 23:48:34 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:46.781 23:48:34 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:14:46.781 [2024-11-26 23:48:34.648494] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:14:46.781 23:48:34 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:46.781 23:48:34 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:14:46.781 "name": "raid_bdev1", 00:14:46.781 "aliases": [ 00:14:46.781 "d8455956-bc2b-4f8f-91e4-f39d8f6d3738" 00:14:46.781 ], 00:14:46.781 "product_name": "Raid Volume", 00:14:46.781 "block_size": 4096, 00:14:46.781 "num_blocks": 7936, 00:14:46.781 "uuid": "d8455956-bc2b-4f8f-91e4-f39d8f6d3738", 00:14:46.781 "assigned_rate_limits": { 00:14:46.781 "rw_ios_per_sec": 0, 00:14:46.781 "rw_mbytes_per_sec": 0, 00:14:46.781 "r_mbytes_per_sec": 0, 00:14:46.781 "w_mbytes_per_sec": 0 00:14:46.781 }, 00:14:46.781 "claimed": false, 00:14:46.781 "zoned": false, 00:14:46.781 "supported_io_types": { 00:14:46.781 "read": true, 00:14:46.781 "write": true, 00:14:46.781 "unmap": false, 00:14:46.781 "flush": false, 00:14:46.781 "reset": true, 00:14:46.781 "nvme_admin": false, 00:14:46.781 "nvme_io": false, 00:14:46.781 "nvme_io_md": false, 00:14:46.781 "write_zeroes": true, 00:14:46.781 "zcopy": false, 00:14:46.781 "get_zone_info": false, 00:14:46.781 "zone_management": false, 00:14:46.781 "zone_append": false, 00:14:46.781 "compare": false, 00:14:46.781 "compare_and_write": false, 00:14:46.781 "abort": false, 00:14:46.781 "seek_hole": false, 00:14:46.781 "seek_data": false, 00:14:46.781 "copy": false, 00:14:46.781 "nvme_iov_md": false 00:14:46.781 }, 00:14:46.781 "memory_domains": [ 00:14:46.781 { 00:14:46.781 "dma_device_id": "system", 00:14:46.781 "dma_device_type": 1 00:14:46.781 }, 00:14:46.781 { 00:14:46.781 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:46.781 "dma_device_type": 2 00:14:46.781 }, 00:14:46.781 { 00:14:46.781 "dma_device_id": "system", 00:14:46.781 "dma_device_type": 1 00:14:46.781 }, 00:14:46.781 { 00:14:46.781 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:46.781 "dma_device_type": 2 00:14:46.781 } 00:14:46.781 ], 00:14:46.781 "driver_specific": { 00:14:46.781 "raid": { 00:14:46.781 "uuid": "d8455956-bc2b-4f8f-91e4-f39d8f6d3738", 00:14:46.781 "strip_size_kb": 0, 00:14:46.781 "state": "online", 00:14:46.781 "raid_level": "raid1", 00:14:46.781 "superblock": true, 00:14:46.781 "num_base_bdevs": 2, 00:14:46.781 "num_base_bdevs_discovered": 2, 00:14:46.781 "num_base_bdevs_operational": 2, 00:14:46.781 "base_bdevs_list": [ 00:14:46.781 { 00:14:46.781 "name": "pt1", 00:14:46.781 "uuid": "00000000-0000-0000-0000-000000000001", 00:14:46.781 "is_configured": true, 00:14:46.781 "data_offset": 256, 00:14:46.781 "data_size": 7936 00:14:46.781 }, 00:14:46.781 { 00:14:46.781 "name": "pt2", 00:14:46.781 "uuid": "00000000-0000-0000-0000-000000000002", 00:14:46.781 "is_configured": true, 00:14:46.781 "data_offset": 256, 00:14:46.781 "data_size": 7936 00:14:46.781 } 00:14:46.781 ] 00:14:46.781 } 00:14:46.781 } 00:14:46.781 }' 00:14:46.781 23:48:34 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:14:46.781 23:48:34 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:14:46.781 pt2' 00:14:46.781 23:48:34 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:46.781 23:48:34 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='4096 ' 00:14:46.781 23:48:34 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:46.781 23:48:34 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:46.781 23:48:34 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:14:46.781 23:48:34 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:46.781 23:48:34 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:14:46.781 23:48:34 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:46.781 23:48:34 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 ' 00:14:46.781 23:48:34 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@193 -- # [[ 4096 == \4\0\9\6\ \ \ ]] 00:14:46.781 23:48:34 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:46.781 23:48:34 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:46.781 23:48:34 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:14:46.781 23:48:34 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:46.781 23:48:34 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:14:46.781 23:48:34 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:46.781 23:48:34 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 ' 00:14:46.781 23:48:34 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@193 -- # [[ 4096 == \4\0\9\6\ \ \ ]] 00:14:46.781 23:48:34 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:14:46.781 23:48:34 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:46.781 23:48:34 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:14:46.781 23:48:34 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:14:46.781 [2024-11-26 23:48:34.852020] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:14:46.781 23:48:34 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:46.781 23:48:34 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=d8455956-bc2b-4f8f-91e4-f39d8f6d3738 00:14:46.781 23:48:34 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@436 -- # '[' -z d8455956-bc2b-4f8f-91e4-f39d8f6d3738 ']' 00:14:46.781 23:48:34 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:14:46.781 23:48:34 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:46.781 23:48:34 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:14:46.781 [2024-11-26 23:48:34.891732] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:14:46.781 [2024-11-26 23:48:34.891756] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:14:46.781 [2024-11-26 23:48:34.891829] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:14:46.782 [2024-11-26 23:48:34.891892] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:14:46.782 [2024-11-26 23:48:34.891901] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:14:46.782 23:48:34 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:46.782 23:48:34 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:46.782 23:48:34 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:46.782 23:48:34 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:14:46.782 23:48:34 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:14:47.042 23:48:34 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:47.043 23:48:34 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:14:47.043 23:48:34 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:14:47.043 23:48:34 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:14:47.043 23:48:34 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:14:47.043 23:48:34 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:47.043 23:48:34 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:14:47.043 23:48:34 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:47.043 23:48:34 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:14:47.043 23:48:34 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:14:47.043 23:48:34 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:47.043 23:48:34 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:14:47.043 23:48:34 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:47.043 23:48:34 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:14:47.043 23:48:34 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:47.043 23:48:34 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:14:47.043 23:48:34 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:14:47.043 23:48:34 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:47.043 23:48:34 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:14:47.043 23:48:34 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:14:47.043 23:48:35 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@652 -- # local es=0 00:14:47.043 23:48:35 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:14:47.043 23:48:35 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:14:47.043 23:48:35 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:14:47.043 23:48:35 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:14:47.043 23:48:35 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:14:47.043 23:48:35 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:14:47.043 23:48:35 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:47.043 23:48:35 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:14:47.043 [2024-11-26 23:48:35.011542] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:14:47.043 [2024-11-26 23:48:35.013296] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:14:47.043 [2024-11-26 23:48:35.013398] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:14:47.043 [2024-11-26 23:48:35.013461] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:14:47.043 [2024-11-26 23:48:35.013477] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:14:47.043 [2024-11-26 23:48:35.013486] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state configuring 00:14:47.043 request: 00:14:47.043 { 00:14:47.043 "name": "raid_bdev1", 00:14:47.043 "raid_level": "raid1", 00:14:47.043 "base_bdevs": [ 00:14:47.043 "malloc1", 00:14:47.043 "malloc2" 00:14:47.043 ], 00:14:47.043 "superblock": false, 00:14:47.043 "method": "bdev_raid_create", 00:14:47.043 "req_id": 1 00:14:47.043 } 00:14:47.043 Got JSON-RPC error response 00:14:47.043 response: 00:14:47.043 { 00:14:47.043 "code": -17, 00:14:47.043 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:14:47.043 } 00:14:47.043 23:48:35 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:14:47.043 23:48:35 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@655 -- # es=1 00:14:47.043 23:48:35 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:14:47.043 23:48:35 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:14:47.043 23:48:35 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:14:47.043 23:48:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:14:47.043 23:48:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:47.043 23:48:35 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:47.043 23:48:35 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:14:47.043 23:48:35 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:47.043 23:48:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:14:47.043 23:48:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:14:47.043 23:48:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:14:47.043 23:48:35 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:47.043 23:48:35 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:14:47.043 [2024-11-26 23:48:35.055465] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:14:47.043 [2024-11-26 23:48:35.055555] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:47.043 [2024-11-26 23:48:35.055608] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:14:47.043 [2024-11-26 23:48:35.055640] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:47.043 [2024-11-26 23:48:35.057726] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:47.043 [2024-11-26 23:48:35.057791] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:14:47.043 [2024-11-26 23:48:35.057889] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:14:47.043 [2024-11-26 23:48:35.057942] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:14:47.043 pt1 00:14:47.043 23:48:35 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:47.043 23:48:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 2 00:14:47.043 23:48:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:47.043 23:48:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:47.043 23:48:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:14:47.043 23:48:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:14:47.043 23:48:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:14:47.043 23:48:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:47.043 23:48:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:47.043 23:48:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:47.043 23:48:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:47.043 23:48:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:47.043 23:48:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:47.043 23:48:35 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:47.043 23:48:35 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:14:47.043 23:48:35 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:47.043 23:48:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:47.043 "name": "raid_bdev1", 00:14:47.043 "uuid": "d8455956-bc2b-4f8f-91e4-f39d8f6d3738", 00:14:47.043 "strip_size_kb": 0, 00:14:47.043 "state": "configuring", 00:14:47.043 "raid_level": "raid1", 00:14:47.043 "superblock": true, 00:14:47.043 "num_base_bdevs": 2, 00:14:47.043 "num_base_bdevs_discovered": 1, 00:14:47.043 "num_base_bdevs_operational": 2, 00:14:47.043 "base_bdevs_list": [ 00:14:47.043 { 00:14:47.043 "name": "pt1", 00:14:47.043 "uuid": "00000000-0000-0000-0000-000000000001", 00:14:47.043 "is_configured": true, 00:14:47.043 "data_offset": 256, 00:14:47.043 "data_size": 7936 00:14:47.043 }, 00:14:47.043 { 00:14:47.043 "name": null, 00:14:47.043 "uuid": "00000000-0000-0000-0000-000000000002", 00:14:47.043 "is_configured": false, 00:14:47.043 "data_offset": 256, 00:14:47.043 "data_size": 7936 00:14:47.043 } 00:14:47.043 ] 00:14:47.043 }' 00:14:47.043 23:48:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:47.043 23:48:35 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:14:47.612 23:48:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@470 -- # '[' 2 -gt 2 ']' 00:14:47.612 23:48:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:14:47.612 23:48:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:14:47.612 23:48:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:14:47.612 23:48:35 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:47.612 23:48:35 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:14:47.612 [2024-11-26 23:48:35.482740] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:14:47.612 [2024-11-26 23:48:35.482793] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:47.612 [2024-11-26 23:48:35.482830] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:14:47.612 [2024-11-26 23:48:35.482839] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:47.612 [2024-11-26 23:48:35.483191] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:47.612 [2024-11-26 23:48:35.483207] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:14:47.612 [2024-11-26 23:48:35.483269] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:14:47.612 [2024-11-26 23:48:35.483288] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:14:47.612 [2024-11-26 23:48:35.483437] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:14:47.612 [2024-11-26 23:48:35.483464] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:14:47.612 [2024-11-26 23:48:35.483717] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:14:47.612 [2024-11-26 23:48:35.483861] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:14:47.612 [2024-11-26 23:48:35.483902] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:14:47.612 [2024-11-26 23:48:35.484031] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:47.612 pt2 00:14:47.612 23:48:35 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:47.612 23:48:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:14:47.612 23:48:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:14:47.612 23:48:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:14:47.612 23:48:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:47.612 23:48:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:47.612 23:48:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:14:47.612 23:48:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:14:47.612 23:48:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:14:47.612 23:48:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:47.612 23:48:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:47.612 23:48:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:47.612 23:48:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:47.612 23:48:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:47.612 23:48:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:47.612 23:48:35 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:47.612 23:48:35 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:14:47.612 23:48:35 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:47.612 23:48:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:47.612 "name": "raid_bdev1", 00:14:47.612 "uuid": "d8455956-bc2b-4f8f-91e4-f39d8f6d3738", 00:14:47.612 "strip_size_kb": 0, 00:14:47.612 "state": "online", 00:14:47.612 "raid_level": "raid1", 00:14:47.612 "superblock": true, 00:14:47.612 "num_base_bdevs": 2, 00:14:47.612 "num_base_bdevs_discovered": 2, 00:14:47.612 "num_base_bdevs_operational": 2, 00:14:47.612 "base_bdevs_list": [ 00:14:47.612 { 00:14:47.612 "name": "pt1", 00:14:47.612 "uuid": "00000000-0000-0000-0000-000000000001", 00:14:47.612 "is_configured": true, 00:14:47.612 "data_offset": 256, 00:14:47.612 "data_size": 7936 00:14:47.612 }, 00:14:47.612 { 00:14:47.612 "name": "pt2", 00:14:47.612 "uuid": "00000000-0000-0000-0000-000000000002", 00:14:47.612 "is_configured": true, 00:14:47.612 "data_offset": 256, 00:14:47.612 "data_size": 7936 00:14:47.612 } 00:14:47.612 ] 00:14:47.612 }' 00:14:47.612 23:48:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:47.612 23:48:35 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:14:47.872 23:48:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:14:47.872 23:48:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:14:47.872 23:48:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:14:47.872 23:48:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:14:47.872 23:48:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@184 -- # local name 00:14:47.872 23:48:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:14:47.872 23:48:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:14:47.872 23:48:35 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:47.872 23:48:35 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:14:47.872 23:48:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:14:47.872 [2024-11-26 23:48:35.894288] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:14:47.872 23:48:35 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:47.872 23:48:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:14:47.872 "name": "raid_bdev1", 00:14:47.872 "aliases": [ 00:14:47.872 "d8455956-bc2b-4f8f-91e4-f39d8f6d3738" 00:14:47.872 ], 00:14:47.872 "product_name": "Raid Volume", 00:14:47.872 "block_size": 4096, 00:14:47.872 "num_blocks": 7936, 00:14:47.872 "uuid": "d8455956-bc2b-4f8f-91e4-f39d8f6d3738", 00:14:47.872 "assigned_rate_limits": { 00:14:47.872 "rw_ios_per_sec": 0, 00:14:47.872 "rw_mbytes_per_sec": 0, 00:14:47.872 "r_mbytes_per_sec": 0, 00:14:47.872 "w_mbytes_per_sec": 0 00:14:47.872 }, 00:14:47.872 "claimed": false, 00:14:47.872 "zoned": false, 00:14:47.872 "supported_io_types": { 00:14:47.872 "read": true, 00:14:47.872 "write": true, 00:14:47.872 "unmap": false, 00:14:47.872 "flush": false, 00:14:47.872 "reset": true, 00:14:47.872 "nvme_admin": false, 00:14:47.872 "nvme_io": false, 00:14:47.872 "nvme_io_md": false, 00:14:47.872 "write_zeroes": true, 00:14:47.872 "zcopy": false, 00:14:47.872 "get_zone_info": false, 00:14:47.872 "zone_management": false, 00:14:47.872 "zone_append": false, 00:14:47.872 "compare": false, 00:14:47.872 "compare_and_write": false, 00:14:47.872 "abort": false, 00:14:47.872 "seek_hole": false, 00:14:47.872 "seek_data": false, 00:14:47.872 "copy": false, 00:14:47.872 "nvme_iov_md": false 00:14:47.872 }, 00:14:47.872 "memory_domains": [ 00:14:47.872 { 00:14:47.872 "dma_device_id": "system", 00:14:47.872 "dma_device_type": 1 00:14:47.872 }, 00:14:47.872 { 00:14:47.872 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:47.872 "dma_device_type": 2 00:14:47.872 }, 00:14:47.872 { 00:14:47.872 "dma_device_id": "system", 00:14:47.872 "dma_device_type": 1 00:14:47.872 }, 00:14:47.872 { 00:14:47.872 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:47.872 "dma_device_type": 2 00:14:47.872 } 00:14:47.872 ], 00:14:47.872 "driver_specific": { 00:14:47.872 "raid": { 00:14:47.872 "uuid": "d8455956-bc2b-4f8f-91e4-f39d8f6d3738", 00:14:47.872 "strip_size_kb": 0, 00:14:47.872 "state": "online", 00:14:47.872 "raid_level": "raid1", 00:14:47.872 "superblock": true, 00:14:47.872 "num_base_bdevs": 2, 00:14:47.872 "num_base_bdevs_discovered": 2, 00:14:47.872 "num_base_bdevs_operational": 2, 00:14:47.872 "base_bdevs_list": [ 00:14:47.872 { 00:14:47.872 "name": "pt1", 00:14:47.872 "uuid": "00000000-0000-0000-0000-000000000001", 00:14:47.872 "is_configured": true, 00:14:47.872 "data_offset": 256, 00:14:47.872 "data_size": 7936 00:14:47.872 }, 00:14:47.872 { 00:14:47.872 "name": "pt2", 00:14:47.872 "uuid": "00000000-0000-0000-0000-000000000002", 00:14:47.872 "is_configured": true, 00:14:47.872 "data_offset": 256, 00:14:47.872 "data_size": 7936 00:14:47.872 } 00:14:47.872 ] 00:14:47.872 } 00:14:47.872 } 00:14:47.872 }' 00:14:47.872 23:48:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:14:47.872 23:48:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:14:47.872 pt2' 00:14:47.872 23:48:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:47.872 23:48:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='4096 ' 00:14:47.872 23:48:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:48.131 23:48:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:48.131 23:48:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:14:48.131 23:48:36 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:48.131 23:48:36 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:14:48.131 23:48:36 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:48.131 23:48:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 ' 00:14:48.131 23:48:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@193 -- # [[ 4096 == \4\0\9\6\ \ \ ]] 00:14:48.131 23:48:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:48.131 23:48:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:14:48.131 23:48:36 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:48.131 23:48:36 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:14:48.131 23:48:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:48.131 23:48:36 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:48.131 23:48:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 ' 00:14:48.131 23:48:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@193 -- # [[ 4096 == \4\0\9\6\ \ \ ]] 00:14:48.131 23:48:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:14:48.131 23:48:36 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:48.131 23:48:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:14:48.131 23:48:36 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:14:48.131 [2024-11-26 23:48:36.117874] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:14:48.131 23:48:36 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:48.131 23:48:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@487 -- # '[' d8455956-bc2b-4f8f-91e4-f39d8f6d3738 '!=' d8455956-bc2b-4f8f-91e4-f39d8f6d3738 ']' 00:14:48.131 23:48:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@491 -- # has_redundancy raid1 00:14:48.131 23:48:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@198 -- # case $1 in 00:14:48.131 23:48:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@199 -- # return 0 00:14:48.131 23:48:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@493 -- # rpc_cmd bdev_passthru_delete pt1 00:14:48.131 23:48:36 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:48.131 23:48:36 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:14:48.131 [2024-11-26 23:48:36.161593] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt1 00:14:48.131 23:48:36 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:48.131 23:48:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@496 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:14:48.131 23:48:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:48.131 23:48:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:48.131 23:48:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:14:48.131 23:48:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:14:48.131 23:48:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:14:48.131 23:48:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:48.131 23:48:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:48.131 23:48:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:48.131 23:48:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:48.131 23:48:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:48.131 23:48:36 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:48.131 23:48:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:48.131 23:48:36 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:14:48.131 23:48:36 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:48.131 23:48:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:48.131 "name": "raid_bdev1", 00:14:48.131 "uuid": "d8455956-bc2b-4f8f-91e4-f39d8f6d3738", 00:14:48.131 "strip_size_kb": 0, 00:14:48.131 "state": "online", 00:14:48.131 "raid_level": "raid1", 00:14:48.131 "superblock": true, 00:14:48.131 "num_base_bdevs": 2, 00:14:48.131 "num_base_bdevs_discovered": 1, 00:14:48.131 "num_base_bdevs_operational": 1, 00:14:48.131 "base_bdevs_list": [ 00:14:48.131 { 00:14:48.131 "name": null, 00:14:48.131 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:48.131 "is_configured": false, 00:14:48.131 "data_offset": 0, 00:14:48.131 "data_size": 7936 00:14:48.131 }, 00:14:48.131 { 00:14:48.131 "name": "pt2", 00:14:48.131 "uuid": "00000000-0000-0000-0000-000000000002", 00:14:48.131 "is_configured": true, 00:14:48.131 "data_offset": 256, 00:14:48.131 "data_size": 7936 00:14:48.131 } 00:14:48.131 ] 00:14:48.131 }' 00:14:48.131 23:48:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:48.131 23:48:36 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:14:48.700 23:48:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@499 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:14:48.700 23:48:36 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:48.700 23:48:36 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:14:48.700 [2024-11-26 23:48:36.648842] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:14:48.700 [2024-11-26 23:48:36.648919] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:14:48.700 [2024-11-26 23:48:36.649032] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:14:48.700 [2024-11-26 23:48:36.649109] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:14:48.700 [2024-11-26 23:48:36.649140] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:14:48.700 23:48:36 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:48.700 23:48:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@500 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:48.700 23:48:36 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:48.700 23:48:36 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:14:48.700 23:48:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@500 -- # jq -r '.[]' 00:14:48.700 23:48:36 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:48.700 23:48:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@500 -- # raid_bdev= 00:14:48.700 23:48:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@501 -- # '[' -n '' ']' 00:14:48.700 23:48:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@506 -- # (( i = 1 )) 00:14:48.700 23:48:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:14:48.700 23:48:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt2 00:14:48.700 23:48:36 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:48.700 23:48:36 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:14:48.700 23:48:36 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:48.700 23:48:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:14:48.700 23:48:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:14:48.700 23:48:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@511 -- # (( i = 1 )) 00:14:48.700 23:48:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:14:48.700 23:48:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@519 -- # i=1 00:14:48.700 23:48:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@520 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:14:48.700 23:48:36 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:48.700 23:48:36 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:14:48.700 [2024-11-26 23:48:36.720720] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:14:48.700 [2024-11-26 23:48:36.720806] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:48.700 [2024-11-26 23:48:36.720853] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008780 00:14:48.700 [2024-11-26 23:48:36.720879] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:48.700 [2024-11-26 23:48:36.722947] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:48.700 [2024-11-26 23:48:36.723027] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:14:48.700 [2024-11-26 23:48:36.723112] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:14:48.700 [2024-11-26 23:48:36.723158] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:14:48.700 [2024-11-26 23:48:36.723244] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:14:48.700 [2024-11-26 23:48:36.723285] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:14:48.700 [2024-11-26 23:48:36.723532] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:14:48.700 [2024-11-26 23:48:36.723678] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:14:48.700 [2024-11-26 23:48:36.723718] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001c80 00:14:48.700 [2024-11-26 23:48:36.723844] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:48.700 pt2 00:14:48.700 23:48:36 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:48.700 23:48:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@523 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:14:48.700 23:48:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:48.700 23:48:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:48.700 23:48:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:14:48.700 23:48:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:14:48.700 23:48:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:14:48.700 23:48:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:48.700 23:48:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:48.700 23:48:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:48.700 23:48:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:48.700 23:48:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:48.700 23:48:36 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:48.700 23:48:36 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:14:48.700 23:48:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:48.700 23:48:36 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:48.700 23:48:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:48.700 "name": "raid_bdev1", 00:14:48.700 "uuid": "d8455956-bc2b-4f8f-91e4-f39d8f6d3738", 00:14:48.700 "strip_size_kb": 0, 00:14:48.700 "state": "online", 00:14:48.700 "raid_level": "raid1", 00:14:48.700 "superblock": true, 00:14:48.700 "num_base_bdevs": 2, 00:14:48.700 "num_base_bdevs_discovered": 1, 00:14:48.700 "num_base_bdevs_operational": 1, 00:14:48.700 "base_bdevs_list": [ 00:14:48.700 { 00:14:48.700 "name": null, 00:14:48.700 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:48.700 "is_configured": false, 00:14:48.700 "data_offset": 256, 00:14:48.700 "data_size": 7936 00:14:48.700 }, 00:14:48.700 { 00:14:48.700 "name": "pt2", 00:14:48.700 "uuid": "00000000-0000-0000-0000-000000000002", 00:14:48.700 "is_configured": true, 00:14:48.700 "data_offset": 256, 00:14:48.700 "data_size": 7936 00:14:48.700 } 00:14:48.700 ] 00:14:48.700 }' 00:14:48.700 23:48:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:48.700 23:48:36 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:14:49.270 23:48:37 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@526 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:14:49.270 23:48:37 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:49.270 23:48:37 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:14:49.270 [2024-11-26 23:48:37.171978] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:14:49.270 [2024-11-26 23:48:37.172049] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:14:49.270 [2024-11-26 23:48:37.172115] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:14:49.270 [2024-11-26 23:48:37.172151] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:14:49.270 [2024-11-26 23:48:37.172161] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name raid_bdev1, state offline 00:14:49.270 23:48:37 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:49.270 23:48:37 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@527 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:49.270 23:48:37 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:49.270 23:48:37 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@527 -- # jq -r '.[]' 00:14:49.270 23:48:37 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:14:49.270 23:48:37 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:49.270 23:48:37 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@527 -- # raid_bdev= 00:14:49.270 23:48:37 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@528 -- # '[' -n '' ']' 00:14:49.270 23:48:37 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@532 -- # '[' 2 -gt 2 ']' 00:14:49.270 23:48:37 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@540 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:14:49.270 23:48:37 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:49.270 23:48:37 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:14:49.270 [2024-11-26 23:48:37.231871] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:14:49.270 [2024-11-26 23:48:37.231984] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:49.270 [2024-11-26 23:48:37.232016] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008d80 00:14:49.270 [2024-11-26 23:48:37.232045] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:49.270 [2024-11-26 23:48:37.234073] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:49.270 [2024-11-26 23:48:37.234140] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:14:49.270 [2024-11-26 23:48:37.234233] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:14:49.270 [2024-11-26 23:48:37.234281] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:14:49.270 [2024-11-26 23:48:37.234392] bdev_raid.c:3685:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev pt2 (4) greater than existing raid bdev raid_bdev1 (2) 00:14:49.270 [2024-11-26 23:48:37.234443] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:14:49.270 [2024-11-26 23:48:37.234490] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002000 name raid_bdev1, state configuring 00:14:49.270 [2024-11-26 23:48:37.234574] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:14:49.270 [2024-11-26 23:48:37.234670] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000002380 00:14:49.270 [2024-11-26 23:48:37.234713] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:14:49.270 [2024-11-26 23:48:37.234946] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:14:49.270 [2024-11-26 23:48:37.235085] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000002380 00:14:49.270 [2024-11-26 23:48:37.235121] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000002380 00:14:49.270 [2024-11-26 23:48:37.235255] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:49.270 pt1 00:14:49.270 23:48:37 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:49.270 23:48:37 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@542 -- # '[' 2 -gt 2 ']' 00:14:49.270 23:48:37 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@554 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:14:49.270 23:48:37 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:49.270 23:48:37 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:49.270 23:48:37 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:14:49.270 23:48:37 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:14:49.270 23:48:37 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:14:49.270 23:48:37 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:49.270 23:48:37 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:49.270 23:48:37 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:49.270 23:48:37 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:49.270 23:48:37 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:49.270 23:48:37 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:49.270 23:48:37 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:49.270 23:48:37 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:14:49.270 23:48:37 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:49.270 23:48:37 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:49.270 "name": "raid_bdev1", 00:14:49.270 "uuid": "d8455956-bc2b-4f8f-91e4-f39d8f6d3738", 00:14:49.270 "strip_size_kb": 0, 00:14:49.270 "state": "online", 00:14:49.270 "raid_level": "raid1", 00:14:49.270 "superblock": true, 00:14:49.270 "num_base_bdevs": 2, 00:14:49.270 "num_base_bdevs_discovered": 1, 00:14:49.270 "num_base_bdevs_operational": 1, 00:14:49.270 "base_bdevs_list": [ 00:14:49.270 { 00:14:49.270 "name": null, 00:14:49.270 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:49.270 "is_configured": false, 00:14:49.270 "data_offset": 256, 00:14:49.270 "data_size": 7936 00:14:49.270 }, 00:14:49.270 { 00:14:49.270 "name": "pt2", 00:14:49.270 "uuid": "00000000-0000-0000-0000-000000000002", 00:14:49.270 "is_configured": true, 00:14:49.270 "data_offset": 256, 00:14:49.270 "data_size": 7936 00:14:49.270 } 00:14:49.270 ] 00:14:49.270 }' 00:14:49.270 23:48:37 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:49.270 23:48:37 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:14:49.838 23:48:37 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@555 -- # rpc_cmd bdev_raid_get_bdevs online 00:14:49.838 23:48:37 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@555 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:14:49.838 23:48:37 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:49.838 23:48:37 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:14:49.838 23:48:37 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:49.838 23:48:37 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@555 -- # [[ false == \f\a\l\s\e ]] 00:14:49.838 23:48:37 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@558 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:14:49.838 23:48:37 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@558 -- # jq -r '.[] | .uuid' 00:14:49.838 23:48:37 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:49.838 23:48:37 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:14:49.838 [2024-11-26 23:48:37.719336] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:14:49.838 23:48:37 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:49.838 23:48:37 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@558 -- # '[' d8455956-bc2b-4f8f-91e4-f39d8f6d3738 '!=' d8455956-bc2b-4f8f-91e4-f39d8f6d3738 ']' 00:14:49.838 23:48:37 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@563 -- # killprocess 96208 00:14:49.838 23:48:37 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@954 -- # '[' -z 96208 ']' 00:14:49.838 23:48:37 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@958 -- # kill -0 96208 00:14:49.838 23:48:37 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@959 -- # uname 00:14:49.838 23:48:37 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:14:49.838 23:48:37 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 96208 00:14:49.838 23:48:37 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:14:49.838 23:48:37 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:14:49.838 killing process with pid 96208 00:14:49.839 23:48:37 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@972 -- # echo 'killing process with pid 96208' 00:14:49.839 23:48:37 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@973 -- # kill 96208 00:14:49.839 [2024-11-26 23:48:37.801764] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:14:49.839 [2024-11-26 23:48:37.801832] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:14:49.839 [2024-11-26 23:48:37.801872] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:14:49.839 [2024-11-26 23:48:37.801880] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002380 name raid_bdev1, state offline 00:14:49.839 23:48:37 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@978 -- # wait 96208 00:14:49.839 [2024-11-26 23:48:37.824626] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:14:50.097 ************************************ 00:14:50.097 END TEST raid_superblock_test_4k 00:14:50.097 ************************************ 00:14:50.097 23:48:38 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@565 -- # return 0 00:14:50.097 00:14:50.097 real 0m4.805s 00:14:50.097 user 0m7.826s 00:14:50.097 sys 0m1.050s 00:14:50.097 23:48:38 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@1130 -- # xtrace_disable 00:14:50.097 23:48:38 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:14:50.097 23:48:38 bdev_raid -- bdev/bdev_raid.sh@999 -- # '[' true = true ']' 00:14:50.097 23:48:38 bdev_raid -- bdev/bdev_raid.sh@1000 -- # run_test raid_rebuild_test_sb_4k raid_rebuild_test raid1 2 true false true 00:14:50.097 23:48:38 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:14:50.097 23:48:38 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:14:50.097 23:48:38 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:14:50.097 ************************************ 00:14:50.097 START TEST raid_rebuild_test_sb_4k 00:14:50.098 ************************************ 00:14:50.098 23:48:38 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@1129 -- # raid_rebuild_test raid1 2 true false true 00:14:50.098 23:48:38 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:14:50.098 23:48:38 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=2 00:14:50.098 23:48:38 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@571 -- # local superblock=true 00:14:50.098 23:48:38 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:14:50.098 23:48:38 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@573 -- # local verify=true 00:14:50.098 23:48:38 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:14:50.098 23:48:38 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:14:50.098 23:48:38 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:14:50.098 23:48:38 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:14:50.098 23:48:38 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:14:50.098 23:48:38 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:14:50.098 23:48:38 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:14:50.098 23:48:38 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:14:50.098 23:48:38 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:14:50.098 23:48:38 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:14:50.098 23:48:38 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:14:50.098 23:48:38 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@576 -- # local strip_size 00:14:50.098 23:48:38 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@577 -- # local create_arg 00:14:50.098 23:48:38 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:14:50.098 23:48:38 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@579 -- # local data_offset 00:14:50.098 23:48:38 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:14:50.098 23:48:38 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:14:50.098 23:48:38 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@592 -- # '[' true = true ']' 00:14:50.098 23:48:38 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@593 -- # create_arg+=' -s' 00:14:50.098 23:48:38 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@597 -- # raid_pid=96521 00:14:50.098 23:48:38 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:14:50.098 23:48:38 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@598 -- # waitforlisten 96521 00:14:50.098 23:48:38 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@835 -- # '[' -z 96521 ']' 00:14:50.098 23:48:38 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:14:50.098 23:48:38 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@840 -- # local max_retries=100 00:14:50.098 23:48:38 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:14:50.098 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:14:50.098 23:48:38 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@844 -- # xtrace_disable 00:14:50.098 23:48:38 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:14:50.098 [2024-11-26 23:48:38.221623] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:14:50.098 [2024-11-26 23:48:38.221814] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.ealI/O size of 3145728 is greater than zero copy threshold (65536). 00:14:50.098 Zero copy mechanism will not be used. 00:14:50.098 :6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid96521 ] 00:14:50.357 [2024-11-26 23:48:38.379681] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:14:50.357 [2024-11-26 23:48:38.405113] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:14:50.357 [2024-11-26 23:48:38.447493] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:14:50.357 [2024-11-26 23:48:38.447612] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:14:50.928 23:48:39 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:14:50.928 23:48:39 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@868 -- # return 0 00:14:50.928 23:48:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:14:50.928 23:48:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 4096 -b BaseBdev1_malloc 00:14:50.928 23:48:39 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:50.928 23:48:39 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:14:50.928 BaseBdev1_malloc 00:14:50.928 23:48:39 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:50.928 23:48:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:14:50.928 23:48:39 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:50.928 23:48:39 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:14:50.928 [2024-11-26 23:48:39.046771] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:14:50.928 [2024-11-26 23:48:39.046849] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:50.928 [2024-11-26 23:48:39.046876] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:14:50.928 [2024-11-26 23:48:39.046887] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:50.928 [2024-11-26 23:48:39.048932] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:50.928 [2024-11-26 23:48:39.048979] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:14:50.928 BaseBdev1 00:14:50.928 23:48:39 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:50.928 23:48:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:14:50.928 23:48:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 4096 -b BaseBdev2_malloc 00:14:50.928 23:48:39 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:50.928 23:48:39 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:14:51.189 BaseBdev2_malloc 00:14:51.189 23:48:39 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:51.189 23:48:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:14:51.189 23:48:39 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:51.189 23:48:39 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:14:51.189 [2024-11-26 23:48:39.075173] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:14:51.189 [2024-11-26 23:48:39.075320] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:51.189 [2024-11-26 23:48:39.075360] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:14:51.189 [2024-11-26 23:48:39.075370] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:51.189 [2024-11-26 23:48:39.077362] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:51.189 [2024-11-26 23:48:39.077413] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:14:51.189 BaseBdev2 00:14:51.189 23:48:39 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:51.189 23:48:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 4096 -b spare_malloc 00:14:51.189 23:48:39 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:51.189 23:48:39 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:14:51.189 spare_malloc 00:14:51.189 23:48:39 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:51.189 23:48:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:14:51.189 23:48:39 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:51.189 23:48:39 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:14:51.189 spare_delay 00:14:51.189 23:48:39 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:51.189 23:48:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:14:51.189 23:48:39 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:51.189 23:48:39 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:14:51.189 [2024-11-26 23:48:39.115437] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:14:51.189 [2024-11-26 23:48:39.115485] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:51.189 [2024-11-26 23:48:39.115517] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:14:51.189 [2024-11-26 23:48:39.115525] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:51.189 [2024-11-26 23:48:39.117503] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:51.189 [2024-11-26 23:48:39.117536] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:14:51.189 spare 00:14:51.189 23:48:39 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:51.189 23:48:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 00:14:51.189 23:48:39 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:51.189 23:48:39 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:14:51.189 [2024-11-26 23:48:39.127475] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:14:51.189 [2024-11-26 23:48:39.129232] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:14:51.189 [2024-11-26 23:48:39.129387] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:14:51.189 [2024-11-26 23:48:39.129400] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:14:51.189 [2024-11-26 23:48:39.129654] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:14:51.189 [2024-11-26 23:48:39.129796] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:14:51.189 [2024-11-26 23:48:39.129816] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:14:51.189 [2024-11-26 23:48:39.129915] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:51.189 23:48:39 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:51.189 23:48:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:14:51.189 23:48:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:51.189 23:48:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:51.189 23:48:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:14:51.189 23:48:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:14:51.189 23:48:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:14:51.189 23:48:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:51.189 23:48:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:51.189 23:48:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:51.189 23:48:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:51.189 23:48:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:51.189 23:48:39 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:51.189 23:48:39 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:14:51.189 23:48:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:51.189 23:48:39 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:51.189 23:48:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:51.189 "name": "raid_bdev1", 00:14:51.189 "uuid": "ff5af955-a414-4ba0-8235-7545475fb4f4", 00:14:51.189 "strip_size_kb": 0, 00:14:51.189 "state": "online", 00:14:51.189 "raid_level": "raid1", 00:14:51.189 "superblock": true, 00:14:51.189 "num_base_bdevs": 2, 00:14:51.190 "num_base_bdevs_discovered": 2, 00:14:51.190 "num_base_bdevs_operational": 2, 00:14:51.190 "base_bdevs_list": [ 00:14:51.190 { 00:14:51.190 "name": "BaseBdev1", 00:14:51.190 "uuid": "7d05ac63-88b0-560b-aae6-2210a5544c4c", 00:14:51.190 "is_configured": true, 00:14:51.190 "data_offset": 256, 00:14:51.190 "data_size": 7936 00:14:51.190 }, 00:14:51.190 { 00:14:51.190 "name": "BaseBdev2", 00:14:51.190 "uuid": "6f077572-90b7-568a-919c-93be71941c99", 00:14:51.190 "is_configured": true, 00:14:51.190 "data_offset": 256, 00:14:51.190 "data_size": 7936 00:14:51.190 } 00:14:51.190 ] 00:14:51.190 }' 00:14:51.190 23:48:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:51.190 23:48:39 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:14:51.448 23:48:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:14:51.448 23:48:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:14:51.448 23:48:39 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:51.448 23:48:39 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:14:51.448 [2024-11-26 23:48:39.558936] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:14:51.707 23:48:39 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:51.708 23:48:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=7936 00:14:51.708 23:48:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:51.708 23:48:39 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:51.708 23:48:39 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:14:51.708 23:48:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:14:51.708 23:48:39 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:51.708 23:48:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@619 -- # data_offset=256 00:14:51.708 23:48:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:14:51.708 23:48:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:14:51.708 23:48:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:14:51.708 23:48:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:14:51.708 23:48:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:14:51.708 23:48:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:14:51.708 23:48:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@10 -- # local bdev_list 00:14:51.708 23:48:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:14:51.708 23:48:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@11 -- # local nbd_list 00:14:51.708 23:48:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@12 -- # local i 00:14:51.708 23:48:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:14:51.708 23:48:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:14:51.708 23:48:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:14:51.708 [2024-11-26 23:48:39.826308] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:14:51.967 /dev/nbd0 00:14:51.968 23:48:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:14:51.968 23:48:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:14:51.968 23:48:39 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:14:51.968 23:48:39 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@873 -- # local i 00:14:51.968 23:48:39 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:14:51.968 23:48:39 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:14:51.968 23:48:39 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:14:51.968 23:48:39 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@877 -- # break 00:14:51.968 23:48:39 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:14:51.968 23:48:39 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:14:51.968 23:48:39 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:14:51.968 1+0 records in 00:14:51.968 1+0 records out 00:14:51.968 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000575123 s, 7.1 MB/s 00:14:51.968 23:48:39 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:51.968 23:48:39 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@890 -- # size=4096 00:14:51.968 23:48:39 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:51.968 23:48:39 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:14:51.968 23:48:39 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@893 -- # return 0 00:14:51.968 23:48:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:14:51.968 23:48:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:14:51.968 23:48:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@629 -- # '[' raid1 = raid5f ']' 00:14:51.968 23:48:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@633 -- # write_unit_size=1 00:14:51.968 23:48:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=4096 count=7936 oflag=direct 00:14:52.536 7936+0 records in 00:14:52.536 7936+0 records out 00:14:52.536 32505856 bytes (33 MB, 31 MiB) copied, 0.608853 s, 53.4 MB/s 00:14:52.536 23:48:40 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:14:52.536 23:48:40 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:14:52.536 23:48:40 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:14:52.536 23:48:40 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@50 -- # local nbd_list 00:14:52.536 23:48:40 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@51 -- # local i 00:14:52.536 23:48:40 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:14:52.536 23:48:40 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:14:52.797 [2024-11-26 23:48:40.727356] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:52.797 23:48:40 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:14:52.797 23:48:40 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:14:52.797 23:48:40 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:14:52.797 23:48:40 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:14:52.797 23:48:40 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:14:52.797 23:48:40 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:14:52.797 23:48:40 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@41 -- # break 00:14:52.797 23:48:40 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@45 -- # return 0 00:14:52.797 23:48:40 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:14:52.797 23:48:40 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:52.797 23:48:40 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:14:52.797 [2024-11-26 23:48:40.755400] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:14:52.797 23:48:40 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:52.797 23:48:40 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:14:52.797 23:48:40 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:52.797 23:48:40 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:52.797 23:48:40 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:14:52.797 23:48:40 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:14:52.797 23:48:40 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:14:52.797 23:48:40 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:52.797 23:48:40 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:52.797 23:48:40 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:52.797 23:48:40 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:52.797 23:48:40 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:52.797 23:48:40 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:52.797 23:48:40 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:52.797 23:48:40 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:14:52.797 23:48:40 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:52.797 23:48:40 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:52.797 "name": "raid_bdev1", 00:14:52.797 "uuid": "ff5af955-a414-4ba0-8235-7545475fb4f4", 00:14:52.797 "strip_size_kb": 0, 00:14:52.797 "state": "online", 00:14:52.797 "raid_level": "raid1", 00:14:52.797 "superblock": true, 00:14:52.797 "num_base_bdevs": 2, 00:14:52.797 "num_base_bdevs_discovered": 1, 00:14:52.797 "num_base_bdevs_operational": 1, 00:14:52.797 "base_bdevs_list": [ 00:14:52.797 { 00:14:52.797 "name": null, 00:14:52.797 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:52.797 "is_configured": false, 00:14:52.797 "data_offset": 0, 00:14:52.797 "data_size": 7936 00:14:52.797 }, 00:14:52.797 { 00:14:52.797 "name": "BaseBdev2", 00:14:52.797 "uuid": "6f077572-90b7-568a-919c-93be71941c99", 00:14:52.797 "is_configured": true, 00:14:52.797 "data_offset": 256, 00:14:52.797 "data_size": 7936 00:14:52.797 } 00:14:52.797 ] 00:14:52.797 }' 00:14:52.797 23:48:40 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:52.797 23:48:40 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:14:53.366 23:48:41 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:14:53.366 23:48:41 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:53.366 23:48:41 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:14:53.366 [2024-11-26 23:48:41.242633] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:14:53.366 [2024-11-26 23:48:41.247567] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d00019c960 00:14:53.366 23:48:41 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:53.366 23:48:41 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@647 -- # sleep 1 00:14:53.366 [2024-11-26 23:48:41.249448] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:14:54.308 23:48:42 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:54.308 23:48:42 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:54.308 23:48:42 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:54.309 23:48:42 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:54.309 23:48:42 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:54.309 23:48:42 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:54.309 23:48:42 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:54.309 23:48:42 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:54.309 23:48:42 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:14:54.309 23:48:42 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:54.309 23:48:42 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:54.309 "name": "raid_bdev1", 00:14:54.309 "uuid": "ff5af955-a414-4ba0-8235-7545475fb4f4", 00:14:54.309 "strip_size_kb": 0, 00:14:54.309 "state": "online", 00:14:54.309 "raid_level": "raid1", 00:14:54.309 "superblock": true, 00:14:54.309 "num_base_bdevs": 2, 00:14:54.309 "num_base_bdevs_discovered": 2, 00:14:54.309 "num_base_bdevs_operational": 2, 00:14:54.309 "process": { 00:14:54.309 "type": "rebuild", 00:14:54.309 "target": "spare", 00:14:54.309 "progress": { 00:14:54.309 "blocks": 2560, 00:14:54.309 "percent": 32 00:14:54.309 } 00:14:54.309 }, 00:14:54.309 "base_bdevs_list": [ 00:14:54.309 { 00:14:54.309 "name": "spare", 00:14:54.309 "uuid": "094bedf0-80ad-5887-8b84-da8234f7bf8c", 00:14:54.309 "is_configured": true, 00:14:54.309 "data_offset": 256, 00:14:54.309 "data_size": 7936 00:14:54.309 }, 00:14:54.309 { 00:14:54.309 "name": "BaseBdev2", 00:14:54.309 "uuid": "6f077572-90b7-568a-919c-93be71941c99", 00:14:54.309 "is_configured": true, 00:14:54.309 "data_offset": 256, 00:14:54.309 "data_size": 7936 00:14:54.309 } 00:14:54.309 ] 00:14:54.309 }' 00:14:54.309 23:48:42 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:54.309 23:48:42 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:54.309 23:48:42 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:54.309 23:48:42 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:54.309 23:48:42 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:14:54.309 23:48:42 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:54.309 23:48:42 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:14:54.309 [2024-11-26 23:48:42.389854] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:14:54.570 [2024-11-26 23:48:42.453755] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:14:54.570 [2024-11-26 23:48:42.453813] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:54.570 [2024-11-26 23:48:42.453833] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:14:54.570 [2024-11-26 23:48:42.453841] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:14:54.570 23:48:42 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:54.570 23:48:42 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:14:54.570 23:48:42 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:54.570 23:48:42 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:54.570 23:48:42 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:14:54.570 23:48:42 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:14:54.570 23:48:42 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:14:54.570 23:48:42 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:54.570 23:48:42 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:54.570 23:48:42 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:54.570 23:48:42 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:54.570 23:48:42 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:54.570 23:48:42 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:54.570 23:48:42 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:54.570 23:48:42 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:14:54.570 23:48:42 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:54.570 23:48:42 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:54.570 "name": "raid_bdev1", 00:14:54.570 "uuid": "ff5af955-a414-4ba0-8235-7545475fb4f4", 00:14:54.570 "strip_size_kb": 0, 00:14:54.570 "state": "online", 00:14:54.570 "raid_level": "raid1", 00:14:54.570 "superblock": true, 00:14:54.570 "num_base_bdevs": 2, 00:14:54.570 "num_base_bdevs_discovered": 1, 00:14:54.570 "num_base_bdevs_operational": 1, 00:14:54.570 "base_bdevs_list": [ 00:14:54.570 { 00:14:54.570 "name": null, 00:14:54.570 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:54.570 "is_configured": false, 00:14:54.570 "data_offset": 0, 00:14:54.570 "data_size": 7936 00:14:54.570 }, 00:14:54.570 { 00:14:54.570 "name": "BaseBdev2", 00:14:54.570 "uuid": "6f077572-90b7-568a-919c-93be71941c99", 00:14:54.570 "is_configured": true, 00:14:54.570 "data_offset": 256, 00:14:54.570 "data_size": 7936 00:14:54.570 } 00:14:54.570 ] 00:14:54.570 }' 00:14:54.570 23:48:42 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:54.570 23:48:42 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:14:54.831 23:48:42 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:14:54.831 23:48:42 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:54.831 23:48:42 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:14:54.831 23:48:42 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=none 00:14:54.831 23:48:42 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:54.831 23:48:42 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:54.831 23:48:42 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:54.831 23:48:42 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:54.831 23:48:42 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:14:54.831 23:48:42 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:54.831 23:48:42 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:54.831 "name": "raid_bdev1", 00:14:54.831 "uuid": "ff5af955-a414-4ba0-8235-7545475fb4f4", 00:14:54.831 "strip_size_kb": 0, 00:14:54.831 "state": "online", 00:14:54.831 "raid_level": "raid1", 00:14:54.831 "superblock": true, 00:14:54.831 "num_base_bdevs": 2, 00:14:54.831 "num_base_bdevs_discovered": 1, 00:14:54.831 "num_base_bdevs_operational": 1, 00:14:54.831 "base_bdevs_list": [ 00:14:54.831 { 00:14:54.831 "name": null, 00:14:54.831 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:54.831 "is_configured": false, 00:14:54.831 "data_offset": 0, 00:14:54.831 "data_size": 7936 00:14:54.831 }, 00:14:54.831 { 00:14:54.831 "name": "BaseBdev2", 00:14:54.831 "uuid": "6f077572-90b7-568a-919c-93be71941c99", 00:14:54.831 "is_configured": true, 00:14:54.831 "data_offset": 256, 00:14:54.831 "data_size": 7936 00:14:54.831 } 00:14:54.831 ] 00:14:54.831 }' 00:14:54.831 23:48:42 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:54.831 23:48:42 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:14:54.831 23:48:42 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:55.091 23:48:42 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:14:55.091 23:48:42 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:14:55.091 23:48:42 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:55.091 23:48:42 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:14:55.091 [2024-11-26 23:48:42.997461] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:14:55.091 [2024-11-26 23:48:43.001794] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d00019ca30 00:14:55.091 23:48:43 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:55.091 23:48:43 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@663 -- # sleep 1 00:14:55.091 [2024-11-26 23:48:43.003625] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:14:56.034 23:48:44 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:56.034 23:48:44 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:56.034 23:48:44 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:56.034 23:48:44 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:56.034 23:48:44 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:56.034 23:48:44 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:56.034 23:48:44 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:56.034 23:48:44 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:56.034 23:48:44 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:14:56.034 23:48:44 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:56.034 23:48:44 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:56.034 "name": "raid_bdev1", 00:14:56.034 "uuid": "ff5af955-a414-4ba0-8235-7545475fb4f4", 00:14:56.034 "strip_size_kb": 0, 00:14:56.034 "state": "online", 00:14:56.034 "raid_level": "raid1", 00:14:56.034 "superblock": true, 00:14:56.034 "num_base_bdevs": 2, 00:14:56.034 "num_base_bdevs_discovered": 2, 00:14:56.034 "num_base_bdevs_operational": 2, 00:14:56.034 "process": { 00:14:56.034 "type": "rebuild", 00:14:56.034 "target": "spare", 00:14:56.034 "progress": { 00:14:56.034 "blocks": 2560, 00:14:56.034 "percent": 32 00:14:56.034 } 00:14:56.034 }, 00:14:56.034 "base_bdevs_list": [ 00:14:56.034 { 00:14:56.034 "name": "spare", 00:14:56.034 "uuid": "094bedf0-80ad-5887-8b84-da8234f7bf8c", 00:14:56.034 "is_configured": true, 00:14:56.034 "data_offset": 256, 00:14:56.034 "data_size": 7936 00:14:56.034 }, 00:14:56.034 { 00:14:56.034 "name": "BaseBdev2", 00:14:56.034 "uuid": "6f077572-90b7-568a-919c-93be71941c99", 00:14:56.034 "is_configured": true, 00:14:56.034 "data_offset": 256, 00:14:56.034 "data_size": 7936 00:14:56.034 } 00:14:56.034 ] 00:14:56.034 }' 00:14:56.034 23:48:44 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:56.034 23:48:44 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:56.034 23:48:44 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:56.034 23:48:44 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:56.034 23:48:44 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@666 -- # '[' true = true ']' 00:14:56.034 23:48:44 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@666 -- # '[' = false ']' 00:14:56.034 /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh: line 666: [: =: unary operator expected 00:14:56.034 23:48:44 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=2 00:14:56.034 23:48:44 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:14:56.034 23:48:44 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@693 -- # '[' 2 -gt 2 ']' 00:14:56.034 23:48:44 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@706 -- # local timeout=556 00:14:56.034 23:48:44 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:56.034 23:48:44 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:56.034 23:48:44 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:56.034 23:48:44 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:56.034 23:48:44 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:56.034 23:48:44 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:56.034 23:48:44 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:56.034 23:48:44 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:56.034 23:48:44 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:56.034 23:48:44 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:14:56.295 23:48:44 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:56.295 23:48:44 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:56.295 "name": "raid_bdev1", 00:14:56.295 "uuid": "ff5af955-a414-4ba0-8235-7545475fb4f4", 00:14:56.295 "strip_size_kb": 0, 00:14:56.295 "state": "online", 00:14:56.295 "raid_level": "raid1", 00:14:56.295 "superblock": true, 00:14:56.295 "num_base_bdevs": 2, 00:14:56.295 "num_base_bdevs_discovered": 2, 00:14:56.295 "num_base_bdevs_operational": 2, 00:14:56.295 "process": { 00:14:56.295 "type": "rebuild", 00:14:56.295 "target": "spare", 00:14:56.295 "progress": { 00:14:56.295 "blocks": 2816, 00:14:56.295 "percent": 35 00:14:56.295 } 00:14:56.295 }, 00:14:56.295 "base_bdevs_list": [ 00:14:56.295 { 00:14:56.295 "name": "spare", 00:14:56.295 "uuid": "094bedf0-80ad-5887-8b84-da8234f7bf8c", 00:14:56.295 "is_configured": true, 00:14:56.295 "data_offset": 256, 00:14:56.295 "data_size": 7936 00:14:56.295 }, 00:14:56.295 { 00:14:56.295 "name": "BaseBdev2", 00:14:56.295 "uuid": "6f077572-90b7-568a-919c-93be71941c99", 00:14:56.295 "is_configured": true, 00:14:56.295 "data_offset": 256, 00:14:56.295 "data_size": 7936 00:14:56.295 } 00:14:56.295 ] 00:14:56.295 }' 00:14:56.295 23:48:44 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:56.295 23:48:44 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:56.295 23:48:44 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:56.295 23:48:44 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:56.295 23:48:44 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@711 -- # sleep 1 00:14:57.235 23:48:45 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:57.235 23:48:45 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:57.235 23:48:45 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:57.235 23:48:45 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:57.235 23:48:45 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:57.235 23:48:45 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:57.235 23:48:45 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:57.235 23:48:45 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:57.235 23:48:45 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:57.235 23:48:45 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:14:57.235 23:48:45 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:57.235 23:48:45 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:57.235 "name": "raid_bdev1", 00:14:57.235 "uuid": "ff5af955-a414-4ba0-8235-7545475fb4f4", 00:14:57.235 "strip_size_kb": 0, 00:14:57.235 "state": "online", 00:14:57.235 "raid_level": "raid1", 00:14:57.235 "superblock": true, 00:14:57.235 "num_base_bdevs": 2, 00:14:57.235 "num_base_bdevs_discovered": 2, 00:14:57.235 "num_base_bdevs_operational": 2, 00:14:57.235 "process": { 00:14:57.235 "type": "rebuild", 00:14:57.235 "target": "spare", 00:14:57.235 "progress": { 00:14:57.235 "blocks": 5632, 00:14:57.235 "percent": 70 00:14:57.235 } 00:14:57.235 }, 00:14:57.235 "base_bdevs_list": [ 00:14:57.235 { 00:14:57.235 "name": "spare", 00:14:57.235 "uuid": "094bedf0-80ad-5887-8b84-da8234f7bf8c", 00:14:57.235 "is_configured": true, 00:14:57.235 "data_offset": 256, 00:14:57.235 "data_size": 7936 00:14:57.235 }, 00:14:57.235 { 00:14:57.235 "name": "BaseBdev2", 00:14:57.235 "uuid": "6f077572-90b7-568a-919c-93be71941c99", 00:14:57.235 "is_configured": true, 00:14:57.235 "data_offset": 256, 00:14:57.235 "data_size": 7936 00:14:57.235 } 00:14:57.235 ] 00:14:57.235 }' 00:14:57.235 23:48:45 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:57.495 23:48:45 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:57.495 23:48:45 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:57.495 23:48:45 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:57.495 23:48:45 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@711 -- # sleep 1 00:14:58.064 [2024-11-26 23:48:46.113371] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:14:58.064 [2024-11-26 23:48:46.113449] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:14:58.064 [2024-11-26 23:48:46.113537] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:58.325 23:48:46 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:58.325 23:48:46 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:58.325 23:48:46 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:58.325 23:48:46 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:58.325 23:48:46 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:58.325 23:48:46 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:58.325 23:48:46 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:58.325 23:48:46 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:58.325 23:48:46 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:58.325 23:48:46 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:14:58.325 23:48:46 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:58.585 23:48:46 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:58.585 "name": "raid_bdev1", 00:14:58.585 "uuid": "ff5af955-a414-4ba0-8235-7545475fb4f4", 00:14:58.585 "strip_size_kb": 0, 00:14:58.585 "state": "online", 00:14:58.585 "raid_level": "raid1", 00:14:58.585 "superblock": true, 00:14:58.585 "num_base_bdevs": 2, 00:14:58.585 "num_base_bdevs_discovered": 2, 00:14:58.585 "num_base_bdevs_operational": 2, 00:14:58.585 "base_bdevs_list": [ 00:14:58.585 { 00:14:58.585 "name": "spare", 00:14:58.585 "uuid": "094bedf0-80ad-5887-8b84-da8234f7bf8c", 00:14:58.585 "is_configured": true, 00:14:58.585 "data_offset": 256, 00:14:58.585 "data_size": 7936 00:14:58.585 }, 00:14:58.585 { 00:14:58.585 "name": "BaseBdev2", 00:14:58.585 "uuid": "6f077572-90b7-568a-919c-93be71941c99", 00:14:58.585 "is_configured": true, 00:14:58.585 "data_offset": 256, 00:14:58.585 "data_size": 7936 00:14:58.585 } 00:14:58.585 ] 00:14:58.585 }' 00:14:58.585 23:48:46 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:58.585 23:48:46 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:14:58.585 23:48:46 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:58.585 23:48:46 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:14:58.585 23:48:46 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@709 -- # break 00:14:58.585 23:48:46 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:14:58.585 23:48:46 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:58.585 23:48:46 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:14:58.585 23:48:46 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=none 00:14:58.585 23:48:46 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:58.585 23:48:46 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:58.585 23:48:46 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:58.585 23:48:46 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:58.585 23:48:46 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:14:58.586 23:48:46 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:58.586 23:48:46 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:58.586 "name": "raid_bdev1", 00:14:58.586 "uuid": "ff5af955-a414-4ba0-8235-7545475fb4f4", 00:14:58.586 "strip_size_kb": 0, 00:14:58.586 "state": "online", 00:14:58.586 "raid_level": "raid1", 00:14:58.586 "superblock": true, 00:14:58.586 "num_base_bdevs": 2, 00:14:58.586 "num_base_bdevs_discovered": 2, 00:14:58.586 "num_base_bdevs_operational": 2, 00:14:58.586 "base_bdevs_list": [ 00:14:58.586 { 00:14:58.586 "name": "spare", 00:14:58.586 "uuid": "094bedf0-80ad-5887-8b84-da8234f7bf8c", 00:14:58.586 "is_configured": true, 00:14:58.586 "data_offset": 256, 00:14:58.586 "data_size": 7936 00:14:58.586 }, 00:14:58.586 { 00:14:58.586 "name": "BaseBdev2", 00:14:58.586 "uuid": "6f077572-90b7-568a-919c-93be71941c99", 00:14:58.586 "is_configured": true, 00:14:58.586 "data_offset": 256, 00:14:58.586 "data_size": 7936 00:14:58.586 } 00:14:58.586 ] 00:14:58.586 }' 00:14:58.586 23:48:46 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:58.586 23:48:46 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:14:58.586 23:48:46 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:58.586 23:48:46 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:14:58.586 23:48:46 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:14:58.586 23:48:46 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:58.586 23:48:46 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:58.586 23:48:46 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:14:58.586 23:48:46 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:14:58.586 23:48:46 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:14:58.586 23:48:46 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:58.586 23:48:46 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:58.586 23:48:46 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:58.586 23:48:46 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:58.586 23:48:46 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:58.586 23:48:46 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:58.586 23:48:46 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:58.586 23:48:46 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:14:58.586 23:48:46 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:58.852 23:48:46 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:58.852 "name": "raid_bdev1", 00:14:58.852 "uuid": "ff5af955-a414-4ba0-8235-7545475fb4f4", 00:14:58.852 "strip_size_kb": 0, 00:14:58.852 "state": "online", 00:14:58.852 "raid_level": "raid1", 00:14:58.852 "superblock": true, 00:14:58.852 "num_base_bdevs": 2, 00:14:58.852 "num_base_bdevs_discovered": 2, 00:14:58.852 "num_base_bdevs_operational": 2, 00:14:58.852 "base_bdevs_list": [ 00:14:58.852 { 00:14:58.852 "name": "spare", 00:14:58.852 "uuid": "094bedf0-80ad-5887-8b84-da8234f7bf8c", 00:14:58.852 "is_configured": true, 00:14:58.852 "data_offset": 256, 00:14:58.852 "data_size": 7936 00:14:58.852 }, 00:14:58.852 { 00:14:58.852 "name": "BaseBdev2", 00:14:58.852 "uuid": "6f077572-90b7-568a-919c-93be71941c99", 00:14:58.852 "is_configured": true, 00:14:58.852 "data_offset": 256, 00:14:58.852 "data_size": 7936 00:14:58.852 } 00:14:58.852 ] 00:14:58.852 }' 00:14:58.852 23:48:46 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:58.852 23:48:46 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:14:59.113 23:48:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:14:59.113 23:48:47 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:59.114 23:48:47 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:14:59.114 [2024-11-26 23:48:47.127889] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:14:59.114 [2024-11-26 23:48:47.127917] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:14:59.114 [2024-11-26 23:48:47.127994] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:14:59.114 [2024-11-26 23:48:47.128051] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:14:59.114 [2024-11-26 23:48:47.128068] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:14:59.114 23:48:47 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:59.114 23:48:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:59.114 23:48:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@720 -- # jq length 00:14:59.114 23:48:47 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:59.114 23:48:47 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:14:59.114 23:48:47 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:59.114 23:48:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:14:59.114 23:48:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:14:59.114 23:48:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:14:59.114 23:48:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:14:59.114 23:48:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:14:59.114 23:48:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:14:59.114 23:48:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@10 -- # local bdev_list 00:14:59.114 23:48:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:14:59.114 23:48:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@11 -- # local nbd_list 00:14:59.114 23:48:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@12 -- # local i 00:14:59.114 23:48:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:14:59.114 23:48:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:14:59.114 23:48:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:14:59.374 /dev/nbd0 00:14:59.374 23:48:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:14:59.374 23:48:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:14:59.374 23:48:47 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:14:59.374 23:48:47 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@873 -- # local i 00:14:59.374 23:48:47 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:14:59.374 23:48:47 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:14:59.374 23:48:47 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:14:59.374 23:48:47 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@877 -- # break 00:14:59.374 23:48:47 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:14:59.374 23:48:47 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:14:59.374 23:48:47 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:14:59.374 1+0 records in 00:14:59.374 1+0 records out 00:14:59.374 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000463978 s, 8.8 MB/s 00:14:59.374 23:48:47 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:59.374 23:48:47 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@890 -- # size=4096 00:14:59.374 23:48:47 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:59.374 23:48:47 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:14:59.374 23:48:47 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@893 -- # return 0 00:14:59.374 23:48:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:14:59.374 23:48:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:14:59.374 23:48:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:14:59.634 /dev/nbd1 00:14:59.634 23:48:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:14:59.634 23:48:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:14:59.634 23:48:47 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:14:59.634 23:48:47 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@873 -- # local i 00:14:59.634 23:48:47 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:14:59.634 23:48:47 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:14:59.634 23:48:47 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:14:59.634 23:48:47 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@877 -- # break 00:14:59.634 23:48:47 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:14:59.634 23:48:47 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:14:59.634 23:48:47 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:14:59.634 1+0 records in 00:14:59.634 1+0 records out 00:14:59.634 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000428917 s, 9.5 MB/s 00:14:59.634 23:48:47 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:59.634 23:48:47 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@890 -- # size=4096 00:14:59.634 23:48:47 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:59.634 23:48:47 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:14:59.634 23:48:47 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@893 -- # return 0 00:14:59.635 23:48:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:14:59.635 23:48:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:14:59.635 23:48:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@738 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:14:59.635 23:48:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:14:59.635 23:48:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:14:59.635 23:48:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:14:59.635 23:48:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@50 -- # local nbd_list 00:14:59.635 23:48:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@51 -- # local i 00:14:59.635 23:48:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:14:59.635 23:48:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:14:59.894 23:48:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:14:59.894 23:48:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:14:59.894 23:48:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:14:59.894 23:48:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:14:59.894 23:48:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:14:59.894 23:48:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:14:59.894 23:48:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@41 -- # break 00:14:59.894 23:48:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@45 -- # return 0 00:14:59.894 23:48:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:14:59.894 23:48:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:15:00.155 23:48:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:15:00.155 23:48:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:15:00.155 23:48:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:15:00.155 23:48:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:15:00.155 23:48:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:15:00.155 23:48:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:15:00.155 23:48:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@41 -- # break 00:15:00.155 23:48:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@45 -- # return 0 00:15:00.155 23:48:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@743 -- # '[' true = true ']' 00:15:00.155 23:48:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@745 -- # rpc_cmd bdev_passthru_delete spare 00:15:00.155 23:48:48 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:00.155 23:48:48 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:00.155 23:48:48 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:00.155 23:48:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@746 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:15:00.155 23:48:48 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:00.155 23:48:48 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:00.155 [2024-11-26 23:48:48.159852] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:15:00.155 [2024-11-26 23:48:48.160287] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:00.155 [2024-11-26 23:48:48.160376] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:15:00.155 [2024-11-26 23:48:48.160442] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:00.155 [2024-11-26 23:48:48.162594] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:00.155 [2024-11-26 23:48:48.162731] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:15:00.155 [2024-11-26 23:48:48.162914] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:15:00.155 [2024-11-26 23:48:48.162960] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:15:00.155 [2024-11-26 23:48:48.163080] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:15:00.155 spare 00:15:00.155 23:48:48 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:00.155 23:48:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@747 -- # rpc_cmd bdev_wait_for_examine 00:15:00.155 23:48:48 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:00.155 23:48:48 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:00.155 [2024-11-26 23:48:48.262970] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001580 00:15:00.155 [2024-11-26 23:48:48.262994] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:15:00.155 [2024-11-26 23:48:48.263236] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0001bb1b0 00:15:00.155 [2024-11-26 23:48:48.263382] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001580 00:15:00.155 [2024-11-26 23:48:48.263398] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001580 00:15:00.155 [2024-11-26 23:48:48.263523] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:00.155 23:48:48 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:00.155 23:48:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@749 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:15:00.155 23:48:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:00.155 23:48:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:00.155 23:48:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:00.155 23:48:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:00.155 23:48:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:00.155 23:48:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:00.155 23:48:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:00.155 23:48:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:00.155 23:48:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:00.155 23:48:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:00.155 23:48:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:00.155 23:48:48 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:00.155 23:48:48 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:00.414 23:48:48 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:00.414 23:48:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:00.414 "name": "raid_bdev1", 00:15:00.414 "uuid": "ff5af955-a414-4ba0-8235-7545475fb4f4", 00:15:00.414 "strip_size_kb": 0, 00:15:00.414 "state": "online", 00:15:00.414 "raid_level": "raid1", 00:15:00.414 "superblock": true, 00:15:00.414 "num_base_bdevs": 2, 00:15:00.414 "num_base_bdevs_discovered": 2, 00:15:00.414 "num_base_bdevs_operational": 2, 00:15:00.415 "base_bdevs_list": [ 00:15:00.415 { 00:15:00.415 "name": "spare", 00:15:00.415 "uuid": "094bedf0-80ad-5887-8b84-da8234f7bf8c", 00:15:00.415 "is_configured": true, 00:15:00.415 "data_offset": 256, 00:15:00.415 "data_size": 7936 00:15:00.415 }, 00:15:00.415 { 00:15:00.415 "name": "BaseBdev2", 00:15:00.415 "uuid": "6f077572-90b7-568a-919c-93be71941c99", 00:15:00.415 "is_configured": true, 00:15:00.415 "data_offset": 256, 00:15:00.415 "data_size": 7936 00:15:00.415 } 00:15:00.415 ] 00:15:00.415 }' 00:15:00.415 23:48:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:00.415 23:48:48 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:00.674 23:48:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@750 -- # verify_raid_bdev_process raid_bdev1 none none 00:15:00.674 23:48:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:00.674 23:48:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:15:00.674 23:48:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=none 00:15:00.674 23:48:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:00.674 23:48:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:00.674 23:48:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:00.674 23:48:48 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:00.674 23:48:48 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:00.674 23:48:48 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:00.674 23:48:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:00.674 "name": "raid_bdev1", 00:15:00.674 "uuid": "ff5af955-a414-4ba0-8235-7545475fb4f4", 00:15:00.674 "strip_size_kb": 0, 00:15:00.674 "state": "online", 00:15:00.674 "raid_level": "raid1", 00:15:00.674 "superblock": true, 00:15:00.674 "num_base_bdevs": 2, 00:15:00.674 "num_base_bdevs_discovered": 2, 00:15:00.674 "num_base_bdevs_operational": 2, 00:15:00.674 "base_bdevs_list": [ 00:15:00.674 { 00:15:00.674 "name": "spare", 00:15:00.674 "uuid": "094bedf0-80ad-5887-8b84-da8234f7bf8c", 00:15:00.675 "is_configured": true, 00:15:00.675 "data_offset": 256, 00:15:00.675 "data_size": 7936 00:15:00.675 }, 00:15:00.675 { 00:15:00.675 "name": "BaseBdev2", 00:15:00.675 "uuid": "6f077572-90b7-568a-919c-93be71941c99", 00:15:00.675 "is_configured": true, 00:15:00.675 "data_offset": 256, 00:15:00.675 "data_size": 7936 00:15:00.675 } 00:15:00.675 ] 00:15:00.675 }' 00:15:00.675 23:48:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:00.675 23:48:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:15:00.675 23:48:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:00.934 23:48:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:15:00.934 23:48:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@751 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:00.934 23:48:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@751 -- # jq -r '.[].base_bdevs_list[0].name' 00:15:00.934 23:48:48 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:00.934 23:48:48 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:00.934 23:48:48 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:00.934 23:48:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@751 -- # [[ spare == \s\p\a\r\e ]] 00:15:00.934 23:48:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@754 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:15:00.934 23:48:48 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:00.934 23:48:48 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:00.934 [2024-11-26 23:48:48.870784] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:00.934 23:48:48 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:00.935 23:48:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@755 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:15:00.935 23:48:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:00.935 23:48:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:00.935 23:48:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:00.935 23:48:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:00.935 23:48:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:15:00.935 23:48:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:00.935 23:48:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:00.935 23:48:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:00.935 23:48:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:00.935 23:48:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:00.935 23:48:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:00.935 23:48:48 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:00.935 23:48:48 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:00.935 23:48:48 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:00.935 23:48:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:00.935 "name": "raid_bdev1", 00:15:00.935 "uuid": "ff5af955-a414-4ba0-8235-7545475fb4f4", 00:15:00.935 "strip_size_kb": 0, 00:15:00.935 "state": "online", 00:15:00.935 "raid_level": "raid1", 00:15:00.935 "superblock": true, 00:15:00.935 "num_base_bdevs": 2, 00:15:00.935 "num_base_bdevs_discovered": 1, 00:15:00.935 "num_base_bdevs_operational": 1, 00:15:00.935 "base_bdevs_list": [ 00:15:00.935 { 00:15:00.935 "name": null, 00:15:00.935 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:00.935 "is_configured": false, 00:15:00.935 "data_offset": 0, 00:15:00.935 "data_size": 7936 00:15:00.935 }, 00:15:00.935 { 00:15:00.935 "name": "BaseBdev2", 00:15:00.935 "uuid": "6f077572-90b7-568a-919c-93be71941c99", 00:15:00.935 "is_configured": true, 00:15:00.935 "data_offset": 256, 00:15:00.935 "data_size": 7936 00:15:00.935 } 00:15:00.935 ] 00:15:00.935 }' 00:15:00.935 23:48:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:00.935 23:48:48 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:01.205 23:48:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@756 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:15:01.205 23:48:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:01.206 23:48:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:01.206 [2024-11-26 23:48:49.298138] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:15:01.206 [2024-11-26 23:48:49.298368] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:15:01.206 [2024-11-26 23:48:49.298428] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:15:01.206 [2024-11-26 23:48:49.298754] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:15:01.206 [2024-11-26 23:48:49.303537] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0001bb280 00:15:01.206 23:48:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:01.206 23:48:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@757 -- # sleep 1 00:15:01.206 [2024-11-26 23:48:49.305390] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:15:02.601 23:48:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@758 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:02.601 23:48:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:02.601 23:48:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:02.601 23:48:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:02.601 23:48:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:02.601 23:48:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:02.601 23:48:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:02.601 23:48:50 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:02.601 23:48:50 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:02.601 23:48:50 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:02.601 23:48:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:02.601 "name": "raid_bdev1", 00:15:02.601 "uuid": "ff5af955-a414-4ba0-8235-7545475fb4f4", 00:15:02.601 "strip_size_kb": 0, 00:15:02.601 "state": "online", 00:15:02.601 "raid_level": "raid1", 00:15:02.601 "superblock": true, 00:15:02.601 "num_base_bdevs": 2, 00:15:02.601 "num_base_bdevs_discovered": 2, 00:15:02.601 "num_base_bdevs_operational": 2, 00:15:02.601 "process": { 00:15:02.601 "type": "rebuild", 00:15:02.601 "target": "spare", 00:15:02.601 "progress": { 00:15:02.601 "blocks": 2560, 00:15:02.601 "percent": 32 00:15:02.601 } 00:15:02.601 }, 00:15:02.601 "base_bdevs_list": [ 00:15:02.601 { 00:15:02.601 "name": "spare", 00:15:02.601 "uuid": "094bedf0-80ad-5887-8b84-da8234f7bf8c", 00:15:02.601 "is_configured": true, 00:15:02.601 "data_offset": 256, 00:15:02.601 "data_size": 7936 00:15:02.601 }, 00:15:02.601 { 00:15:02.601 "name": "BaseBdev2", 00:15:02.601 "uuid": "6f077572-90b7-568a-919c-93be71941c99", 00:15:02.601 "is_configured": true, 00:15:02.601 "data_offset": 256, 00:15:02.601 "data_size": 7936 00:15:02.601 } 00:15:02.601 ] 00:15:02.601 }' 00:15:02.601 23:48:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:02.601 23:48:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:02.601 23:48:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:02.601 23:48:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:02.601 23:48:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@761 -- # rpc_cmd bdev_passthru_delete spare 00:15:02.601 23:48:50 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:02.601 23:48:50 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:02.601 [2024-11-26 23:48:50.441617] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:02.601 [2024-11-26 23:48:50.509302] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:15:02.601 [2024-11-26 23:48:50.509673] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:02.601 [2024-11-26 23:48:50.509701] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:02.601 [2024-11-26 23:48:50.509709] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:15:02.601 23:48:50 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:02.601 23:48:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@762 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:15:02.601 23:48:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:02.601 23:48:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:02.601 23:48:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:02.601 23:48:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:02.601 23:48:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:15:02.601 23:48:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:02.601 23:48:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:02.601 23:48:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:02.601 23:48:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:02.601 23:48:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:02.601 23:48:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:02.601 23:48:50 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:02.601 23:48:50 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:02.601 23:48:50 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:02.601 23:48:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:02.601 "name": "raid_bdev1", 00:15:02.601 "uuid": "ff5af955-a414-4ba0-8235-7545475fb4f4", 00:15:02.601 "strip_size_kb": 0, 00:15:02.601 "state": "online", 00:15:02.601 "raid_level": "raid1", 00:15:02.601 "superblock": true, 00:15:02.601 "num_base_bdevs": 2, 00:15:02.601 "num_base_bdevs_discovered": 1, 00:15:02.601 "num_base_bdevs_operational": 1, 00:15:02.601 "base_bdevs_list": [ 00:15:02.601 { 00:15:02.601 "name": null, 00:15:02.602 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:02.602 "is_configured": false, 00:15:02.602 "data_offset": 0, 00:15:02.602 "data_size": 7936 00:15:02.602 }, 00:15:02.602 { 00:15:02.602 "name": "BaseBdev2", 00:15:02.602 "uuid": "6f077572-90b7-568a-919c-93be71941c99", 00:15:02.602 "is_configured": true, 00:15:02.602 "data_offset": 256, 00:15:02.602 "data_size": 7936 00:15:02.602 } 00:15:02.602 ] 00:15:02.602 }' 00:15:02.602 23:48:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:02.602 23:48:50 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:02.862 23:48:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@763 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:15:02.862 23:48:50 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:02.862 23:48:50 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:02.862 [2024-11-26 23:48:50.921327] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:15:02.862 [2024-11-26 23:48:50.921485] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:02.862 [2024-11-26 23:48:50.921543] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009f80 00:15:02.862 [2024-11-26 23:48:50.921604] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:02.862 [2024-11-26 23:48:50.922072] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:02.862 [2024-11-26 23:48:50.922205] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:15:02.862 [2024-11-26 23:48:50.922334] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:15:02.862 [2024-11-26 23:48:50.922366] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:15:02.862 [2024-11-26 23:48:50.922392] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:15:02.862 [2024-11-26 23:48:50.922505] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:15:02.862 [2024-11-26 23:48:50.926580] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0001bb350 00:15:02.862 spare 00:15:02.862 23:48:50 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:02.862 23:48:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@764 -- # sleep 1 00:15:02.862 [2024-11-26 23:48:50.928431] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:15:04.245 23:48:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@765 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:04.245 23:48:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:04.245 23:48:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:04.245 23:48:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:04.245 23:48:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:04.245 23:48:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:04.245 23:48:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:04.245 23:48:51 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:04.245 23:48:51 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:04.245 23:48:51 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:04.245 23:48:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:04.245 "name": "raid_bdev1", 00:15:04.245 "uuid": "ff5af955-a414-4ba0-8235-7545475fb4f4", 00:15:04.245 "strip_size_kb": 0, 00:15:04.245 "state": "online", 00:15:04.245 "raid_level": "raid1", 00:15:04.245 "superblock": true, 00:15:04.245 "num_base_bdevs": 2, 00:15:04.245 "num_base_bdevs_discovered": 2, 00:15:04.245 "num_base_bdevs_operational": 2, 00:15:04.245 "process": { 00:15:04.245 "type": "rebuild", 00:15:04.245 "target": "spare", 00:15:04.245 "progress": { 00:15:04.245 "blocks": 2560, 00:15:04.245 "percent": 32 00:15:04.245 } 00:15:04.245 }, 00:15:04.245 "base_bdevs_list": [ 00:15:04.245 { 00:15:04.245 "name": "spare", 00:15:04.245 "uuid": "094bedf0-80ad-5887-8b84-da8234f7bf8c", 00:15:04.245 "is_configured": true, 00:15:04.245 "data_offset": 256, 00:15:04.245 "data_size": 7936 00:15:04.245 }, 00:15:04.245 { 00:15:04.245 "name": "BaseBdev2", 00:15:04.245 "uuid": "6f077572-90b7-568a-919c-93be71941c99", 00:15:04.245 "is_configured": true, 00:15:04.245 "data_offset": 256, 00:15:04.245 "data_size": 7936 00:15:04.245 } 00:15:04.245 ] 00:15:04.245 }' 00:15:04.245 23:48:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:04.245 23:48:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:04.245 23:48:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:04.245 23:48:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:04.245 23:48:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@768 -- # rpc_cmd bdev_passthru_delete spare 00:15:04.245 23:48:52 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:04.245 23:48:52 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:04.245 [2024-11-26 23:48:52.068503] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:04.245 [2024-11-26 23:48:52.132299] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:15:04.245 [2024-11-26 23:48:52.132651] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:04.245 [2024-11-26 23:48:52.132669] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:04.245 [2024-11-26 23:48:52.132679] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:15:04.245 23:48:52 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:04.245 23:48:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@769 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:15:04.245 23:48:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:04.245 23:48:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:04.245 23:48:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:04.245 23:48:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:04.245 23:48:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:15:04.245 23:48:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:04.245 23:48:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:04.245 23:48:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:04.245 23:48:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:04.245 23:48:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:04.245 23:48:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:04.245 23:48:52 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:04.245 23:48:52 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:04.245 23:48:52 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:04.245 23:48:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:04.245 "name": "raid_bdev1", 00:15:04.245 "uuid": "ff5af955-a414-4ba0-8235-7545475fb4f4", 00:15:04.245 "strip_size_kb": 0, 00:15:04.245 "state": "online", 00:15:04.245 "raid_level": "raid1", 00:15:04.245 "superblock": true, 00:15:04.245 "num_base_bdevs": 2, 00:15:04.245 "num_base_bdevs_discovered": 1, 00:15:04.245 "num_base_bdevs_operational": 1, 00:15:04.245 "base_bdevs_list": [ 00:15:04.245 { 00:15:04.245 "name": null, 00:15:04.245 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:04.245 "is_configured": false, 00:15:04.245 "data_offset": 0, 00:15:04.245 "data_size": 7936 00:15:04.245 }, 00:15:04.245 { 00:15:04.245 "name": "BaseBdev2", 00:15:04.245 "uuid": "6f077572-90b7-568a-919c-93be71941c99", 00:15:04.245 "is_configured": true, 00:15:04.245 "data_offset": 256, 00:15:04.245 "data_size": 7936 00:15:04.245 } 00:15:04.245 ] 00:15:04.245 }' 00:15:04.245 23:48:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:04.245 23:48:52 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:04.504 23:48:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@770 -- # verify_raid_bdev_process raid_bdev1 none none 00:15:04.504 23:48:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:04.504 23:48:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:15:04.504 23:48:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=none 00:15:04.504 23:48:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:04.504 23:48:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:04.504 23:48:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:04.504 23:48:52 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:04.504 23:48:52 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:04.504 23:48:52 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:04.504 23:48:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:04.504 "name": "raid_bdev1", 00:15:04.504 "uuid": "ff5af955-a414-4ba0-8235-7545475fb4f4", 00:15:04.504 "strip_size_kb": 0, 00:15:04.504 "state": "online", 00:15:04.504 "raid_level": "raid1", 00:15:04.504 "superblock": true, 00:15:04.504 "num_base_bdevs": 2, 00:15:04.504 "num_base_bdevs_discovered": 1, 00:15:04.504 "num_base_bdevs_operational": 1, 00:15:04.504 "base_bdevs_list": [ 00:15:04.504 { 00:15:04.504 "name": null, 00:15:04.504 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:04.504 "is_configured": false, 00:15:04.504 "data_offset": 0, 00:15:04.504 "data_size": 7936 00:15:04.504 }, 00:15:04.504 { 00:15:04.504 "name": "BaseBdev2", 00:15:04.504 "uuid": "6f077572-90b7-568a-919c-93be71941c99", 00:15:04.504 "is_configured": true, 00:15:04.504 "data_offset": 256, 00:15:04.504 "data_size": 7936 00:15:04.504 } 00:15:04.504 ] 00:15:04.504 }' 00:15:04.504 23:48:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:04.763 23:48:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:15:04.763 23:48:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:04.763 23:48:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:15:04.763 23:48:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@773 -- # rpc_cmd bdev_passthru_delete BaseBdev1 00:15:04.763 23:48:52 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:04.763 23:48:52 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:04.763 23:48:52 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:04.763 23:48:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@774 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:15:04.763 23:48:52 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:04.763 23:48:52 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:04.763 [2024-11-26 23:48:52.719935] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:15:04.763 [2024-11-26 23:48:52.720145] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:04.763 [2024-11-26 23:48:52.720202] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a580 00:15:04.763 [2024-11-26 23:48:52.720251] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:04.763 [2024-11-26 23:48:52.720675] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:04.763 [2024-11-26 23:48:52.720796] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:15:04.763 [2024-11-26 23:48:52.720889] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:15:04.763 [2024-11-26 23:48:52.721004] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:15:04.763 [2024-11-26 23:48:52.721043] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:15:04.763 [2024-11-26 23:48:52.721102] bdev_raid.c:3894:raid_bdev_examine_done: *ERROR*: Failed to examine bdev BaseBdev1: Invalid argument 00:15:04.763 BaseBdev1 00:15:04.763 23:48:52 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:04.763 23:48:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@775 -- # sleep 1 00:15:05.702 23:48:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@776 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:15:05.702 23:48:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:05.702 23:48:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:05.702 23:48:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:05.702 23:48:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:05.702 23:48:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:15:05.702 23:48:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:05.702 23:48:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:05.702 23:48:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:05.702 23:48:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:05.702 23:48:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:05.702 23:48:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:05.702 23:48:53 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:05.702 23:48:53 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:05.702 23:48:53 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:05.702 23:48:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:05.702 "name": "raid_bdev1", 00:15:05.702 "uuid": "ff5af955-a414-4ba0-8235-7545475fb4f4", 00:15:05.702 "strip_size_kb": 0, 00:15:05.702 "state": "online", 00:15:05.702 "raid_level": "raid1", 00:15:05.702 "superblock": true, 00:15:05.702 "num_base_bdevs": 2, 00:15:05.702 "num_base_bdevs_discovered": 1, 00:15:05.702 "num_base_bdevs_operational": 1, 00:15:05.702 "base_bdevs_list": [ 00:15:05.702 { 00:15:05.702 "name": null, 00:15:05.702 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:05.702 "is_configured": false, 00:15:05.702 "data_offset": 0, 00:15:05.702 "data_size": 7936 00:15:05.702 }, 00:15:05.702 { 00:15:05.702 "name": "BaseBdev2", 00:15:05.702 "uuid": "6f077572-90b7-568a-919c-93be71941c99", 00:15:05.702 "is_configured": true, 00:15:05.702 "data_offset": 256, 00:15:05.702 "data_size": 7936 00:15:05.702 } 00:15:05.702 ] 00:15:05.702 }' 00:15:05.702 23:48:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:05.702 23:48:53 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:06.273 23:48:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@777 -- # verify_raid_bdev_process raid_bdev1 none none 00:15:06.273 23:48:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:06.273 23:48:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:15:06.273 23:48:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=none 00:15:06.273 23:48:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:06.273 23:48:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:06.273 23:48:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:06.273 23:48:54 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:06.274 23:48:54 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:06.274 23:48:54 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:06.274 23:48:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:06.274 "name": "raid_bdev1", 00:15:06.274 "uuid": "ff5af955-a414-4ba0-8235-7545475fb4f4", 00:15:06.274 "strip_size_kb": 0, 00:15:06.274 "state": "online", 00:15:06.274 "raid_level": "raid1", 00:15:06.274 "superblock": true, 00:15:06.274 "num_base_bdevs": 2, 00:15:06.274 "num_base_bdevs_discovered": 1, 00:15:06.274 "num_base_bdevs_operational": 1, 00:15:06.274 "base_bdevs_list": [ 00:15:06.274 { 00:15:06.274 "name": null, 00:15:06.274 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:06.274 "is_configured": false, 00:15:06.274 "data_offset": 0, 00:15:06.274 "data_size": 7936 00:15:06.274 }, 00:15:06.274 { 00:15:06.274 "name": "BaseBdev2", 00:15:06.274 "uuid": "6f077572-90b7-568a-919c-93be71941c99", 00:15:06.274 "is_configured": true, 00:15:06.274 "data_offset": 256, 00:15:06.274 "data_size": 7936 00:15:06.274 } 00:15:06.274 ] 00:15:06.274 }' 00:15:06.274 23:48:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:06.274 23:48:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:15:06.274 23:48:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:06.274 23:48:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:15:06.274 23:48:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@778 -- # NOT rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:15:06.274 23:48:54 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@652 -- # local es=0 00:15:06.274 23:48:54 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:15:06.274 23:48:54 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:15:06.274 23:48:54 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:15:06.274 23:48:54 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:15:06.274 23:48:54 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:15:06.274 23:48:54 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:15:06.274 23:48:54 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:06.274 23:48:54 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:06.274 [2024-11-26 23:48:54.289283] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:15:06.274 [2024-11-26 23:48:54.289505] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:15:06.274 [2024-11-26 23:48:54.289521] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:15:06.274 request: 00:15:06.274 { 00:15:06.274 "base_bdev": "BaseBdev1", 00:15:06.274 "raid_bdev": "raid_bdev1", 00:15:06.274 "method": "bdev_raid_add_base_bdev", 00:15:06.274 "req_id": 1 00:15:06.274 } 00:15:06.274 Got JSON-RPC error response 00:15:06.274 response: 00:15:06.274 { 00:15:06.274 "code": -22, 00:15:06.274 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:15:06.274 } 00:15:06.274 23:48:54 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:15:06.274 23:48:54 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@655 -- # es=1 00:15:06.274 23:48:54 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:15:06.274 23:48:54 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:15:06.274 23:48:54 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:15:06.274 23:48:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@779 -- # sleep 1 00:15:07.214 23:48:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@780 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:15:07.214 23:48:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:07.214 23:48:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:07.214 23:48:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:07.214 23:48:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:07.214 23:48:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:15:07.214 23:48:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:07.214 23:48:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:07.214 23:48:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:07.214 23:48:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:07.214 23:48:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:07.214 23:48:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:07.214 23:48:55 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:07.214 23:48:55 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:07.214 23:48:55 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:07.474 23:48:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:07.474 "name": "raid_bdev1", 00:15:07.474 "uuid": "ff5af955-a414-4ba0-8235-7545475fb4f4", 00:15:07.474 "strip_size_kb": 0, 00:15:07.474 "state": "online", 00:15:07.474 "raid_level": "raid1", 00:15:07.474 "superblock": true, 00:15:07.474 "num_base_bdevs": 2, 00:15:07.474 "num_base_bdevs_discovered": 1, 00:15:07.474 "num_base_bdevs_operational": 1, 00:15:07.474 "base_bdevs_list": [ 00:15:07.474 { 00:15:07.474 "name": null, 00:15:07.474 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:07.474 "is_configured": false, 00:15:07.474 "data_offset": 0, 00:15:07.474 "data_size": 7936 00:15:07.474 }, 00:15:07.474 { 00:15:07.474 "name": "BaseBdev2", 00:15:07.474 "uuid": "6f077572-90b7-568a-919c-93be71941c99", 00:15:07.474 "is_configured": true, 00:15:07.474 "data_offset": 256, 00:15:07.474 "data_size": 7936 00:15:07.474 } 00:15:07.474 ] 00:15:07.474 }' 00:15:07.474 23:48:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:07.474 23:48:55 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:07.735 23:48:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@781 -- # verify_raid_bdev_process raid_bdev1 none none 00:15:07.735 23:48:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:07.735 23:48:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:15:07.735 23:48:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=none 00:15:07.735 23:48:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:07.735 23:48:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:07.735 23:48:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:07.735 23:48:55 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:07.735 23:48:55 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:07.735 23:48:55 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:07.735 23:48:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:07.735 "name": "raid_bdev1", 00:15:07.735 "uuid": "ff5af955-a414-4ba0-8235-7545475fb4f4", 00:15:07.735 "strip_size_kb": 0, 00:15:07.735 "state": "online", 00:15:07.735 "raid_level": "raid1", 00:15:07.735 "superblock": true, 00:15:07.735 "num_base_bdevs": 2, 00:15:07.735 "num_base_bdevs_discovered": 1, 00:15:07.735 "num_base_bdevs_operational": 1, 00:15:07.735 "base_bdevs_list": [ 00:15:07.735 { 00:15:07.735 "name": null, 00:15:07.735 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:07.735 "is_configured": false, 00:15:07.735 "data_offset": 0, 00:15:07.735 "data_size": 7936 00:15:07.735 }, 00:15:07.735 { 00:15:07.735 "name": "BaseBdev2", 00:15:07.735 "uuid": "6f077572-90b7-568a-919c-93be71941c99", 00:15:07.735 "is_configured": true, 00:15:07.735 "data_offset": 256, 00:15:07.735 "data_size": 7936 00:15:07.735 } 00:15:07.735 ] 00:15:07.735 }' 00:15:07.735 23:48:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:07.735 23:48:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:15:07.735 23:48:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:07.735 23:48:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:15:07.995 23:48:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@784 -- # killprocess 96521 00:15:07.995 23:48:55 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@954 -- # '[' -z 96521 ']' 00:15:07.995 23:48:55 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@958 -- # kill -0 96521 00:15:07.995 23:48:55 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@959 -- # uname 00:15:07.995 23:48:55 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:15:07.995 23:48:55 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 96521 00:15:07.995 23:48:55 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:15:07.995 killing process with pid 96521 00:15:07.995 Received shutdown signal, test time was about 60.000000 seconds 00:15:07.995 00:15:07.995 Latency(us) 00:15:07.995 [2024-11-26T23:48:56.127Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:15:07.995 [2024-11-26T23:48:56.127Z] =================================================================================================================== 00:15:07.995 [2024-11-26T23:48:56.127Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:15:07.995 23:48:55 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:15:07.995 23:48:55 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@972 -- # echo 'killing process with pid 96521' 00:15:07.995 23:48:55 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@973 -- # kill 96521 00:15:07.995 [2024-11-26 23:48:55.908221] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:15:07.995 [2024-11-26 23:48:55.908329] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:07.995 [2024-11-26 23:48:55.908391] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:07.995 [2024-11-26 23:48:55.908401] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state offline 00:15:07.995 23:48:55 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@978 -- # wait 96521 00:15:07.995 [2024-11-26 23:48:55.938636] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:15:08.256 23:48:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@786 -- # return 0 00:15:08.256 00:15:08.256 real 0m18.018s 00:15:08.256 user 0m23.781s 00:15:08.256 sys 0m2.546s 00:15:08.256 23:48:56 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@1130 -- # xtrace_disable 00:15:08.256 23:48:56 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:08.256 ************************************ 00:15:08.256 END TEST raid_rebuild_test_sb_4k 00:15:08.256 ************************************ 00:15:08.256 23:48:56 bdev_raid -- bdev/bdev_raid.sh@1003 -- # base_malloc_params='-m 32' 00:15:08.256 23:48:56 bdev_raid -- bdev/bdev_raid.sh@1004 -- # run_test raid_state_function_test_sb_md_separate raid_state_function_test raid1 2 true 00:15:08.256 23:48:56 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:15:08.256 23:48:56 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:15:08.256 23:48:56 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:15:08.256 ************************************ 00:15:08.256 START TEST raid_state_function_test_sb_md_separate 00:15:08.256 ************************************ 00:15:08.256 23:48:56 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@1129 -- # raid_state_function_test raid1 2 true 00:15:08.256 23:48:56 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@205 -- # local raid_level=raid1 00:15:08.256 23:48:56 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=2 00:15:08.256 23:48:56 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:15:08.256 23:48:56 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:15:08.256 23:48:56 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:15:08.256 23:48:56 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:15:08.256 23:48:56 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:15:08.256 23:48:56 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:15:08.256 23:48:56 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:15:08.256 23:48:56 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:15:08.256 23:48:56 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:15:08.256 23:48:56 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:15:08.256 23:48:56 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:15:08.256 23:48:56 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:15:08.256 23:48:56 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:15:08.256 23:48:56 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@211 -- # local strip_size 00:15:08.256 23:48:56 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:15:08.256 23:48:56 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:15:08.256 23:48:56 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@215 -- # '[' raid1 '!=' raid1 ']' 00:15:08.256 23:48:56 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@219 -- # strip_size=0 00:15:08.256 23:48:56 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:15:08.256 23:48:56 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:15:08.256 Process raid pid: 97195 00:15:08.256 23:48:56 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@229 -- # raid_pid=97195 00:15:08.256 23:48:56 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:15:08.256 23:48:56 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 97195' 00:15:08.256 23:48:56 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@231 -- # waitforlisten 97195 00:15:08.256 23:48:56 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@835 -- # '[' -z 97195 ']' 00:15:08.256 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:15:08.256 23:48:56 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:15:08.256 23:48:56 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@840 -- # local max_retries=100 00:15:08.256 23:48:56 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:15:08.256 23:48:56 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@844 -- # xtrace_disable 00:15:08.256 23:48:56 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:08.256 [2024-11-26 23:48:56.307235] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:15:08.256 [2024-11-26 23:48:56.307397] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:15:08.520 [2024-11-26 23:48:56.464484] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:15:08.520 [2024-11-26 23:48:56.489354] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:15:08.520 [2024-11-26 23:48:56.531503] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:15:08.520 [2024-11-26 23:48:56.531539] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:15:09.091 23:48:57 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:15:09.091 23:48:57 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@868 -- # return 0 00:15:09.091 23:48:57 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:15:09.091 23:48:57 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:09.091 23:48:57 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:09.091 [2024-11-26 23:48:57.134066] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:15:09.091 [2024-11-26 23:48:57.134118] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:15:09.091 [2024-11-26 23:48:57.134134] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:15:09.091 [2024-11-26 23:48:57.134145] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:15:09.091 23:48:57 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:09.091 23:48:57 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:15:09.091 23:48:57 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:09.091 23:48:57 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:09.091 23:48:57 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:09.091 23:48:57 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:09.091 23:48:57 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:09.091 23:48:57 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:09.091 23:48:57 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:09.091 23:48:57 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:09.091 23:48:57 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:09.091 23:48:57 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:09.091 23:48:57 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:09.091 23:48:57 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:09.091 23:48:57 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:09.091 23:48:57 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:09.091 23:48:57 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:09.091 "name": "Existed_Raid", 00:15:09.091 "uuid": "9a0ee0ae-5fbc-4634-945b-687e5b2517bf", 00:15:09.091 "strip_size_kb": 0, 00:15:09.091 "state": "configuring", 00:15:09.091 "raid_level": "raid1", 00:15:09.091 "superblock": true, 00:15:09.091 "num_base_bdevs": 2, 00:15:09.091 "num_base_bdevs_discovered": 0, 00:15:09.091 "num_base_bdevs_operational": 2, 00:15:09.091 "base_bdevs_list": [ 00:15:09.091 { 00:15:09.091 "name": "BaseBdev1", 00:15:09.091 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:09.091 "is_configured": false, 00:15:09.091 "data_offset": 0, 00:15:09.091 "data_size": 0 00:15:09.091 }, 00:15:09.091 { 00:15:09.091 "name": "BaseBdev2", 00:15:09.091 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:09.091 "is_configured": false, 00:15:09.091 "data_offset": 0, 00:15:09.091 "data_size": 0 00:15:09.091 } 00:15:09.091 ] 00:15:09.091 }' 00:15:09.091 23:48:57 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:09.091 23:48:57 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:09.660 23:48:57 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:15:09.660 23:48:57 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:09.660 23:48:57 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:09.660 [2024-11-26 23:48:57.541280] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:15:09.660 [2024-11-26 23:48:57.541388] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:15:09.660 23:48:57 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:09.660 23:48:57 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:15:09.660 23:48:57 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:09.660 23:48:57 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:09.660 [2024-11-26 23:48:57.553289] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:15:09.660 [2024-11-26 23:48:57.553397] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:15:09.660 [2024-11-26 23:48:57.553427] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:15:09.660 [2024-11-26 23:48:57.553461] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:15:09.660 23:48:57 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:09.660 23:48:57 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -b BaseBdev1 00:15:09.660 23:48:57 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:09.660 23:48:57 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:09.660 [2024-11-26 23:48:57.574446] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:15:09.660 BaseBdev1 00:15:09.660 23:48:57 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:09.660 23:48:57 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:15:09.660 23:48:57 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:15:09.661 23:48:57 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:15:09.661 23:48:57 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@905 -- # local i 00:15:09.661 23:48:57 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:15:09.661 23:48:57 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:15:09.661 23:48:57 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:15:09.661 23:48:57 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:09.661 23:48:57 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:09.661 23:48:57 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:09.661 23:48:57 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:15:09.661 23:48:57 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:09.661 23:48:57 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:09.661 [ 00:15:09.661 { 00:15:09.661 "name": "BaseBdev1", 00:15:09.661 "aliases": [ 00:15:09.661 "d985e8e0-b088-4b6c-8230-70c1447fcab4" 00:15:09.661 ], 00:15:09.661 "product_name": "Malloc disk", 00:15:09.661 "block_size": 4096, 00:15:09.661 "num_blocks": 8192, 00:15:09.661 "uuid": "d985e8e0-b088-4b6c-8230-70c1447fcab4", 00:15:09.661 "md_size": 32, 00:15:09.661 "md_interleave": false, 00:15:09.661 "dif_type": 0, 00:15:09.661 "assigned_rate_limits": { 00:15:09.661 "rw_ios_per_sec": 0, 00:15:09.661 "rw_mbytes_per_sec": 0, 00:15:09.661 "r_mbytes_per_sec": 0, 00:15:09.661 "w_mbytes_per_sec": 0 00:15:09.661 }, 00:15:09.661 "claimed": true, 00:15:09.661 "claim_type": "exclusive_write", 00:15:09.661 "zoned": false, 00:15:09.661 "supported_io_types": { 00:15:09.661 "read": true, 00:15:09.661 "write": true, 00:15:09.661 "unmap": true, 00:15:09.661 "flush": true, 00:15:09.661 "reset": true, 00:15:09.661 "nvme_admin": false, 00:15:09.661 "nvme_io": false, 00:15:09.661 "nvme_io_md": false, 00:15:09.661 "write_zeroes": true, 00:15:09.661 "zcopy": true, 00:15:09.661 "get_zone_info": false, 00:15:09.661 "zone_management": false, 00:15:09.661 "zone_append": false, 00:15:09.661 "compare": false, 00:15:09.661 "compare_and_write": false, 00:15:09.661 "abort": true, 00:15:09.661 "seek_hole": false, 00:15:09.661 "seek_data": false, 00:15:09.661 "copy": true, 00:15:09.661 "nvme_iov_md": false 00:15:09.661 }, 00:15:09.661 "memory_domains": [ 00:15:09.661 { 00:15:09.661 "dma_device_id": "system", 00:15:09.661 "dma_device_type": 1 00:15:09.661 }, 00:15:09.661 { 00:15:09.661 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:09.661 "dma_device_type": 2 00:15:09.661 } 00:15:09.661 ], 00:15:09.661 "driver_specific": {} 00:15:09.661 } 00:15:09.661 ] 00:15:09.661 23:48:57 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:09.661 23:48:57 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@911 -- # return 0 00:15:09.661 23:48:57 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:15:09.661 23:48:57 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:09.661 23:48:57 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:09.661 23:48:57 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:09.661 23:48:57 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:09.661 23:48:57 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:09.661 23:48:57 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:09.661 23:48:57 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:09.661 23:48:57 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:09.661 23:48:57 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:09.661 23:48:57 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:09.661 23:48:57 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:09.661 23:48:57 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:09.661 23:48:57 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:09.661 23:48:57 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:09.661 23:48:57 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:09.661 "name": "Existed_Raid", 00:15:09.661 "uuid": "1daea59e-0fa4-4a3b-92f7-d65071dd6f58", 00:15:09.661 "strip_size_kb": 0, 00:15:09.661 "state": "configuring", 00:15:09.661 "raid_level": "raid1", 00:15:09.661 "superblock": true, 00:15:09.661 "num_base_bdevs": 2, 00:15:09.661 "num_base_bdevs_discovered": 1, 00:15:09.661 "num_base_bdevs_operational": 2, 00:15:09.661 "base_bdevs_list": [ 00:15:09.661 { 00:15:09.661 "name": "BaseBdev1", 00:15:09.661 "uuid": "d985e8e0-b088-4b6c-8230-70c1447fcab4", 00:15:09.661 "is_configured": true, 00:15:09.661 "data_offset": 256, 00:15:09.661 "data_size": 7936 00:15:09.661 }, 00:15:09.661 { 00:15:09.661 "name": "BaseBdev2", 00:15:09.661 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:09.661 "is_configured": false, 00:15:09.661 "data_offset": 0, 00:15:09.661 "data_size": 0 00:15:09.661 } 00:15:09.661 ] 00:15:09.661 }' 00:15:09.661 23:48:57 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:09.661 23:48:57 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:10.232 23:48:58 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:15:10.232 23:48:58 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:10.232 23:48:58 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:10.232 [2024-11-26 23:48:58.081616] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:15:10.232 [2024-11-26 23:48:58.081704] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:15:10.232 23:48:58 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:10.232 23:48:58 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:15:10.232 23:48:58 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:10.232 23:48:58 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:10.232 [2024-11-26 23:48:58.093645] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:15:10.232 [2024-11-26 23:48:58.095538] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:15:10.232 [2024-11-26 23:48:58.095579] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:15:10.232 23:48:58 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:10.232 23:48:58 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:15:10.232 23:48:58 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:15:10.232 23:48:58 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:15:10.232 23:48:58 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:10.232 23:48:58 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:10.232 23:48:58 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:10.232 23:48:58 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:10.232 23:48:58 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:10.232 23:48:58 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:10.232 23:48:58 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:10.232 23:48:58 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:10.232 23:48:58 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:10.232 23:48:58 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:10.232 23:48:58 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:10.232 23:48:58 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:10.232 23:48:58 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:10.232 23:48:58 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:10.232 23:48:58 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:10.232 "name": "Existed_Raid", 00:15:10.232 "uuid": "12daefc0-ced2-47cc-9cf7-ccf1e7a77d87", 00:15:10.232 "strip_size_kb": 0, 00:15:10.232 "state": "configuring", 00:15:10.232 "raid_level": "raid1", 00:15:10.232 "superblock": true, 00:15:10.232 "num_base_bdevs": 2, 00:15:10.232 "num_base_bdevs_discovered": 1, 00:15:10.232 "num_base_bdevs_operational": 2, 00:15:10.232 "base_bdevs_list": [ 00:15:10.232 { 00:15:10.232 "name": "BaseBdev1", 00:15:10.232 "uuid": "d985e8e0-b088-4b6c-8230-70c1447fcab4", 00:15:10.232 "is_configured": true, 00:15:10.232 "data_offset": 256, 00:15:10.232 "data_size": 7936 00:15:10.232 }, 00:15:10.232 { 00:15:10.232 "name": "BaseBdev2", 00:15:10.232 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:10.232 "is_configured": false, 00:15:10.232 "data_offset": 0, 00:15:10.232 "data_size": 0 00:15:10.232 } 00:15:10.232 ] 00:15:10.232 }' 00:15:10.232 23:48:58 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:10.232 23:48:58 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:10.505 23:48:58 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -b BaseBdev2 00:15:10.505 23:48:58 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:10.506 23:48:58 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:10.506 [2024-11-26 23:48:58.532316] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:15:10.506 [2024-11-26 23:48:58.532602] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:15:10.506 [2024-11-26 23:48:58.532649] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:15:10.506 [2024-11-26 23:48:58.532765] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:15:10.506 [2024-11-26 23:48:58.532896] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:15:10.506 [2024-11-26 23:48:58.532942] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:15:10.506 [2024-11-26 23:48:58.533041] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:10.506 BaseBdev2 00:15:10.506 23:48:58 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:10.506 23:48:58 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:15:10.506 23:48:58 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:15:10.506 23:48:58 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:15:10.506 23:48:58 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@905 -- # local i 00:15:10.506 23:48:58 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:15:10.506 23:48:58 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:15:10.506 23:48:58 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:15:10.506 23:48:58 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:10.507 23:48:58 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:10.507 23:48:58 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:10.507 23:48:58 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:15:10.507 23:48:58 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:10.507 23:48:58 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:10.507 [ 00:15:10.507 { 00:15:10.507 "name": "BaseBdev2", 00:15:10.507 "aliases": [ 00:15:10.507 "7f95ca5c-8914-454c-b0ba-27eeafae640e" 00:15:10.507 ], 00:15:10.507 "product_name": "Malloc disk", 00:15:10.507 "block_size": 4096, 00:15:10.507 "num_blocks": 8192, 00:15:10.507 "uuid": "7f95ca5c-8914-454c-b0ba-27eeafae640e", 00:15:10.507 "md_size": 32, 00:15:10.507 "md_interleave": false, 00:15:10.507 "dif_type": 0, 00:15:10.507 "assigned_rate_limits": { 00:15:10.507 "rw_ios_per_sec": 0, 00:15:10.507 "rw_mbytes_per_sec": 0, 00:15:10.507 "r_mbytes_per_sec": 0, 00:15:10.507 "w_mbytes_per_sec": 0 00:15:10.507 }, 00:15:10.507 "claimed": true, 00:15:10.507 "claim_type": "exclusive_write", 00:15:10.507 "zoned": false, 00:15:10.507 "supported_io_types": { 00:15:10.507 "read": true, 00:15:10.507 "write": true, 00:15:10.507 "unmap": true, 00:15:10.507 "flush": true, 00:15:10.507 "reset": true, 00:15:10.507 "nvme_admin": false, 00:15:10.507 "nvme_io": false, 00:15:10.507 "nvme_io_md": false, 00:15:10.507 "write_zeroes": true, 00:15:10.507 "zcopy": true, 00:15:10.507 "get_zone_info": false, 00:15:10.507 "zone_management": false, 00:15:10.507 "zone_append": false, 00:15:10.507 "compare": false, 00:15:10.507 "compare_and_write": false, 00:15:10.507 "abort": true, 00:15:10.507 "seek_hole": false, 00:15:10.508 "seek_data": false, 00:15:10.508 "copy": true, 00:15:10.508 "nvme_iov_md": false 00:15:10.508 }, 00:15:10.508 "memory_domains": [ 00:15:10.508 { 00:15:10.508 "dma_device_id": "system", 00:15:10.508 "dma_device_type": 1 00:15:10.508 }, 00:15:10.508 { 00:15:10.508 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:10.508 "dma_device_type": 2 00:15:10.508 } 00:15:10.508 ], 00:15:10.508 "driver_specific": {} 00:15:10.508 } 00:15:10.508 ] 00:15:10.508 23:48:58 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:10.508 23:48:58 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@911 -- # return 0 00:15:10.508 23:48:58 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:15:10.508 23:48:58 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:15:10.508 23:48:58 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid1 0 2 00:15:10.508 23:48:58 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:10.508 23:48:58 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:10.508 23:48:58 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:10.508 23:48:58 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:10.508 23:48:58 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:10.508 23:48:58 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:10.508 23:48:58 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:10.508 23:48:58 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:10.508 23:48:58 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:10.508 23:48:58 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:10.508 23:48:58 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:10.508 23:48:58 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:10.508 23:48:58 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:10.508 23:48:58 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:10.508 23:48:58 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:10.509 "name": "Existed_Raid", 00:15:10.509 "uuid": "12daefc0-ced2-47cc-9cf7-ccf1e7a77d87", 00:15:10.509 "strip_size_kb": 0, 00:15:10.509 "state": "online", 00:15:10.509 "raid_level": "raid1", 00:15:10.509 "superblock": true, 00:15:10.509 "num_base_bdevs": 2, 00:15:10.509 "num_base_bdevs_discovered": 2, 00:15:10.509 "num_base_bdevs_operational": 2, 00:15:10.509 "base_bdevs_list": [ 00:15:10.509 { 00:15:10.509 "name": "BaseBdev1", 00:15:10.509 "uuid": "d985e8e0-b088-4b6c-8230-70c1447fcab4", 00:15:10.509 "is_configured": true, 00:15:10.509 "data_offset": 256, 00:15:10.509 "data_size": 7936 00:15:10.509 }, 00:15:10.509 { 00:15:10.509 "name": "BaseBdev2", 00:15:10.509 "uuid": "7f95ca5c-8914-454c-b0ba-27eeafae640e", 00:15:10.509 "is_configured": true, 00:15:10.509 "data_offset": 256, 00:15:10.509 "data_size": 7936 00:15:10.509 } 00:15:10.509 ] 00:15:10.509 }' 00:15:10.509 23:48:58 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:10.509 23:48:58 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:11.087 23:48:58 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:15:11.087 23:48:58 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:15:11.087 23:48:58 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:15:11.087 23:48:58 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:15:11.087 23:48:58 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@184 -- # local name 00:15:11.087 23:48:58 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:15:11.087 23:48:58 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:15:11.087 23:48:58 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:15:11.087 23:48:58 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:11.087 23:48:58 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:11.087 [2024-11-26 23:48:58.995808] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:11.087 23:48:59 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:11.087 23:48:59 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:15:11.087 "name": "Existed_Raid", 00:15:11.087 "aliases": [ 00:15:11.087 "12daefc0-ced2-47cc-9cf7-ccf1e7a77d87" 00:15:11.087 ], 00:15:11.087 "product_name": "Raid Volume", 00:15:11.087 "block_size": 4096, 00:15:11.087 "num_blocks": 7936, 00:15:11.087 "uuid": "12daefc0-ced2-47cc-9cf7-ccf1e7a77d87", 00:15:11.087 "md_size": 32, 00:15:11.087 "md_interleave": false, 00:15:11.087 "dif_type": 0, 00:15:11.087 "assigned_rate_limits": { 00:15:11.087 "rw_ios_per_sec": 0, 00:15:11.087 "rw_mbytes_per_sec": 0, 00:15:11.087 "r_mbytes_per_sec": 0, 00:15:11.087 "w_mbytes_per_sec": 0 00:15:11.087 }, 00:15:11.087 "claimed": false, 00:15:11.087 "zoned": false, 00:15:11.087 "supported_io_types": { 00:15:11.087 "read": true, 00:15:11.087 "write": true, 00:15:11.087 "unmap": false, 00:15:11.087 "flush": false, 00:15:11.087 "reset": true, 00:15:11.087 "nvme_admin": false, 00:15:11.087 "nvme_io": false, 00:15:11.087 "nvme_io_md": false, 00:15:11.087 "write_zeroes": true, 00:15:11.087 "zcopy": false, 00:15:11.087 "get_zone_info": false, 00:15:11.087 "zone_management": false, 00:15:11.087 "zone_append": false, 00:15:11.087 "compare": false, 00:15:11.087 "compare_and_write": false, 00:15:11.087 "abort": false, 00:15:11.087 "seek_hole": false, 00:15:11.087 "seek_data": false, 00:15:11.087 "copy": false, 00:15:11.087 "nvme_iov_md": false 00:15:11.087 }, 00:15:11.087 "memory_domains": [ 00:15:11.087 { 00:15:11.087 "dma_device_id": "system", 00:15:11.087 "dma_device_type": 1 00:15:11.087 }, 00:15:11.087 { 00:15:11.087 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:11.087 "dma_device_type": 2 00:15:11.087 }, 00:15:11.087 { 00:15:11.087 "dma_device_id": "system", 00:15:11.087 "dma_device_type": 1 00:15:11.087 }, 00:15:11.087 { 00:15:11.087 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:11.087 "dma_device_type": 2 00:15:11.087 } 00:15:11.087 ], 00:15:11.087 "driver_specific": { 00:15:11.087 "raid": { 00:15:11.087 "uuid": "12daefc0-ced2-47cc-9cf7-ccf1e7a77d87", 00:15:11.087 "strip_size_kb": 0, 00:15:11.087 "state": "online", 00:15:11.087 "raid_level": "raid1", 00:15:11.087 "superblock": true, 00:15:11.087 "num_base_bdevs": 2, 00:15:11.087 "num_base_bdevs_discovered": 2, 00:15:11.087 "num_base_bdevs_operational": 2, 00:15:11.087 "base_bdevs_list": [ 00:15:11.087 { 00:15:11.087 "name": "BaseBdev1", 00:15:11.087 "uuid": "d985e8e0-b088-4b6c-8230-70c1447fcab4", 00:15:11.087 "is_configured": true, 00:15:11.087 "data_offset": 256, 00:15:11.087 "data_size": 7936 00:15:11.087 }, 00:15:11.087 { 00:15:11.087 "name": "BaseBdev2", 00:15:11.087 "uuid": "7f95ca5c-8914-454c-b0ba-27eeafae640e", 00:15:11.087 "is_configured": true, 00:15:11.087 "data_offset": 256, 00:15:11.087 "data_size": 7936 00:15:11.087 } 00:15:11.087 ] 00:15:11.087 } 00:15:11.087 } 00:15:11.087 }' 00:15:11.087 23:48:59 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:15:11.087 23:48:59 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:15:11.087 BaseBdev2' 00:15:11.087 23:48:59 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:11.087 23:48:59 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='4096 32 false 0' 00:15:11.087 23:48:59 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:11.087 23:48:59 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:15:11.087 23:48:59 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:11.088 23:48:59 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:11.088 23:48:59 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:11.088 23:48:59 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:11.088 23:48:59 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 32 false 0' 00:15:11.088 23:48:59 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@193 -- # [[ 4096 32 false 0 == \4\0\9\6\ \3\2\ \f\a\l\s\e\ \0 ]] 00:15:11.088 23:48:59 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:11.088 23:48:59 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:11.088 23:48:59 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:15:11.088 23:48:59 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:11.088 23:48:59 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:11.088 23:48:59 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:11.088 23:48:59 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 32 false 0' 00:15:11.088 23:48:59 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@193 -- # [[ 4096 32 false 0 == \4\0\9\6\ \3\2\ \f\a\l\s\e\ \0 ]] 00:15:11.088 23:48:59 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:15:11.088 23:48:59 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:11.088 23:48:59 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:11.088 [2024-11-26 23:48:59.199249] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:15:11.088 23:48:59 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:11.088 23:48:59 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@260 -- # local expected_state 00:15:11.088 23:48:59 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@261 -- # has_redundancy raid1 00:15:11.088 23:48:59 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@198 -- # case $1 in 00:15:11.088 23:48:59 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@199 -- # return 0 00:15:11.088 23:48:59 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:15:11.088 23:48:59 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid1 0 1 00:15:11.088 23:48:59 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:11.088 23:48:59 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:11.088 23:48:59 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:11.088 23:48:59 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:11.088 23:48:59 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:15:11.088 23:48:59 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:11.088 23:48:59 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:11.088 23:48:59 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:11.346 23:48:59 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:11.346 23:48:59 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:11.347 23:48:59 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:11.347 23:48:59 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:11.347 23:48:59 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:11.347 23:48:59 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:11.347 23:48:59 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:11.347 "name": "Existed_Raid", 00:15:11.347 "uuid": "12daefc0-ced2-47cc-9cf7-ccf1e7a77d87", 00:15:11.347 "strip_size_kb": 0, 00:15:11.347 "state": "online", 00:15:11.347 "raid_level": "raid1", 00:15:11.347 "superblock": true, 00:15:11.347 "num_base_bdevs": 2, 00:15:11.347 "num_base_bdevs_discovered": 1, 00:15:11.347 "num_base_bdevs_operational": 1, 00:15:11.347 "base_bdevs_list": [ 00:15:11.347 { 00:15:11.347 "name": null, 00:15:11.347 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:11.347 "is_configured": false, 00:15:11.347 "data_offset": 0, 00:15:11.347 "data_size": 7936 00:15:11.347 }, 00:15:11.347 { 00:15:11.347 "name": "BaseBdev2", 00:15:11.347 "uuid": "7f95ca5c-8914-454c-b0ba-27eeafae640e", 00:15:11.347 "is_configured": true, 00:15:11.347 "data_offset": 256, 00:15:11.347 "data_size": 7936 00:15:11.347 } 00:15:11.347 ] 00:15:11.347 }' 00:15:11.347 23:48:59 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:11.347 23:48:59 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:11.606 23:48:59 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:15:11.606 23:48:59 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:15:11.606 23:48:59 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:15:11.606 23:48:59 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:11.606 23:48:59 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:11.606 23:48:59 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:11.866 23:48:59 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:11.866 23:48:59 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:15:11.866 23:48:59 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:15:11.866 23:48:59 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:15:11.866 23:48:59 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:11.866 23:48:59 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:11.866 [2024-11-26 23:48:59.754299] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:15:11.866 [2024-11-26 23:48:59.754464] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:15:11.866 [2024-11-26 23:48:59.766418] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:11.866 [2024-11-26 23:48:59.766532] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:11.866 [2024-11-26 23:48:59.766575] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:15:11.866 23:48:59 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:11.866 23:48:59 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:15:11.866 23:48:59 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:15:11.866 23:48:59 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:11.866 23:48:59 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:11.866 23:48:59 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:11.866 23:48:59 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:15:11.866 23:48:59 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:11.866 23:48:59 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:15:11.866 23:48:59 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:15:11.867 23:48:59 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@284 -- # '[' 2 -gt 2 ']' 00:15:11.867 23:48:59 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@326 -- # killprocess 97195 00:15:11.867 23:48:59 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@954 -- # '[' -z 97195 ']' 00:15:11.867 23:48:59 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@958 -- # kill -0 97195 00:15:11.867 23:48:59 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@959 -- # uname 00:15:11.867 23:48:59 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:15:11.867 23:48:59 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 97195 00:15:11.867 23:48:59 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:15:11.867 killing process with pid 97195 00:15:11.867 23:48:59 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:15:11.867 23:48:59 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@972 -- # echo 'killing process with pid 97195' 00:15:11.867 23:48:59 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@973 -- # kill 97195 00:15:11.867 [2024-11-26 23:48:59.847523] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:15:11.867 23:48:59 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@978 -- # wait 97195 00:15:11.867 [2024-11-26 23:48:59.848496] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:15:12.127 23:49:00 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@328 -- # return 0 00:15:12.127 00:15:12.127 real 0m3.847s 00:15:12.127 user 0m6.056s 00:15:12.127 sys 0m0.815s 00:15:12.127 23:49:00 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@1130 -- # xtrace_disable 00:15:12.127 ************************************ 00:15:12.127 END TEST raid_state_function_test_sb_md_separate 00:15:12.127 ************************************ 00:15:12.127 23:49:00 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:12.127 23:49:00 bdev_raid -- bdev/bdev_raid.sh@1005 -- # run_test raid_superblock_test_md_separate raid_superblock_test raid1 2 00:15:12.127 23:49:00 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:15:12.127 23:49:00 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:15:12.127 23:49:00 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:15:12.127 ************************************ 00:15:12.127 START TEST raid_superblock_test_md_separate 00:15:12.127 ************************************ 00:15:12.127 23:49:00 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@1129 -- # raid_superblock_test raid1 2 00:15:12.127 23:49:00 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@393 -- # local raid_level=raid1 00:15:12.127 23:49:00 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=2 00:15:12.127 23:49:00 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:15:12.127 23:49:00 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:15:12.127 23:49:00 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:15:12.127 23:49:00 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:15:12.127 23:49:00 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:15:12.127 23:49:00 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:15:12.127 23:49:00 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:15:12.127 23:49:00 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@399 -- # local strip_size 00:15:12.127 23:49:00 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:15:12.127 23:49:00 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:15:12.127 23:49:00 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:15:12.127 23:49:00 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@404 -- # '[' raid1 '!=' raid1 ']' 00:15:12.127 23:49:00 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@408 -- # strip_size=0 00:15:12.127 23:49:00 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@412 -- # raid_pid=97430 00:15:12.127 23:49:00 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:15:12.127 23:49:00 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@413 -- # waitforlisten 97430 00:15:12.127 23:49:00 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@835 -- # '[' -z 97430 ']' 00:15:12.127 23:49:00 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:15:12.127 23:49:00 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@840 -- # local max_retries=100 00:15:12.128 23:49:00 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:15:12.128 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:15:12.128 23:49:00 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@844 -- # xtrace_disable 00:15:12.128 23:49:00 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:12.128 [2024-11-26 23:49:00.222565] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:15:12.128 [2024-11-26 23:49:00.223128] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid97430 ] 00:15:12.387 [2024-11-26 23:49:00.377567] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:15:12.388 [2024-11-26 23:49:00.402849] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:15:12.388 [2024-11-26 23:49:00.445297] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:15:12.388 [2024-11-26 23:49:00.445460] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:15:12.957 23:49:01 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:15:12.957 23:49:01 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@868 -- # return 0 00:15:12.957 23:49:01 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:15:12.958 23:49:01 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:15:12.958 23:49:01 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:15:12.958 23:49:01 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:15:12.958 23:49:01 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:15:12.958 23:49:01 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:15:12.958 23:49:01 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:15:12.958 23:49:01 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:15:12.958 23:49:01 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -b malloc1 00:15:12.958 23:49:01 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:12.958 23:49:01 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:12.958 malloc1 00:15:12.958 23:49:01 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:12.958 23:49:01 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:15:12.958 23:49:01 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:12.958 23:49:01 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:12.958 [2024-11-26 23:49:01.056810] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:15:12.958 [2024-11-26 23:49:01.056890] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:12.958 [2024-11-26 23:49:01.056914] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:15:12.958 [2024-11-26 23:49:01.056925] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:12.958 [2024-11-26 23:49:01.058815] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:12.958 [2024-11-26 23:49:01.058857] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:15:12.958 pt1 00:15:12.958 23:49:01 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:12.958 23:49:01 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:15:12.958 23:49:01 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:15:12.958 23:49:01 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:15:12.958 23:49:01 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:15:12.958 23:49:01 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:15:12.958 23:49:01 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:15:12.958 23:49:01 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:15:12.958 23:49:01 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:15:12.958 23:49:01 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -b malloc2 00:15:12.958 23:49:01 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:12.958 23:49:01 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:12.958 malloc2 00:15:12.958 23:49:01 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:12.958 23:49:01 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:15:12.958 23:49:01 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:12.958 23:49:01 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:12.958 [2024-11-26 23:49:01.085746] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:15:12.958 [2024-11-26 23:49:01.085876] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:12.958 [2024-11-26 23:49:01.085907] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:15:12.958 [2024-11-26 23:49:01.085934] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:13.217 [2024-11-26 23:49:01.087772] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:13.217 [2024-11-26 23:49:01.087848] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:15:13.217 pt2 00:15:13.217 23:49:01 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:13.217 23:49:01 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:15:13.217 23:49:01 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:15:13.217 23:49:01 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''pt1 pt2'\''' -n raid_bdev1 -s 00:15:13.217 23:49:01 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:13.217 23:49:01 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:13.217 [2024-11-26 23:49:01.097766] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:15:13.217 [2024-11-26 23:49:01.099588] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:15:13.217 [2024-11-26 23:49:01.099778] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:15:13.217 [2024-11-26 23:49:01.099829] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:15:13.217 [2024-11-26 23:49:01.099926] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:15:13.217 [2024-11-26 23:49:01.100064] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:15:13.217 [2024-11-26 23:49:01.100104] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:15:13.217 [2024-11-26 23:49:01.100230] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:13.217 23:49:01 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:13.217 23:49:01 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:15:13.217 23:49:01 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:13.217 23:49:01 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:13.217 23:49:01 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:13.217 23:49:01 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:13.217 23:49:01 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:13.217 23:49:01 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:13.217 23:49:01 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:13.217 23:49:01 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:13.217 23:49:01 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:13.217 23:49:01 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:13.217 23:49:01 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:13.217 23:49:01 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:13.217 23:49:01 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:13.217 23:49:01 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:13.217 23:49:01 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:13.217 "name": "raid_bdev1", 00:15:13.217 "uuid": "8df82abf-2f78-4e89-91f0-502c5a51c3ea", 00:15:13.217 "strip_size_kb": 0, 00:15:13.217 "state": "online", 00:15:13.217 "raid_level": "raid1", 00:15:13.217 "superblock": true, 00:15:13.217 "num_base_bdevs": 2, 00:15:13.217 "num_base_bdevs_discovered": 2, 00:15:13.217 "num_base_bdevs_operational": 2, 00:15:13.217 "base_bdevs_list": [ 00:15:13.217 { 00:15:13.217 "name": "pt1", 00:15:13.217 "uuid": "00000000-0000-0000-0000-000000000001", 00:15:13.217 "is_configured": true, 00:15:13.217 "data_offset": 256, 00:15:13.217 "data_size": 7936 00:15:13.217 }, 00:15:13.217 { 00:15:13.217 "name": "pt2", 00:15:13.217 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:13.217 "is_configured": true, 00:15:13.217 "data_offset": 256, 00:15:13.217 "data_size": 7936 00:15:13.217 } 00:15:13.217 ] 00:15:13.217 }' 00:15:13.217 23:49:01 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:13.217 23:49:01 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:13.495 23:49:01 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:15:13.495 23:49:01 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:15:13.495 23:49:01 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:15:13.495 23:49:01 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:15:13.495 23:49:01 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@184 -- # local name 00:15:13.495 23:49:01 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:15:13.495 23:49:01 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:15:13.495 23:49:01 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:15:13.495 23:49:01 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:13.495 23:49:01 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:13.495 [2024-11-26 23:49:01.569206] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:13.495 23:49:01 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:13.495 23:49:01 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:15:13.495 "name": "raid_bdev1", 00:15:13.495 "aliases": [ 00:15:13.495 "8df82abf-2f78-4e89-91f0-502c5a51c3ea" 00:15:13.495 ], 00:15:13.495 "product_name": "Raid Volume", 00:15:13.495 "block_size": 4096, 00:15:13.495 "num_blocks": 7936, 00:15:13.495 "uuid": "8df82abf-2f78-4e89-91f0-502c5a51c3ea", 00:15:13.495 "md_size": 32, 00:15:13.495 "md_interleave": false, 00:15:13.495 "dif_type": 0, 00:15:13.495 "assigned_rate_limits": { 00:15:13.495 "rw_ios_per_sec": 0, 00:15:13.495 "rw_mbytes_per_sec": 0, 00:15:13.495 "r_mbytes_per_sec": 0, 00:15:13.495 "w_mbytes_per_sec": 0 00:15:13.495 }, 00:15:13.495 "claimed": false, 00:15:13.495 "zoned": false, 00:15:13.495 "supported_io_types": { 00:15:13.495 "read": true, 00:15:13.495 "write": true, 00:15:13.495 "unmap": false, 00:15:13.495 "flush": false, 00:15:13.495 "reset": true, 00:15:13.495 "nvme_admin": false, 00:15:13.495 "nvme_io": false, 00:15:13.495 "nvme_io_md": false, 00:15:13.495 "write_zeroes": true, 00:15:13.495 "zcopy": false, 00:15:13.495 "get_zone_info": false, 00:15:13.495 "zone_management": false, 00:15:13.495 "zone_append": false, 00:15:13.495 "compare": false, 00:15:13.495 "compare_and_write": false, 00:15:13.495 "abort": false, 00:15:13.495 "seek_hole": false, 00:15:13.495 "seek_data": false, 00:15:13.495 "copy": false, 00:15:13.495 "nvme_iov_md": false 00:15:13.495 }, 00:15:13.495 "memory_domains": [ 00:15:13.495 { 00:15:13.495 "dma_device_id": "system", 00:15:13.495 "dma_device_type": 1 00:15:13.495 }, 00:15:13.495 { 00:15:13.495 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:13.495 "dma_device_type": 2 00:15:13.495 }, 00:15:13.495 { 00:15:13.495 "dma_device_id": "system", 00:15:13.495 "dma_device_type": 1 00:15:13.495 }, 00:15:13.495 { 00:15:13.495 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:13.495 "dma_device_type": 2 00:15:13.495 } 00:15:13.495 ], 00:15:13.495 "driver_specific": { 00:15:13.495 "raid": { 00:15:13.495 "uuid": "8df82abf-2f78-4e89-91f0-502c5a51c3ea", 00:15:13.495 "strip_size_kb": 0, 00:15:13.495 "state": "online", 00:15:13.495 "raid_level": "raid1", 00:15:13.495 "superblock": true, 00:15:13.495 "num_base_bdevs": 2, 00:15:13.495 "num_base_bdevs_discovered": 2, 00:15:13.495 "num_base_bdevs_operational": 2, 00:15:13.495 "base_bdevs_list": [ 00:15:13.495 { 00:15:13.495 "name": "pt1", 00:15:13.495 "uuid": "00000000-0000-0000-0000-000000000001", 00:15:13.495 "is_configured": true, 00:15:13.495 "data_offset": 256, 00:15:13.495 "data_size": 7936 00:15:13.495 }, 00:15:13.495 { 00:15:13.495 "name": "pt2", 00:15:13.495 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:13.495 "is_configured": true, 00:15:13.495 "data_offset": 256, 00:15:13.495 "data_size": 7936 00:15:13.495 } 00:15:13.495 ] 00:15:13.495 } 00:15:13.495 } 00:15:13.495 }' 00:15:13.495 23:49:01 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:15:13.755 23:49:01 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:15:13.755 pt2' 00:15:13.755 23:49:01 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:13.756 23:49:01 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='4096 32 false 0' 00:15:13.756 23:49:01 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:13.756 23:49:01 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:15:13.756 23:49:01 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:13.756 23:49:01 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:13.756 23:49:01 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:13.756 23:49:01 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:13.756 23:49:01 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 32 false 0' 00:15:13.756 23:49:01 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@193 -- # [[ 4096 32 false 0 == \4\0\9\6\ \3\2\ \f\a\l\s\e\ \0 ]] 00:15:13.756 23:49:01 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:13.756 23:49:01 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:15:13.756 23:49:01 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:13.756 23:49:01 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:13.756 23:49:01 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:13.756 23:49:01 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:13.756 23:49:01 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 32 false 0' 00:15:13.756 23:49:01 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@193 -- # [[ 4096 32 false 0 == \4\0\9\6\ \3\2\ \f\a\l\s\e\ \0 ]] 00:15:13.756 23:49:01 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:15:13.756 23:49:01 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:13.756 23:49:01 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:15:13.756 23:49:01 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:13.756 [2024-11-26 23:49:01.812692] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:13.756 23:49:01 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:13.756 23:49:01 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=8df82abf-2f78-4e89-91f0-502c5a51c3ea 00:15:13.756 23:49:01 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@436 -- # '[' -z 8df82abf-2f78-4e89-91f0-502c5a51c3ea ']' 00:15:13.756 23:49:01 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:15:13.756 23:49:01 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:13.756 23:49:01 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:13.756 [2024-11-26 23:49:01.856441] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:15:13.756 [2024-11-26 23:49:01.856510] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:15:13.756 [2024-11-26 23:49:01.856616] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:13.756 [2024-11-26 23:49:01.856689] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:13.756 [2024-11-26 23:49:01.856723] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:15:13.756 23:49:01 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:13.756 23:49:01 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:15:13.756 23:49:01 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:13.756 23:49:01 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:13.756 23:49:01 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:13.756 23:49:01 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:14.016 23:49:01 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:15:14.016 23:49:01 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:15:14.016 23:49:01 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:15:14.016 23:49:01 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:15:14.016 23:49:01 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:14.016 23:49:01 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:14.016 23:49:01 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:14.016 23:49:01 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:15:14.016 23:49:01 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:15:14.016 23:49:01 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:14.016 23:49:01 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:14.016 23:49:01 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:14.016 23:49:01 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:15:14.016 23:49:01 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:15:14.016 23:49:01 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:14.016 23:49:01 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:14.016 23:49:01 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:14.016 23:49:01 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:15:14.016 23:49:01 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:15:14.016 23:49:01 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@652 -- # local es=0 00:15:14.017 23:49:01 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:15:14.017 23:49:01 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:15:14.017 23:49:01 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:15:14.017 23:49:01 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:15:14.017 23:49:01 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:15:14.017 23:49:01 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:15:14.017 23:49:01 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:14.017 23:49:01 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:14.017 [2024-11-26 23:49:01.976239] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:15:14.017 [2024-11-26 23:49:01.978094] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:15:14.017 [2024-11-26 23:49:01.978203] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:15:14.017 [2024-11-26 23:49:01.978276] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:15:14.017 [2024-11-26 23:49:01.978315] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:15:14.017 [2024-11-26 23:49:01.978353] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state configuring 00:15:14.017 request: 00:15:14.017 { 00:15:14.017 "name": "raid_bdev1", 00:15:14.017 "raid_level": "raid1", 00:15:14.017 "base_bdevs": [ 00:15:14.017 "malloc1", 00:15:14.017 "malloc2" 00:15:14.017 ], 00:15:14.017 "superblock": false, 00:15:14.017 "method": "bdev_raid_create", 00:15:14.017 "req_id": 1 00:15:14.017 } 00:15:14.017 Got JSON-RPC error response 00:15:14.017 response: 00:15:14.017 { 00:15:14.017 "code": -17, 00:15:14.017 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:15:14.017 } 00:15:14.017 23:49:01 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:15:14.017 23:49:01 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@655 -- # es=1 00:15:14.017 23:49:01 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:15:14.017 23:49:01 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:15:14.017 23:49:01 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:15:14.017 23:49:01 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:15:14.017 23:49:01 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:14.017 23:49:01 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:14.017 23:49:01 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:14.017 23:49:01 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:14.017 23:49:02 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:15:14.017 23:49:02 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:15:14.017 23:49:02 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:15:14.017 23:49:02 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:14.017 23:49:02 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:14.017 [2024-11-26 23:49:02.024131] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:15:14.017 [2024-11-26 23:49:02.024223] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:14.017 [2024-11-26 23:49:02.024255] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:15:14.017 [2024-11-26 23:49:02.024281] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:14.017 [2024-11-26 23:49:02.026120] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:14.017 [2024-11-26 23:49:02.026183] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:15:14.017 [2024-11-26 23:49:02.026241] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:15:14.017 [2024-11-26 23:49:02.026306] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:15:14.017 pt1 00:15:14.017 23:49:02 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:14.017 23:49:02 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 2 00:15:14.017 23:49:02 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:14.017 23:49:02 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:14.017 23:49:02 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:14.017 23:49:02 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:14.017 23:49:02 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:14.017 23:49:02 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:14.017 23:49:02 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:14.017 23:49:02 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:14.017 23:49:02 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:14.017 23:49:02 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:14.017 23:49:02 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:14.017 23:49:02 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:14.017 23:49:02 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:14.017 23:49:02 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:14.017 23:49:02 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:14.017 "name": "raid_bdev1", 00:15:14.017 "uuid": "8df82abf-2f78-4e89-91f0-502c5a51c3ea", 00:15:14.017 "strip_size_kb": 0, 00:15:14.017 "state": "configuring", 00:15:14.017 "raid_level": "raid1", 00:15:14.017 "superblock": true, 00:15:14.017 "num_base_bdevs": 2, 00:15:14.017 "num_base_bdevs_discovered": 1, 00:15:14.017 "num_base_bdevs_operational": 2, 00:15:14.017 "base_bdevs_list": [ 00:15:14.017 { 00:15:14.017 "name": "pt1", 00:15:14.017 "uuid": "00000000-0000-0000-0000-000000000001", 00:15:14.017 "is_configured": true, 00:15:14.017 "data_offset": 256, 00:15:14.017 "data_size": 7936 00:15:14.017 }, 00:15:14.017 { 00:15:14.017 "name": null, 00:15:14.017 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:14.017 "is_configured": false, 00:15:14.017 "data_offset": 256, 00:15:14.017 "data_size": 7936 00:15:14.017 } 00:15:14.017 ] 00:15:14.017 }' 00:15:14.017 23:49:02 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:14.017 23:49:02 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:14.588 23:49:02 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@470 -- # '[' 2 -gt 2 ']' 00:15:14.588 23:49:02 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:15:14.588 23:49:02 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:15:14.588 23:49:02 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:15:14.588 23:49:02 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:14.588 23:49:02 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:14.588 [2024-11-26 23:49:02.475432] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:15:14.588 [2024-11-26 23:49:02.475479] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:14.588 [2024-11-26 23:49:02.475497] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:15:14.588 [2024-11-26 23:49:02.475505] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:14.588 [2024-11-26 23:49:02.475636] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:14.588 [2024-11-26 23:49:02.475649] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:15:14.588 [2024-11-26 23:49:02.475688] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:15:14.588 [2024-11-26 23:49:02.475708] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:15:14.588 [2024-11-26 23:49:02.475784] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:15:14.588 [2024-11-26 23:49:02.475791] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:15:14.588 [2024-11-26 23:49:02.475852] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:15:14.588 [2024-11-26 23:49:02.475923] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:15:14.588 [2024-11-26 23:49:02.475933] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:15:14.588 [2024-11-26 23:49:02.475986] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:14.588 pt2 00:15:14.588 23:49:02 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:14.588 23:49:02 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:15:14.588 23:49:02 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:15:14.588 23:49:02 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:15:14.588 23:49:02 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:14.588 23:49:02 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:14.588 23:49:02 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:14.588 23:49:02 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:14.588 23:49:02 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:14.588 23:49:02 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:14.588 23:49:02 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:14.588 23:49:02 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:14.588 23:49:02 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:14.588 23:49:02 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:14.588 23:49:02 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:14.588 23:49:02 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:14.588 23:49:02 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:14.588 23:49:02 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:14.588 23:49:02 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:14.588 "name": "raid_bdev1", 00:15:14.588 "uuid": "8df82abf-2f78-4e89-91f0-502c5a51c3ea", 00:15:14.588 "strip_size_kb": 0, 00:15:14.588 "state": "online", 00:15:14.588 "raid_level": "raid1", 00:15:14.588 "superblock": true, 00:15:14.588 "num_base_bdevs": 2, 00:15:14.588 "num_base_bdevs_discovered": 2, 00:15:14.588 "num_base_bdevs_operational": 2, 00:15:14.588 "base_bdevs_list": [ 00:15:14.588 { 00:15:14.588 "name": "pt1", 00:15:14.588 "uuid": "00000000-0000-0000-0000-000000000001", 00:15:14.588 "is_configured": true, 00:15:14.588 "data_offset": 256, 00:15:14.588 "data_size": 7936 00:15:14.588 }, 00:15:14.588 { 00:15:14.588 "name": "pt2", 00:15:14.588 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:14.588 "is_configured": true, 00:15:14.588 "data_offset": 256, 00:15:14.588 "data_size": 7936 00:15:14.588 } 00:15:14.588 ] 00:15:14.588 }' 00:15:14.588 23:49:02 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:14.588 23:49:02 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:14.848 23:49:02 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:15:14.848 23:49:02 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:15:14.849 23:49:02 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:15:14.849 23:49:02 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:15:14.849 23:49:02 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@184 -- # local name 00:15:14.849 23:49:02 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:15:14.849 23:49:02 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:15:14.849 23:49:02 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:15:14.849 23:49:02 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:14.849 23:49:02 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:14.849 [2024-11-26 23:49:02.902978] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:14.849 23:49:02 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:14.849 23:49:02 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:15:14.849 "name": "raid_bdev1", 00:15:14.849 "aliases": [ 00:15:14.849 "8df82abf-2f78-4e89-91f0-502c5a51c3ea" 00:15:14.849 ], 00:15:14.849 "product_name": "Raid Volume", 00:15:14.849 "block_size": 4096, 00:15:14.849 "num_blocks": 7936, 00:15:14.849 "uuid": "8df82abf-2f78-4e89-91f0-502c5a51c3ea", 00:15:14.849 "md_size": 32, 00:15:14.849 "md_interleave": false, 00:15:14.849 "dif_type": 0, 00:15:14.849 "assigned_rate_limits": { 00:15:14.849 "rw_ios_per_sec": 0, 00:15:14.849 "rw_mbytes_per_sec": 0, 00:15:14.849 "r_mbytes_per_sec": 0, 00:15:14.849 "w_mbytes_per_sec": 0 00:15:14.849 }, 00:15:14.849 "claimed": false, 00:15:14.849 "zoned": false, 00:15:14.849 "supported_io_types": { 00:15:14.849 "read": true, 00:15:14.849 "write": true, 00:15:14.849 "unmap": false, 00:15:14.849 "flush": false, 00:15:14.849 "reset": true, 00:15:14.849 "nvme_admin": false, 00:15:14.849 "nvme_io": false, 00:15:14.849 "nvme_io_md": false, 00:15:14.849 "write_zeroes": true, 00:15:14.849 "zcopy": false, 00:15:14.849 "get_zone_info": false, 00:15:14.849 "zone_management": false, 00:15:14.849 "zone_append": false, 00:15:14.849 "compare": false, 00:15:14.849 "compare_and_write": false, 00:15:14.849 "abort": false, 00:15:14.849 "seek_hole": false, 00:15:14.849 "seek_data": false, 00:15:14.849 "copy": false, 00:15:14.849 "nvme_iov_md": false 00:15:14.849 }, 00:15:14.849 "memory_domains": [ 00:15:14.849 { 00:15:14.849 "dma_device_id": "system", 00:15:14.849 "dma_device_type": 1 00:15:14.849 }, 00:15:14.849 { 00:15:14.849 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:14.849 "dma_device_type": 2 00:15:14.849 }, 00:15:14.849 { 00:15:14.849 "dma_device_id": "system", 00:15:14.849 "dma_device_type": 1 00:15:14.849 }, 00:15:14.849 { 00:15:14.849 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:14.849 "dma_device_type": 2 00:15:14.849 } 00:15:14.849 ], 00:15:14.849 "driver_specific": { 00:15:14.849 "raid": { 00:15:14.849 "uuid": "8df82abf-2f78-4e89-91f0-502c5a51c3ea", 00:15:14.849 "strip_size_kb": 0, 00:15:14.849 "state": "online", 00:15:14.849 "raid_level": "raid1", 00:15:14.849 "superblock": true, 00:15:14.849 "num_base_bdevs": 2, 00:15:14.849 "num_base_bdevs_discovered": 2, 00:15:14.849 "num_base_bdevs_operational": 2, 00:15:14.849 "base_bdevs_list": [ 00:15:14.849 { 00:15:14.849 "name": "pt1", 00:15:14.849 "uuid": "00000000-0000-0000-0000-000000000001", 00:15:14.849 "is_configured": true, 00:15:14.849 "data_offset": 256, 00:15:14.849 "data_size": 7936 00:15:14.849 }, 00:15:14.849 { 00:15:14.849 "name": "pt2", 00:15:14.849 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:14.849 "is_configured": true, 00:15:14.849 "data_offset": 256, 00:15:14.849 "data_size": 7936 00:15:14.849 } 00:15:14.849 ] 00:15:14.849 } 00:15:14.849 } 00:15:14.849 }' 00:15:14.849 23:49:02 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:15:15.108 23:49:02 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:15:15.108 pt2' 00:15:15.108 23:49:02 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:15.108 23:49:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='4096 32 false 0' 00:15:15.108 23:49:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:15.108 23:49:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:15.108 23:49:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:15:15.108 23:49:03 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:15.108 23:49:03 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:15.108 23:49:03 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:15.108 23:49:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 32 false 0' 00:15:15.108 23:49:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@193 -- # [[ 4096 32 false 0 == \4\0\9\6\ \3\2\ \f\a\l\s\e\ \0 ]] 00:15:15.108 23:49:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:15.108 23:49:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:15.108 23:49:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:15:15.108 23:49:03 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:15.108 23:49:03 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:15.109 23:49:03 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:15.109 23:49:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 32 false 0' 00:15:15.109 23:49:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@193 -- # [[ 4096 32 false 0 == \4\0\9\6\ \3\2\ \f\a\l\s\e\ \0 ]] 00:15:15.109 23:49:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:15:15.109 23:49:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:15:15.109 23:49:03 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:15.109 23:49:03 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:15.109 [2024-11-26 23:49:03.142626] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:15.109 23:49:03 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:15.109 23:49:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@487 -- # '[' 8df82abf-2f78-4e89-91f0-502c5a51c3ea '!=' 8df82abf-2f78-4e89-91f0-502c5a51c3ea ']' 00:15:15.109 23:49:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@491 -- # has_redundancy raid1 00:15:15.109 23:49:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@198 -- # case $1 in 00:15:15.109 23:49:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@199 -- # return 0 00:15:15.109 23:49:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@493 -- # rpc_cmd bdev_passthru_delete pt1 00:15:15.109 23:49:03 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:15.109 23:49:03 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:15.109 [2024-11-26 23:49:03.186334] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt1 00:15:15.109 23:49:03 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:15.109 23:49:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@496 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:15:15.109 23:49:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:15.109 23:49:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:15.109 23:49:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:15.109 23:49:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:15.109 23:49:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:15:15.109 23:49:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:15.109 23:49:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:15.109 23:49:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:15.109 23:49:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:15.109 23:49:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:15.109 23:49:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:15.109 23:49:03 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:15.109 23:49:03 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:15.109 23:49:03 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:15.368 23:49:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:15.368 "name": "raid_bdev1", 00:15:15.368 "uuid": "8df82abf-2f78-4e89-91f0-502c5a51c3ea", 00:15:15.368 "strip_size_kb": 0, 00:15:15.368 "state": "online", 00:15:15.368 "raid_level": "raid1", 00:15:15.368 "superblock": true, 00:15:15.368 "num_base_bdevs": 2, 00:15:15.368 "num_base_bdevs_discovered": 1, 00:15:15.368 "num_base_bdevs_operational": 1, 00:15:15.368 "base_bdevs_list": [ 00:15:15.368 { 00:15:15.368 "name": null, 00:15:15.368 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:15.368 "is_configured": false, 00:15:15.368 "data_offset": 0, 00:15:15.368 "data_size": 7936 00:15:15.368 }, 00:15:15.368 { 00:15:15.368 "name": "pt2", 00:15:15.368 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:15.368 "is_configured": true, 00:15:15.368 "data_offset": 256, 00:15:15.368 "data_size": 7936 00:15:15.368 } 00:15:15.368 ] 00:15:15.368 }' 00:15:15.368 23:49:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:15.368 23:49:03 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:15.627 23:49:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@499 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:15:15.627 23:49:03 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:15.627 23:49:03 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:15.627 [2024-11-26 23:49:03.625541] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:15:15.627 [2024-11-26 23:49:03.625614] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:15:15.627 [2024-11-26 23:49:03.625701] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:15.627 [2024-11-26 23:49:03.625754] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:15.627 [2024-11-26 23:49:03.625796] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:15:15.627 23:49:03 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:15.627 23:49:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@500 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:15.627 23:49:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@500 -- # jq -r '.[]' 00:15:15.627 23:49:03 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:15.627 23:49:03 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:15.627 23:49:03 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:15.627 23:49:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@500 -- # raid_bdev= 00:15:15.627 23:49:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@501 -- # '[' -n '' ']' 00:15:15.627 23:49:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@506 -- # (( i = 1 )) 00:15:15.627 23:49:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:15:15.627 23:49:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt2 00:15:15.627 23:49:03 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:15.627 23:49:03 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:15.627 23:49:03 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:15.627 23:49:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:15:15.627 23:49:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:15:15.627 23:49:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@511 -- # (( i = 1 )) 00:15:15.627 23:49:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:15:15.627 23:49:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@519 -- # i=1 00:15:15.627 23:49:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@520 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:15:15.627 23:49:03 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:15.627 23:49:03 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:15.627 [2024-11-26 23:49:03.701423] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:15:15.627 [2024-11-26 23:49:03.701463] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:15.627 [2024-11-26 23:49:03.701478] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008780 00:15:15.627 [2024-11-26 23:49:03.701486] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:15.627 [2024-11-26 23:49:03.703339] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:15.627 [2024-11-26 23:49:03.703389] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:15:15.627 [2024-11-26 23:49:03.703436] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:15:15.627 [2024-11-26 23:49:03.703466] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:15:15.627 [2024-11-26 23:49:03.703527] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:15:15.627 [2024-11-26 23:49:03.703535] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:15:15.627 [2024-11-26 23:49:03.703593] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:15:15.627 [2024-11-26 23:49:03.703661] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:15:15.627 [2024-11-26 23:49:03.703670] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001c80 00:15:15.627 [2024-11-26 23:49:03.703724] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:15.627 pt2 00:15:15.627 23:49:03 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:15.627 23:49:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@523 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:15:15.627 23:49:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:15.627 23:49:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:15.627 23:49:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:15.627 23:49:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:15.627 23:49:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:15:15.627 23:49:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:15.627 23:49:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:15.627 23:49:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:15.627 23:49:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:15.627 23:49:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:15.627 23:49:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:15.627 23:49:03 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:15.627 23:49:03 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:15.627 23:49:03 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:15.887 23:49:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:15.887 "name": "raid_bdev1", 00:15:15.887 "uuid": "8df82abf-2f78-4e89-91f0-502c5a51c3ea", 00:15:15.887 "strip_size_kb": 0, 00:15:15.887 "state": "online", 00:15:15.887 "raid_level": "raid1", 00:15:15.887 "superblock": true, 00:15:15.887 "num_base_bdevs": 2, 00:15:15.887 "num_base_bdevs_discovered": 1, 00:15:15.887 "num_base_bdevs_operational": 1, 00:15:15.887 "base_bdevs_list": [ 00:15:15.887 { 00:15:15.887 "name": null, 00:15:15.887 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:15.887 "is_configured": false, 00:15:15.887 "data_offset": 256, 00:15:15.887 "data_size": 7936 00:15:15.887 }, 00:15:15.887 { 00:15:15.887 "name": "pt2", 00:15:15.887 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:15.887 "is_configured": true, 00:15:15.887 "data_offset": 256, 00:15:15.887 "data_size": 7936 00:15:15.887 } 00:15:15.887 ] 00:15:15.887 }' 00:15:15.887 23:49:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:15.887 23:49:03 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:16.147 23:49:04 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@526 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:15:16.147 23:49:04 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:16.147 23:49:04 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:16.147 [2024-11-26 23:49:04.168590] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:15:16.147 [2024-11-26 23:49:04.168662] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:15:16.147 [2024-11-26 23:49:04.168736] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:16.147 [2024-11-26 23:49:04.168787] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:16.147 [2024-11-26 23:49:04.168849] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name raid_bdev1, state offline 00:15:16.147 23:49:04 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:16.147 23:49:04 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@527 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:16.147 23:49:04 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:16.147 23:49:04 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:16.147 23:49:04 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@527 -- # jq -r '.[]' 00:15:16.147 23:49:04 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:16.147 23:49:04 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@527 -- # raid_bdev= 00:15:16.147 23:49:04 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@528 -- # '[' -n '' ']' 00:15:16.147 23:49:04 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@532 -- # '[' 2 -gt 2 ']' 00:15:16.147 23:49:04 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@540 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:15:16.147 23:49:04 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:16.147 23:49:04 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:16.147 [2024-11-26 23:49:04.232499] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:15:16.147 [2024-11-26 23:49:04.232621] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:16.147 [2024-11-26 23:49:04.232654] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008d80 00:15:16.147 [2024-11-26 23:49:04.232687] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:16.147 [2024-11-26 23:49:04.234569] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:16.147 [2024-11-26 23:49:04.234651] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:15:16.147 [2024-11-26 23:49:04.234734] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:15:16.147 [2024-11-26 23:49:04.234802] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:15:16.147 [2024-11-26 23:49:04.234916] bdev_raid.c:3685:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev pt2 (4) greater than existing raid bdev raid_bdev1 (2) 00:15:16.147 [2024-11-26 23:49:04.234969] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:15:16.147 [2024-11-26 23:49:04.235021] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002000 name raid_bdev1, state configuring 00:15:16.147 [2024-11-26 23:49:04.235097] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:15:16.147 [2024-11-26 23:49:04.235188] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000002380 00:15:16.147 [2024-11-26 23:49:04.235231] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:15:16.147 [2024-11-26 23:49:04.235298] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:15:16.147 [2024-11-26 23:49:04.235419] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000002380 00:15:16.147 [2024-11-26 23:49:04.235460] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000002380 00:15:16.147 [2024-11-26 23:49:04.235573] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:16.147 pt1 00:15:16.147 23:49:04 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:16.147 23:49:04 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@542 -- # '[' 2 -gt 2 ']' 00:15:16.147 23:49:04 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@554 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:15:16.147 23:49:04 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:16.147 23:49:04 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:16.147 23:49:04 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:16.147 23:49:04 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:16.147 23:49:04 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:15:16.147 23:49:04 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:16.147 23:49:04 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:16.147 23:49:04 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:16.147 23:49:04 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:16.147 23:49:04 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:16.148 23:49:04 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:16.148 23:49:04 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:16.148 23:49:04 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:16.148 23:49:04 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:16.407 23:49:04 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:16.407 "name": "raid_bdev1", 00:15:16.407 "uuid": "8df82abf-2f78-4e89-91f0-502c5a51c3ea", 00:15:16.407 "strip_size_kb": 0, 00:15:16.407 "state": "online", 00:15:16.407 "raid_level": "raid1", 00:15:16.407 "superblock": true, 00:15:16.407 "num_base_bdevs": 2, 00:15:16.407 "num_base_bdevs_discovered": 1, 00:15:16.407 "num_base_bdevs_operational": 1, 00:15:16.407 "base_bdevs_list": [ 00:15:16.407 { 00:15:16.407 "name": null, 00:15:16.407 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:16.407 "is_configured": false, 00:15:16.407 "data_offset": 256, 00:15:16.407 "data_size": 7936 00:15:16.407 }, 00:15:16.407 { 00:15:16.407 "name": "pt2", 00:15:16.407 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:16.407 "is_configured": true, 00:15:16.407 "data_offset": 256, 00:15:16.407 "data_size": 7936 00:15:16.407 } 00:15:16.407 ] 00:15:16.407 }' 00:15:16.407 23:49:04 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:16.407 23:49:04 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:16.667 23:49:04 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@555 -- # rpc_cmd bdev_raid_get_bdevs online 00:15:16.667 23:49:04 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:16.667 23:49:04 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:16.667 23:49:04 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@555 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:15:16.667 23:49:04 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:16.667 23:49:04 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@555 -- # [[ false == \f\a\l\s\e ]] 00:15:16.667 23:49:04 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@558 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:15:16.667 23:49:04 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@558 -- # jq -r '.[] | .uuid' 00:15:16.667 23:49:04 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:16.667 23:49:04 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:16.667 [2024-11-26 23:49:04.731878] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:16.667 23:49:04 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:16.667 23:49:04 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@558 -- # '[' 8df82abf-2f78-4e89-91f0-502c5a51c3ea '!=' 8df82abf-2f78-4e89-91f0-502c5a51c3ea ']' 00:15:16.667 23:49:04 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@563 -- # killprocess 97430 00:15:16.668 23:49:04 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@954 -- # '[' -z 97430 ']' 00:15:16.668 23:49:04 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@958 -- # kill -0 97430 00:15:16.668 23:49:04 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@959 -- # uname 00:15:16.668 23:49:04 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:15:16.668 23:49:04 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 97430 00:15:16.927 23:49:04 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:15:16.927 23:49:04 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:15:16.927 23:49:04 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@972 -- # echo 'killing process with pid 97430' 00:15:16.927 killing process with pid 97430 00:15:16.927 23:49:04 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@973 -- # kill 97430 00:15:16.927 [2024-11-26 23:49:04.815879] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:15:16.927 [2024-11-26 23:49:04.815945] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:16.927 [2024-11-26 23:49:04.815984] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:16.927 [2024-11-26 23:49:04.815992] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002380 name raid_bdev1, state offline 00:15:16.927 23:49:04 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@978 -- # wait 97430 00:15:16.927 [2024-11-26 23:49:04.839402] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:15:16.927 23:49:05 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@565 -- # return 0 00:15:16.927 00:15:16.927 real 0m4.917s 00:15:16.927 user 0m8.062s 00:15:16.927 sys 0m1.067s 00:15:16.927 23:49:05 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@1130 -- # xtrace_disable 00:15:16.927 ************************************ 00:15:16.927 END TEST raid_superblock_test_md_separate 00:15:16.927 ************************************ 00:15:16.927 23:49:05 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:17.186 23:49:05 bdev_raid -- bdev/bdev_raid.sh@1006 -- # '[' true = true ']' 00:15:17.186 23:49:05 bdev_raid -- bdev/bdev_raid.sh@1007 -- # run_test raid_rebuild_test_sb_md_separate raid_rebuild_test raid1 2 true false true 00:15:17.186 23:49:05 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:15:17.186 23:49:05 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:15:17.186 23:49:05 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:15:17.186 ************************************ 00:15:17.186 START TEST raid_rebuild_test_sb_md_separate 00:15:17.186 ************************************ 00:15:17.186 23:49:05 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@1129 -- # raid_rebuild_test raid1 2 true false true 00:15:17.186 23:49:05 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:15:17.186 23:49:05 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=2 00:15:17.187 23:49:05 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@571 -- # local superblock=true 00:15:17.187 23:49:05 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:15:17.187 23:49:05 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@573 -- # local verify=true 00:15:17.187 23:49:05 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:15:17.187 23:49:05 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:15:17.187 23:49:05 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:15:17.187 23:49:05 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:15:17.187 23:49:05 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:15:17.187 23:49:05 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:15:17.187 23:49:05 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:15:17.187 23:49:05 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:15:17.187 23:49:05 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:15:17.187 23:49:05 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:15:17.187 23:49:05 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:15:17.187 23:49:05 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@576 -- # local strip_size 00:15:17.187 23:49:05 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@577 -- # local create_arg 00:15:17.187 23:49:05 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:15:17.187 23:49:05 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@579 -- # local data_offset 00:15:17.187 23:49:05 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:15:17.187 23:49:05 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:15:17.187 23:49:05 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@592 -- # '[' true = true ']' 00:15:17.187 23:49:05 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@593 -- # create_arg+=' -s' 00:15:17.187 23:49:05 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@597 -- # raid_pid=97742 00:15:17.187 23:49:05 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@598 -- # waitforlisten 97742 00:15:17.187 23:49:05 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:15:17.187 23:49:05 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@835 -- # '[' -z 97742 ']' 00:15:17.187 23:49:05 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:15:17.187 23:49:05 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@840 -- # local max_retries=100 00:15:17.187 23:49:05 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:15:17.187 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:15:17.187 23:49:05 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@844 -- # xtrace_disable 00:15:17.187 23:49:05 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:17.187 [2024-11-26 23:49:05.231979] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:15:17.187 [2024-11-26 23:49:05.232169] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.ealI/O size of 3145728 is greater than zero copy threshold (65536). 00:15:17.187 Zero copy mechanism will not be used. 00:15:17.187 :6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid97742 ] 00:15:17.447 [2024-11-26 23:49:05.387422] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:15:17.447 [2024-11-26 23:49:05.411729] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:15:17.447 [2024-11-26 23:49:05.454006] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:15:17.447 [2024-11-26 23:49:05.454116] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:15:18.029 23:49:06 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:15:18.029 23:49:06 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@868 -- # return 0 00:15:18.029 23:49:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:15:18.029 23:49:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -b BaseBdev1_malloc 00:15:18.029 23:49:06 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:18.029 23:49:06 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:18.029 BaseBdev1_malloc 00:15:18.029 23:49:06 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:18.029 23:49:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:15:18.029 23:49:06 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:18.029 23:49:06 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:18.029 [2024-11-26 23:49:06.069736] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:15:18.029 [2024-11-26 23:49:06.069805] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:18.029 [2024-11-26 23:49:06.069832] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:15:18.029 [2024-11-26 23:49:06.069842] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:18.029 [2024-11-26 23:49:06.071732] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:18.029 [2024-11-26 23:49:06.071842] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:15:18.029 BaseBdev1 00:15:18.029 23:49:06 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:18.029 23:49:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:15:18.029 23:49:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -b BaseBdev2_malloc 00:15:18.029 23:49:06 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:18.029 23:49:06 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:18.029 BaseBdev2_malloc 00:15:18.029 23:49:06 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:18.029 23:49:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:15:18.029 23:49:06 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:18.029 23:49:06 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:18.029 [2024-11-26 23:49:06.098693] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:15:18.029 [2024-11-26 23:49:06.098833] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:18.029 [2024-11-26 23:49:06.098861] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:15:18.029 [2024-11-26 23:49:06.098870] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:18.029 [2024-11-26 23:49:06.100728] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:18.029 [2024-11-26 23:49:06.100763] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:15:18.029 BaseBdev2 00:15:18.029 23:49:06 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:18.029 23:49:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -b spare_malloc 00:15:18.029 23:49:06 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:18.029 23:49:06 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:18.029 spare_malloc 00:15:18.029 23:49:06 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:18.029 23:49:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:15:18.029 23:49:06 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:18.029 23:49:06 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:18.029 spare_delay 00:15:18.029 23:49:06 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:18.029 23:49:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:15:18.029 23:49:06 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:18.029 23:49:06 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:18.353 [2024-11-26 23:49:06.154906] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:15:18.353 [2024-11-26 23:49:06.154966] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:18.353 [2024-11-26 23:49:06.154993] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:15:18.353 [2024-11-26 23:49:06.155005] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:18.353 [2024-11-26 23:49:06.157452] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:18.353 [2024-11-26 23:49:06.157493] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:15:18.353 spare 00:15:18.353 23:49:06 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:18.353 23:49:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 00:15:18.353 23:49:06 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:18.353 23:49:06 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:18.353 [2024-11-26 23:49:06.166909] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:15:18.353 [2024-11-26 23:49:06.168742] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:15:18.353 [2024-11-26 23:49:06.168927] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:15:18.353 [2024-11-26 23:49:06.168976] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:15:18.353 [2024-11-26 23:49:06.169082] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:15:18.353 [2024-11-26 23:49:06.169222] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:15:18.353 [2024-11-26 23:49:06.169264] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:15:18.353 [2024-11-26 23:49:06.169394] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:18.353 23:49:06 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:18.353 23:49:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:15:18.353 23:49:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:18.353 23:49:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:18.353 23:49:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:18.353 23:49:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:18.353 23:49:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:18.353 23:49:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:18.353 23:49:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:18.353 23:49:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:18.353 23:49:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:18.353 23:49:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:18.353 23:49:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:18.353 23:49:06 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:18.353 23:49:06 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:18.353 23:49:06 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:18.353 23:49:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:18.353 "name": "raid_bdev1", 00:15:18.353 "uuid": "1db8de6e-e5f5-48af-8510-2cccfe208f1d", 00:15:18.353 "strip_size_kb": 0, 00:15:18.353 "state": "online", 00:15:18.353 "raid_level": "raid1", 00:15:18.353 "superblock": true, 00:15:18.353 "num_base_bdevs": 2, 00:15:18.353 "num_base_bdevs_discovered": 2, 00:15:18.353 "num_base_bdevs_operational": 2, 00:15:18.353 "base_bdevs_list": [ 00:15:18.353 { 00:15:18.353 "name": "BaseBdev1", 00:15:18.353 "uuid": "fc1b083f-e919-59e1-a87b-2bed9afc4981", 00:15:18.353 "is_configured": true, 00:15:18.353 "data_offset": 256, 00:15:18.353 "data_size": 7936 00:15:18.353 }, 00:15:18.353 { 00:15:18.354 "name": "BaseBdev2", 00:15:18.354 "uuid": "2c118720-c56e-58b2-9ce8-0fe1ca19a44e", 00:15:18.354 "is_configured": true, 00:15:18.354 "data_offset": 256, 00:15:18.354 "data_size": 7936 00:15:18.354 } 00:15:18.354 ] 00:15:18.354 }' 00:15:18.354 23:49:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:18.354 23:49:06 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:18.613 23:49:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:15:18.613 23:49:06 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:18.613 23:49:06 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:18.613 23:49:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:15:18.613 [2024-11-26 23:49:06.562533] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:18.613 23:49:06 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:18.613 23:49:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=7936 00:15:18.613 23:49:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:15:18.613 23:49:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:18.613 23:49:06 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:18.613 23:49:06 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:18.613 23:49:06 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:18.613 23:49:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@619 -- # data_offset=256 00:15:18.613 23:49:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:15:18.613 23:49:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:15:18.613 23:49:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:15:18.613 23:49:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:15:18.613 23:49:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:15:18.613 23:49:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:15:18.613 23:49:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@10 -- # local bdev_list 00:15:18.613 23:49:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:15:18.613 23:49:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@11 -- # local nbd_list 00:15:18.613 23:49:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@12 -- # local i 00:15:18.613 23:49:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:15:18.613 23:49:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:15:18.613 23:49:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:15:18.872 [2024-11-26 23:49:06.837843] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:15:18.872 /dev/nbd0 00:15:18.872 23:49:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:15:18.872 23:49:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:15:18.872 23:49:06 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:15:18.872 23:49:06 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@873 -- # local i 00:15:18.872 23:49:06 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:15:18.872 23:49:06 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:15:18.872 23:49:06 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:15:18.872 23:49:06 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@877 -- # break 00:15:18.872 23:49:06 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:15:18.872 23:49:06 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:15:18.872 23:49:06 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:15:18.872 1+0 records in 00:15:18.872 1+0 records out 00:15:18.872 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000478914 s, 8.6 MB/s 00:15:18.872 23:49:06 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:18.872 23:49:06 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@890 -- # size=4096 00:15:18.872 23:49:06 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:18.872 23:49:06 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:15:18.872 23:49:06 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@893 -- # return 0 00:15:18.872 23:49:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:15:18.872 23:49:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:15:18.872 23:49:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@629 -- # '[' raid1 = raid5f ']' 00:15:18.872 23:49:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@633 -- # write_unit_size=1 00:15:18.872 23:49:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=4096 count=7936 oflag=direct 00:15:19.442 7936+0 records in 00:15:19.442 7936+0 records out 00:15:19.442 32505856 bytes (33 MB, 31 MiB) copied, 0.607944 s, 53.5 MB/s 00:15:19.442 23:49:07 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:15:19.442 23:49:07 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:15:19.442 23:49:07 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:15:19.442 23:49:07 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@50 -- # local nbd_list 00:15:19.442 23:49:07 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@51 -- # local i 00:15:19.442 23:49:07 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:15:19.442 23:49:07 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:15:19.714 23:49:07 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:15:19.714 [2024-11-26 23:49:07.739242] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:19.714 23:49:07 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:15:19.714 23:49:07 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:15:19.714 23:49:07 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:15:19.714 23:49:07 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:15:19.714 23:49:07 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:15:19.714 23:49:07 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@41 -- # break 00:15:19.714 23:49:07 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@45 -- # return 0 00:15:19.714 23:49:07 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:15:19.714 23:49:07 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:19.714 23:49:07 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:19.714 [2024-11-26 23:49:07.756704] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:15:19.714 23:49:07 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:19.714 23:49:07 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:15:19.715 23:49:07 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:19.715 23:49:07 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:19.715 23:49:07 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:19.715 23:49:07 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:19.715 23:49:07 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:15:19.715 23:49:07 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:19.715 23:49:07 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:19.715 23:49:07 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:19.715 23:49:07 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:19.715 23:49:07 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:19.715 23:49:07 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:19.715 23:49:07 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:19.715 23:49:07 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:19.715 23:49:07 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:19.715 23:49:07 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:19.715 "name": "raid_bdev1", 00:15:19.715 "uuid": "1db8de6e-e5f5-48af-8510-2cccfe208f1d", 00:15:19.715 "strip_size_kb": 0, 00:15:19.715 "state": "online", 00:15:19.715 "raid_level": "raid1", 00:15:19.716 "superblock": true, 00:15:19.716 "num_base_bdevs": 2, 00:15:19.716 "num_base_bdevs_discovered": 1, 00:15:19.716 "num_base_bdevs_operational": 1, 00:15:19.716 "base_bdevs_list": [ 00:15:19.716 { 00:15:19.716 "name": null, 00:15:19.716 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:19.716 "is_configured": false, 00:15:19.716 "data_offset": 0, 00:15:19.716 "data_size": 7936 00:15:19.716 }, 00:15:19.716 { 00:15:19.716 "name": "BaseBdev2", 00:15:19.716 "uuid": "2c118720-c56e-58b2-9ce8-0fe1ca19a44e", 00:15:19.716 "is_configured": true, 00:15:19.716 "data_offset": 256, 00:15:19.716 "data_size": 7936 00:15:19.716 } 00:15:19.716 ] 00:15:19.716 }' 00:15:19.716 23:49:07 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:19.716 23:49:07 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:20.290 23:49:08 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:15:20.290 23:49:08 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:20.290 23:49:08 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:20.290 [2024-11-26 23:49:08.243901] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:15:20.290 [2024-11-26 23:49:08.246529] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d00019c960 00:15:20.290 [2024-11-26 23:49:08.248414] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:15:20.290 23:49:08 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:20.290 23:49:08 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@647 -- # sleep 1 00:15:21.230 23:49:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:21.230 23:49:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:21.230 23:49:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:21.230 23:49:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:21.230 23:49:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:21.230 23:49:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:21.230 23:49:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:21.230 23:49:09 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:21.230 23:49:09 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:21.230 23:49:09 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:21.230 23:49:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:21.230 "name": "raid_bdev1", 00:15:21.230 "uuid": "1db8de6e-e5f5-48af-8510-2cccfe208f1d", 00:15:21.230 "strip_size_kb": 0, 00:15:21.230 "state": "online", 00:15:21.230 "raid_level": "raid1", 00:15:21.230 "superblock": true, 00:15:21.230 "num_base_bdevs": 2, 00:15:21.230 "num_base_bdevs_discovered": 2, 00:15:21.230 "num_base_bdevs_operational": 2, 00:15:21.230 "process": { 00:15:21.230 "type": "rebuild", 00:15:21.230 "target": "spare", 00:15:21.230 "progress": { 00:15:21.230 "blocks": 2560, 00:15:21.230 "percent": 32 00:15:21.230 } 00:15:21.230 }, 00:15:21.230 "base_bdevs_list": [ 00:15:21.230 { 00:15:21.230 "name": "spare", 00:15:21.230 "uuid": "6f15d6e8-07d1-535a-b452-9aefb81194d0", 00:15:21.230 "is_configured": true, 00:15:21.230 "data_offset": 256, 00:15:21.230 "data_size": 7936 00:15:21.230 }, 00:15:21.230 { 00:15:21.230 "name": "BaseBdev2", 00:15:21.230 "uuid": "2c118720-c56e-58b2-9ce8-0fe1ca19a44e", 00:15:21.230 "is_configured": true, 00:15:21.230 "data_offset": 256, 00:15:21.230 "data_size": 7936 00:15:21.230 } 00:15:21.230 ] 00:15:21.230 }' 00:15:21.230 23:49:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:21.230 23:49:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:21.230 23:49:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:21.491 23:49:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:21.491 23:49:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:15:21.491 23:49:09 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:21.491 23:49:09 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:21.491 [2024-11-26 23:49:09.371322] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:21.491 [2024-11-26 23:49:09.452946] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:15:21.491 [2024-11-26 23:49:09.453002] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:21.491 [2024-11-26 23:49:09.453019] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:21.491 [2024-11-26 23:49:09.453026] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:15:21.491 23:49:09 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:21.491 23:49:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:15:21.491 23:49:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:21.491 23:49:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:21.491 23:49:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:21.491 23:49:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:21.491 23:49:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:15:21.491 23:49:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:21.491 23:49:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:21.491 23:49:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:21.491 23:49:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:21.491 23:49:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:21.491 23:49:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:21.491 23:49:09 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:21.491 23:49:09 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:21.491 23:49:09 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:21.491 23:49:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:21.491 "name": "raid_bdev1", 00:15:21.491 "uuid": "1db8de6e-e5f5-48af-8510-2cccfe208f1d", 00:15:21.491 "strip_size_kb": 0, 00:15:21.491 "state": "online", 00:15:21.491 "raid_level": "raid1", 00:15:21.491 "superblock": true, 00:15:21.491 "num_base_bdevs": 2, 00:15:21.491 "num_base_bdevs_discovered": 1, 00:15:21.491 "num_base_bdevs_operational": 1, 00:15:21.491 "base_bdevs_list": [ 00:15:21.491 { 00:15:21.491 "name": null, 00:15:21.491 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:21.491 "is_configured": false, 00:15:21.491 "data_offset": 0, 00:15:21.491 "data_size": 7936 00:15:21.491 }, 00:15:21.491 { 00:15:21.491 "name": "BaseBdev2", 00:15:21.491 "uuid": "2c118720-c56e-58b2-9ce8-0fe1ca19a44e", 00:15:21.491 "is_configured": true, 00:15:21.491 "data_offset": 256, 00:15:21.491 "data_size": 7936 00:15:21.491 } 00:15:21.491 ] 00:15:21.491 }' 00:15:21.491 23:49:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:21.491 23:49:09 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:21.752 23:49:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:15:21.752 23:49:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:21.752 23:49:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:15:21.752 23:49:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=none 00:15:21.752 23:49:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:21.752 23:49:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:21.752 23:49:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:21.752 23:49:09 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:21.752 23:49:09 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:22.012 23:49:09 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:22.012 23:49:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:22.012 "name": "raid_bdev1", 00:15:22.012 "uuid": "1db8de6e-e5f5-48af-8510-2cccfe208f1d", 00:15:22.012 "strip_size_kb": 0, 00:15:22.012 "state": "online", 00:15:22.012 "raid_level": "raid1", 00:15:22.012 "superblock": true, 00:15:22.012 "num_base_bdevs": 2, 00:15:22.012 "num_base_bdevs_discovered": 1, 00:15:22.012 "num_base_bdevs_operational": 1, 00:15:22.012 "base_bdevs_list": [ 00:15:22.012 { 00:15:22.012 "name": null, 00:15:22.012 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:22.012 "is_configured": false, 00:15:22.012 "data_offset": 0, 00:15:22.012 "data_size": 7936 00:15:22.012 }, 00:15:22.012 { 00:15:22.012 "name": "BaseBdev2", 00:15:22.012 "uuid": "2c118720-c56e-58b2-9ce8-0fe1ca19a44e", 00:15:22.012 "is_configured": true, 00:15:22.012 "data_offset": 256, 00:15:22.012 "data_size": 7936 00:15:22.012 } 00:15:22.012 ] 00:15:22.012 }' 00:15:22.012 23:49:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:22.012 23:49:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:15:22.012 23:49:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:22.012 23:49:10 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:15:22.012 23:49:10 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:15:22.012 23:49:10 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:22.012 23:49:10 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:22.012 [2024-11-26 23:49:10.011270] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:15:22.012 [2024-11-26 23:49:10.013290] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d00019ca30 00:15:22.012 [2024-11-26 23:49:10.015177] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:15:22.012 23:49:10 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:22.012 23:49:10 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@663 -- # sleep 1 00:15:22.953 23:49:11 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:22.953 23:49:11 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:22.953 23:49:11 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:22.953 23:49:11 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:22.953 23:49:11 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:22.953 23:49:11 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:22.953 23:49:11 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:22.953 23:49:11 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:22.953 23:49:11 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:22.953 23:49:11 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:22.953 23:49:11 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:22.953 "name": "raid_bdev1", 00:15:22.953 "uuid": "1db8de6e-e5f5-48af-8510-2cccfe208f1d", 00:15:22.953 "strip_size_kb": 0, 00:15:22.953 "state": "online", 00:15:22.953 "raid_level": "raid1", 00:15:22.953 "superblock": true, 00:15:22.953 "num_base_bdevs": 2, 00:15:22.953 "num_base_bdevs_discovered": 2, 00:15:22.953 "num_base_bdevs_operational": 2, 00:15:22.953 "process": { 00:15:22.953 "type": "rebuild", 00:15:22.953 "target": "spare", 00:15:22.953 "progress": { 00:15:22.953 "blocks": 2560, 00:15:22.953 "percent": 32 00:15:22.953 } 00:15:22.953 }, 00:15:22.953 "base_bdevs_list": [ 00:15:22.953 { 00:15:22.953 "name": "spare", 00:15:22.953 "uuid": "6f15d6e8-07d1-535a-b452-9aefb81194d0", 00:15:22.953 "is_configured": true, 00:15:22.953 "data_offset": 256, 00:15:22.953 "data_size": 7936 00:15:22.953 }, 00:15:22.953 { 00:15:22.953 "name": "BaseBdev2", 00:15:22.953 "uuid": "2c118720-c56e-58b2-9ce8-0fe1ca19a44e", 00:15:22.953 "is_configured": true, 00:15:22.953 "data_offset": 256, 00:15:22.953 "data_size": 7936 00:15:22.953 } 00:15:22.953 ] 00:15:22.953 }' 00:15:22.953 23:49:11 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:23.214 23:49:11 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:23.214 23:49:11 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:23.214 23:49:11 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:23.214 23:49:11 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@666 -- # '[' true = true ']' 00:15:23.214 23:49:11 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@666 -- # '[' = false ']' 00:15:23.214 /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh: line 666: [: =: unary operator expected 00:15:23.214 23:49:11 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=2 00:15:23.214 23:49:11 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:15:23.214 23:49:11 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@693 -- # '[' 2 -gt 2 ']' 00:15:23.214 23:49:11 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@706 -- # local timeout=583 00:15:23.214 23:49:11 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:23.214 23:49:11 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:23.214 23:49:11 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:23.214 23:49:11 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:23.214 23:49:11 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:23.214 23:49:11 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:23.214 23:49:11 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:23.214 23:49:11 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:23.214 23:49:11 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:23.214 23:49:11 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:23.214 23:49:11 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:23.214 23:49:11 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:23.214 "name": "raid_bdev1", 00:15:23.214 "uuid": "1db8de6e-e5f5-48af-8510-2cccfe208f1d", 00:15:23.215 "strip_size_kb": 0, 00:15:23.215 "state": "online", 00:15:23.215 "raid_level": "raid1", 00:15:23.215 "superblock": true, 00:15:23.215 "num_base_bdevs": 2, 00:15:23.215 "num_base_bdevs_discovered": 2, 00:15:23.215 "num_base_bdevs_operational": 2, 00:15:23.215 "process": { 00:15:23.215 "type": "rebuild", 00:15:23.215 "target": "spare", 00:15:23.215 "progress": { 00:15:23.215 "blocks": 2816, 00:15:23.215 "percent": 35 00:15:23.215 } 00:15:23.215 }, 00:15:23.215 "base_bdevs_list": [ 00:15:23.215 { 00:15:23.215 "name": "spare", 00:15:23.215 "uuid": "6f15d6e8-07d1-535a-b452-9aefb81194d0", 00:15:23.215 "is_configured": true, 00:15:23.215 "data_offset": 256, 00:15:23.215 "data_size": 7936 00:15:23.215 }, 00:15:23.215 { 00:15:23.215 "name": "BaseBdev2", 00:15:23.215 "uuid": "2c118720-c56e-58b2-9ce8-0fe1ca19a44e", 00:15:23.215 "is_configured": true, 00:15:23.215 "data_offset": 256, 00:15:23.215 "data_size": 7936 00:15:23.215 } 00:15:23.215 ] 00:15:23.215 }' 00:15:23.215 23:49:11 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:23.215 23:49:11 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:23.215 23:49:11 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:23.215 23:49:11 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:23.215 23:49:11 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:24.595 23:49:12 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:24.595 23:49:12 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:24.595 23:49:12 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:24.595 23:49:12 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:24.595 23:49:12 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:24.595 23:49:12 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:24.595 23:49:12 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:24.595 23:49:12 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:24.595 23:49:12 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:24.595 23:49:12 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:24.595 23:49:12 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:24.595 23:49:12 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:24.595 "name": "raid_bdev1", 00:15:24.595 "uuid": "1db8de6e-e5f5-48af-8510-2cccfe208f1d", 00:15:24.595 "strip_size_kb": 0, 00:15:24.595 "state": "online", 00:15:24.595 "raid_level": "raid1", 00:15:24.595 "superblock": true, 00:15:24.595 "num_base_bdevs": 2, 00:15:24.595 "num_base_bdevs_discovered": 2, 00:15:24.595 "num_base_bdevs_operational": 2, 00:15:24.595 "process": { 00:15:24.595 "type": "rebuild", 00:15:24.595 "target": "spare", 00:15:24.595 "progress": { 00:15:24.595 "blocks": 5632, 00:15:24.595 "percent": 70 00:15:24.595 } 00:15:24.595 }, 00:15:24.595 "base_bdevs_list": [ 00:15:24.595 { 00:15:24.595 "name": "spare", 00:15:24.595 "uuid": "6f15d6e8-07d1-535a-b452-9aefb81194d0", 00:15:24.595 "is_configured": true, 00:15:24.595 "data_offset": 256, 00:15:24.595 "data_size": 7936 00:15:24.595 }, 00:15:24.595 { 00:15:24.595 "name": "BaseBdev2", 00:15:24.595 "uuid": "2c118720-c56e-58b2-9ce8-0fe1ca19a44e", 00:15:24.595 "is_configured": true, 00:15:24.595 "data_offset": 256, 00:15:24.595 "data_size": 7936 00:15:24.595 } 00:15:24.595 ] 00:15:24.595 }' 00:15:24.595 23:49:12 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:24.595 23:49:12 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:24.596 23:49:12 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:24.596 23:49:12 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:24.596 23:49:12 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:25.165 [2024-11-26 23:49:13.125356] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:15:25.165 [2024-11-26 23:49:13.125506] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:15:25.165 [2024-11-26 23:49:13.125621] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:25.424 23:49:13 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:25.424 23:49:13 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:25.424 23:49:13 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:25.424 23:49:13 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:25.424 23:49:13 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:25.424 23:49:13 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:25.424 23:49:13 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:25.424 23:49:13 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:25.424 23:49:13 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:25.424 23:49:13 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:25.424 23:49:13 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:25.424 23:49:13 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:25.424 "name": "raid_bdev1", 00:15:25.424 "uuid": "1db8de6e-e5f5-48af-8510-2cccfe208f1d", 00:15:25.424 "strip_size_kb": 0, 00:15:25.424 "state": "online", 00:15:25.424 "raid_level": "raid1", 00:15:25.424 "superblock": true, 00:15:25.424 "num_base_bdevs": 2, 00:15:25.424 "num_base_bdevs_discovered": 2, 00:15:25.424 "num_base_bdevs_operational": 2, 00:15:25.424 "base_bdevs_list": [ 00:15:25.424 { 00:15:25.424 "name": "spare", 00:15:25.424 "uuid": "6f15d6e8-07d1-535a-b452-9aefb81194d0", 00:15:25.424 "is_configured": true, 00:15:25.424 "data_offset": 256, 00:15:25.424 "data_size": 7936 00:15:25.424 }, 00:15:25.424 { 00:15:25.424 "name": "BaseBdev2", 00:15:25.424 "uuid": "2c118720-c56e-58b2-9ce8-0fe1ca19a44e", 00:15:25.424 "is_configured": true, 00:15:25.424 "data_offset": 256, 00:15:25.424 "data_size": 7936 00:15:25.424 } 00:15:25.424 ] 00:15:25.424 }' 00:15:25.424 23:49:13 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:25.424 23:49:13 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:15:25.424 23:49:13 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:25.684 23:49:13 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:15:25.684 23:49:13 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@709 -- # break 00:15:25.684 23:49:13 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:15:25.684 23:49:13 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:25.684 23:49:13 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:15:25.684 23:49:13 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=none 00:15:25.684 23:49:13 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:25.684 23:49:13 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:25.684 23:49:13 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:25.684 23:49:13 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:25.684 23:49:13 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:25.684 23:49:13 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:25.684 23:49:13 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:25.684 "name": "raid_bdev1", 00:15:25.684 "uuid": "1db8de6e-e5f5-48af-8510-2cccfe208f1d", 00:15:25.684 "strip_size_kb": 0, 00:15:25.684 "state": "online", 00:15:25.684 "raid_level": "raid1", 00:15:25.684 "superblock": true, 00:15:25.684 "num_base_bdevs": 2, 00:15:25.684 "num_base_bdevs_discovered": 2, 00:15:25.684 "num_base_bdevs_operational": 2, 00:15:25.684 "base_bdevs_list": [ 00:15:25.684 { 00:15:25.684 "name": "spare", 00:15:25.684 "uuid": "6f15d6e8-07d1-535a-b452-9aefb81194d0", 00:15:25.684 "is_configured": true, 00:15:25.684 "data_offset": 256, 00:15:25.684 "data_size": 7936 00:15:25.684 }, 00:15:25.684 { 00:15:25.684 "name": "BaseBdev2", 00:15:25.684 "uuid": "2c118720-c56e-58b2-9ce8-0fe1ca19a44e", 00:15:25.684 "is_configured": true, 00:15:25.684 "data_offset": 256, 00:15:25.684 "data_size": 7936 00:15:25.684 } 00:15:25.684 ] 00:15:25.684 }' 00:15:25.684 23:49:13 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:25.684 23:49:13 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:15:25.684 23:49:13 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:25.684 23:49:13 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:15:25.684 23:49:13 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:15:25.684 23:49:13 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:25.684 23:49:13 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:25.684 23:49:13 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:25.684 23:49:13 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:25.684 23:49:13 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:25.684 23:49:13 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:25.684 23:49:13 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:25.684 23:49:13 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:25.684 23:49:13 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:25.684 23:49:13 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:25.684 23:49:13 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:25.684 23:49:13 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:25.684 23:49:13 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:25.684 23:49:13 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:25.685 23:49:13 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:25.685 "name": "raid_bdev1", 00:15:25.685 "uuid": "1db8de6e-e5f5-48af-8510-2cccfe208f1d", 00:15:25.685 "strip_size_kb": 0, 00:15:25.685 "state": "online", 00:15:25.685 "raid_level": "raid1", 00:15:25.685 "superblock": true, 00:15:25.685 "num_base_bdevs": 2, 00:15:25.685 "num_base_bdevs_discovered": 2, 00:15:25.685 "num_base_bdevs_operational": 2, 00:15:25.685 "base_bdevs_list": [ 00:15:25.685 { 00:15:25.685 "name": "spare", 00:15:25.685 "uuid": "6f15d6e8-07d1-535a-b452-9aefb81194d0", 00:15:25.685 "is_configured": true, 00:15:25.685 "data_offset": 256, 00:15:25.685 "data_size": 7936 00:15:25.685 }, 00:15:25.685 { 00:15:25.685 "name": "BaseBdev2", 00:15:25.685 "uuid": "2c118720-c56e-58b2-9ce8-0fe1ca19a44e", 00:15:25.685 "is_configured": true, 00:15:25.685 "data_offset": 256, 00:15:25.685 "data_size": 7936 00:15:25.685 } 00:15:25.685 ] 00:15:25.685 }' 00:15:25.685 23:49:13 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:25.685 23:49:13 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:26.256 23:49:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:15:26.256 23:49:14 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:26.256 23:49:14 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:26.256 [2024-11-26 23:49:14.158440] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:15:26.256 [2024-11-26 23:49:14.158469] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:15:26.256 [2024-11-26 23:49:14.158552] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:26.256 [2024-11-26 23:49:14.158612] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:26.256 [2024-11-26 23:49:14.158624] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:15:26.256 23:49:14 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:26.256 23:49:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:26.256 23:49:14 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:26.256 23:49:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@720 -- # jq length 00:15:26.256 23:49:14 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:26.256 23:49:14 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:26.256 23:49:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:15:26.256 23:49:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:15:26.256 23:49:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:15:26.256 23:49:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:15:26.256 23:49:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:15:26.256 23:49:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:15:26.256 23:49:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@10 -- # local bdev_list 00:15:26.256 23:49:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:15:26.256 23:49:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@11 -- # local nbd_list 00:15:26.256 23:49:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@12 -- # local i 00:15:26.256 23:49:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:15:26.256 23:49:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:15:26.256 23:49:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:15:26.517 /dev/nbd0 00:15:26.517 23:49:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:15:26.517 23:49:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:15:26.517 23:49:14 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:15:26.517 23:49:14 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@873 -- # local i 00:15:26.517 23:49:14 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:15:26.517 23:49:14 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:15:26.517 23:49:14 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:15:26.517 23:49:14 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@877 -- # break 00:15:26.517 23:49:14 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:15:26.517 23:49:14 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:15:26.517 23:49:14 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:15:26.517 1+0 records in 00:15:26.517 1+0 records out 00:15:26.517 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000439265 s, 9.3 MB/s 00:15:26.517 23:49:14 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:26.517 23:49:14 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@890 -- # size=4096 00:15:26.517 23:49:14 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:26.517 23:49:14 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:15:26.517 23:49:14 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@893 -- # return 0 00:15:26.517 23:49:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:15:26.517 23:49:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:15:26.517 23:49:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:15:26.775 /dev/nbd1 00:15:26.775 23:49:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:15:26.775 23:49:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:15:26.775 23:49:14 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:15:26.775 23:49:14 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@873 -- # local i 00:15:26.775 23:49:14 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:15:26.775 23:49:14 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:15:26.775 23:49:14 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:15:26.775 23:49:14 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@877 -- # break 00:15:26.776 23:49:14 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:15:26.776 23:49:14 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:15:26.776 23:49:14 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:15:26.776 1+0 records in 00:15:26.776 1+0 records out 00:15:26.776 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000399847 s, 10.2 MB/s 00:15:26.776 23:49:14 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:26.776 23:49:14 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@890 -- # size=4096 00:15:26.776 23:49:14 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:26.776 23:49:14 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:15:26.776 23:49:14 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@893 -- # return 0 00:15:26.776 23:49:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:15:26.776 23:49:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:15:26.776 23:49:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@738 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:15:26.776 23:49:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:15:26.776 23:49:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:15:26.776 23:49:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:15:26.776 23:49:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@50 -- # local nbd_list 00:15:26.776 23:49:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@51 -- # local i 00:15:26.776 23:49:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:15:26.776 23:49:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:15:27.035 23:49:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:15:27.035 23:49:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:15:27.035 23:49:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:15:27.035 23:49:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:15:27.035 23:49:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:15:27.035 23:49:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:15:27.035 23:49:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@41 -- # break 00:15:27.035 23:49:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@45 -- # return 0 00:15:27.035 23:49:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:15:27.035 23:49:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:15:27.295 23:49:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:15:27.295 23:49:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:15:27.295 23:49:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:15:27.295 23:49:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:15:27.295 23:49:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:15:27.295 23:49:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:15:27.295 23:49:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@41 -- # break 00:15:27.295 23:49:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@45 -- # return 0 00:15:27.295 23:49:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@743 -- # '[' true = true ']' 00:15:27.295 23:49:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@745 -- # rpc_cmd bdev_passthru_delete spare 00:15:27.295 23:49:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:27.295 23:49:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:27.295 23:49:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:27.295 23:49:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@746 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:15:27.295 23:49:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:27.295 23:49:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:27.295 [2024-11-26 23:49:15.253599] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:15:27.295 [2024-11-26 23:49:15.253654] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:27.295 [2024-11-26 23:49:15.253673] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:15:27.295 [2024-11-26 23:49:15.253686] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:27.295 [2024-11-26 23:49:15.255649] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:27.295 [2024-11-26 23:49:15.255689] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:15:27.295 [2024-11-26 23:49:15.255741] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:15:27.295 [2024-11-26 23:49:15.255791] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:15:27.295 [2024-11-26 23:49:15.255915] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:15:27.295 spare 00:15:27.295 23:49:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:27.295 23:49:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@747 -- # rpc_cmd bdev_wait_for_examine 00:15:27.295 23:49:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:27.295 23:49:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:27.295 [2024-11-26 23:49:15.355801] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001580 00:15:27.295 [2024-11-26 23:49:15.355826] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:15:27.296 [2024-11-26 23:49:15.355904] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0001bb1b0 00:15:27.296 [2024-11-26 23:49:15.355995] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001580 00:15:27.296 [2024-11-26 23:49:15.356005] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001580 00:15:27.296 [2024-11-26 23:49:15.356085] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:27.296 23:49:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:27.296 23:49:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@749 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:15:27.296 23:49:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:27.296 23:49:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:27.296 23:49:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:27.296 23:49:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:27.296 23:49:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:27.296 23:49:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:27.296 23:49:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:27.296 23:49:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:27.296 23:49:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:27.296 23:49:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:27.296 23:49:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:27.296 23:49:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:27.296 23:49:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:27.296 23:49:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:27.296 23:49:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:27.296 "name": "raid_bdev1", 00:15:27.296 "uuid": "1db8de6e-e5f5-48af-8510-2cccfe208f1d", 00:15:27.296 "strip_size_kb": 0, 00:15:27.296 "state": "online", 00:15:27.296 "raid_level": "raid1", 00:15:27.296 "superblock": true, 00:15:27.296 "num_base_bdevs": 2, 00:15:27.296 "num_base_bdevs_discovered": 2, 00:15:27.296 "num_base_bdevs_operational": 2, 00:15:27.296 "base_bdevs_list": [ 00:15:27.296 { 00:15:27.296 "name": "spare", 00:15:27.296 "uuid": "6f15d6e8-07d1-535a-b452-9aefb81194d0", 00:15:27.296 "is_configured": true, 00:15:27.296 "data_offset": 256, 00:15:27.296 "data_size": 7936 00:15:27.296 }, 00:15:27.296 { 00:15:27.296 "name": "BaseBdev2", 00:15:27.296 "uuid": "2c118720-c56e-58b2-9ce8-0fe1ca19a44e", 00:15:27.296 "is_configured": true, 00:15:27.296 "data_offset": 256, 00:15:27.296 "data_size": 7936 00:15:27.296 } 00:15:27.296 ] 00:15:27.296 }' 00:15:27.296 23:49:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:27.296 23:49:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:27.866 23:49:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@750 -- # verify_raid_bdev_process raid_bdev1 none none 00:15:27.866 23:49:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:27.866 23:49:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:15:27.866 23:49:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=none 00:15:27.866 23:49:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:27.866 23:49:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:27.866 23:49:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:27.866 23:49:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:27.866 23:49:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:27.866 23:49:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:27.866 23:49:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:27.866 "name": "raid_bdev1", 00:15:27.866 "uuid": "1db8de6e-e5f5-48af-8510-2cccfe208f1d", 00:15:27.866 "strip_size_kb": 0, 00:15:27.866 "state": "online", 00:15:27.866 "raid_level": "raid1", 00:15:27.866 "superblock": true, 00:15:27.866 "num_base_bdevs": 2, 00:15:27.866 "num_base_bdevs_discovered": 2, 00:15:27.866 "num_base_bdevs_operational": 2, 00:15:27.866 "base_bdevs_list": [ 00:15:27.866 { 00:15:27.866 "name": "spare", 00:15:27.866 "uuid": "6f15d6e8-07d1-535a-b452-9aefb81194d0", 00:15:27.866 "is_configured": true, 00:15:27.866 "data_offset": 256, 00:15:27.866 "data_size": 7936 00:15:27.866 }, 00:15:27.866 { 00:15:27.866 "name": "BaseBdev2", 00:15:27.866 "uuid": "2c118720-c56e-58b2-9ce8-0fe1ca19a44e", 00:15:27.866 "is_configured": true, 00:15:27.866 "data_offset": 256, 00:15:27.866 "data_size": 7936 00:15:27.866 } 00:15:27.866 ] 00:15:27.866 }' 00:15:27.866 23:49:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:27.866 23:49:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:15:27.866 23:49:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:27.866 23:49:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:15:27.866 23:49:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@751 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:27.866 23:49:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@751 -- # jq -r '.[].base_bdevs_list[0].name' 00:15:27.866 23:49:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:27.866 23:49:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:27.866 23:49:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:28.126 23:49:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@751 -- # [[ spare == \s\p\a\r\e ]] 00:15:28.127 23:49:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@754 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:15:28.127 23:49:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:28.127 23:49:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:28.127 [2024-11-26 23:49:16.004468] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:28.127 23:49:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:28.127 23:49:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@755 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:15:28.127 23:49:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:28.127 23:49:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:28.127 23:49:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:28.127 23:49:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:28.127 23:49:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:15:28.127 23:49:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:28.127 23:49:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:28.127 23:49:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:28.127 23:49:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:28.127 23:49:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:28.127 23:49:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:28.127 23:49:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:28.127 23:49:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:28.127 23:49:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:28.127 23:49:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:28.127 "name": "raid_bdev1", 00:15:28.127 "uuid": "1db8de6e-e5f5-48af-8510-2cccfe208f1d", 00:15:28.127 "strip_size_kb": 0, 00:15:28.127 "state": "online", 00:15:28.127 "raid_level": "raid1", 00:15:28.127 "superblock": true, 00:15:28.127 "num_base_bdevs": 2, 00:15:28.127 "num_base_bdevs_discovered": 1, 00:15:28.127 "num_base_bdevs_operational": 1, 00:15:28.127 "base_bdevs_list": [ 00:15:28.127 { 00:15:28.127 "name": null, 00:15:28.127 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:28.127 "is_configured": false, 00:15:28.127 "data_offset": 0, 00:15:28.127 "data_size": 7936 00:15:28.127 }, 00:15:28.127 { 00:15:28.127 "name": "BaseBdev2", 00:15:28.127 "uuid": "2c118720-c56e-58b2-9ce8-0fe1ca19a44e", 00:15:28.127 "is_configured": true, 00:15:28.127 "data_offset": 256, 00:15:28.127 "data_size": 7936 00:15:28.127 } 00:15:28.127 ] 00:15:28.127 }' 00:15:28.127 23:49:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:28.127 23:49:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:28.387 23:49:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@756 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:15:28.387 23:49:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:28.387 23:49:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:28.387 [2024-11-26 23:49:16.427748] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:15:28.387 [2024-11-26 23:49:16.427895] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:15:28.387 [2024-11-26 23:49:16.427909] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:15:28.388 [2024-11-26 23:49:16.427960] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:15:28.388 [2024-11-26 23:49:16.430266] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0001bb280 00:15:28.388 [2024-11-26 23:49:16.432034] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:15:28.388 23:49:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:28.388 23:49:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@757 -- # sleep 1 00:15:29.329 23:49:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@758 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:29.329 23:49:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:29.329 23:49:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:29.329 23:49:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:29.329 23:49:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:29.329 23:49:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:29.329 23:49:17 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:29.329 23:49:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:29.329 23:49:17 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:29.590 23:49:17 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:29.590 23:49:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:29.590 "name": "raid_bdev1", 00:15:29.590 "uuid": "1db8de6e-e5f5-48af-8510-2cccfe208f1d", 00:15:29.590 "strip_size_kb": 0, 00:15:29.590 "state": "online", 00:15:29.590 "raid_level": "raid1", 00:15:29.590 "superblock": true, 00:15:29.590 "num_base_bdevs": 2, 00:15:29.590 "num_base_bdevs_discovered": 2, 00:15:29.590 "num_base_bdevs_operational": 2, 00:15:29.590 "process": { 00:15:29.590 "type": "rebuild", 00:15:29.590 "target": "spare", 00:15:29.590 "progress": { 00:15:29.590 "blocks": 2560, 00:15:29.590 "percent": 32 00:15:29.590 } 00:15:29.590 }, 00:15:29.590 "base_bdevs_list": [ 00:15:29.590 { 00:15:29.590 "name": "spare", 00:15:29.590 "uuid": "6f15d6e8-07d1-535a-b452-9aefb81194d0", 00:15:29.590 "is_configured": true, 00:15:29.590 "data_offset": 256, 00:15:29.590 "data_size": 7936 00:15:29.590 }, 00:15:29.590 { 00:15:29.590 "name": "BaseBdev2", 00:15:29.590 "uuid": "2c118720-c56e-58b2-9ce8-0fe1ca19a44e", 00:15:29.590 "is_configured": true, 00:15:29.590 "data_offset": 256, 00:15:29.590 "data_size": 7936 00:15:29.590 } 00:15:29.590 ] 00:15:29.590 }' 00:15:29.590 23:49:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:29.590 23:49:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:29.590 23:49:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:29.590 23:49:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:29.590 23:49:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@761 -- # rpc_cmd bdev_passthru_delete spare 00:15:29.590 23:49:17 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:29.590 23:49:17 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:29.590 [2024-11-26 23:49:17.598843] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:29.590 [2024-11-26 23:49:17.635964] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:15:29.590 [2024-11-26 23:49:17.636017] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:29.590 [2024-11-26 23:49:17.636034] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:29.590 [2024-11-26 23:49:17.636041] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:15:29.590 23:49:17 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:29.590 23:49:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@762 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:15:29.590 23:49:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:29.590 23:49:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:29.590 23:49:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:29.590 23:49:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:29.590 23:49:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:15:29.590 23:49:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:29.590 23:49:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:29.590 23:49:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:29.590 23:49:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:29.590 23:49:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:29.590 23:49:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:29.590 23:49:17 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:29.590 23:49:17 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:29.590 23:49:17 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:29.590 23:49:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:29.590 "name": "raid_bdev1", 00:15:29.590 "uuid": "1db8de6e-e5f5-48af-8510-2cccfe208f1d", 00:15:29.590 "strip_size_kb": 0, 00:15:29.590 "state": "online", 00:15:29.590 "raid_level": "raid1", 00:15:29.590 "superblock": true, 00:15:29.590 "num_base_bdevs": 2, 00:15:29.590 "num_base_bdevs_discovered": 1, 00:15:29.590 "num_base_bdevs_operational": 1, 00:15:29.590 "base_bdevs_list": [ 00:15:29.590 { 00:15:29.590 "name": null, 00:15:29.590 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:29.590 "is_configured": false, 00:15:29.590 "data_offset": 0, 00:15:29.590 "data_size": 7936 00:15:29.590 }, 00:15:29.590 { 00:15:29.590 "name": "BaseBdev2", 00:15:29.590 "uuid": "2c118720-c56e-58b2-9ce8-0fe1ca19a44e", 00:15:29.590 "is_configured": true, 00:15:29.590 "data_offset": 256, 00:15:29.590 "data_size": 7936 00:15:29.590 } 00:15:29.590 ] 00:15:29.590 }' 00:15:29.590 23:49:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:29.590 23:49:17 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:30.161 23:49:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@763 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:15:30.161 23:49:18 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:30.161 23:49:18 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:30.161 [2024-11-26 23:49:18.117961] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:15:30.161 [2024-11-26 23:49:18.118020] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:30.161 [2024-11-26 23:49:18.118047] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009f80 00:15:30.161 [2024-11-26 23:49:18.118056] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:30.161 [2024-11-26 23:49:18.118267] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:30.161 [2024-11-26 23:49:18.118288] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:15:30.161 [2024-11-26 23:49:18.118352] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:15:30.161 [2024-11-26 23:49:18.118363] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:15:30.161 [2024-11-26 23:49:18.118377] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:15:30.161 [2024-11-26 23:49:18.118396] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:15:30.161 [2024-11-26 23:49:18.120555] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0001bb350 00:15:30.161 [2024-11-26 23:49:18.122464] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:15:30.161 spare 00:15:30.161 23:49:18 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:30.161 23:49:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@764 -- # sleep 1 00:15:31.101 23:49:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@765 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:31.101 23:49:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:31.101 23:49:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:31.101 23:49:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:31.101 23:49:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:31.101 23:49:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:31.101 23:49:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:31.101 23:49:19 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:31.101 23:49:19 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:31.101 23:49:19 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:31.101 23:49:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:31.101 "name": "raid_bdev1", 00:15:31.101 "uuid": "1db8de6e-e5f5-48af-8510-2cccfe208f1d", 00:15:31.101 "strip_size_kb": 0, 00:15:31.101 "state": "online", 00:15:31.101 "raid_level": "raid1", 00:15:31.101 "superblock": true, 00:15:31.101 "num_base_bdevs": 2, 00:15:31.101 "num_base_bdevs_discovered": 2, 00:15:31.101 "num_base_bdevs_operational": 2, 00:15:31.101 "process": { 00:15:31.101 "type": "rebuild", 00:15:31.101 "target": "spare", 00:15:31.101 "progress": { 00:15:31.101 "blocks": 2560, 00:15:31.101 "percent": 32 00:15:31.101 } 00:15:31.101 }, 00:15:31.101 "base_bdevs_list": [ 00:15:31.101 { 00:15:31.101 "name": "spare", 00:15:31.101 "uuid": "6f15d6e8-07d1-535a-b452-9aefb81194d0", 00:15:31.101 "is_configured": true, 00:15:31.101 "data_offset": 256, 00:15:31.101 "data_size": 7936 00:15:31.101 }, 00:15:31.101 { 00:15:31.101 "name": "BaseBdev2", 00:15:31.101 "uuid": "2c118720-c56e-58b2-9ce8-0fe1ca19a44e", 00:15:31.101 "is_configured": true, 00:15:31.101 "data_offset": 256, 00:15:31.101 "data_size": 7936 00:15:31.101 } 00:15:31.101 ] 00:15:31.101 }' 00:15:31.101 23:49:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:31.360 23:49:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:31.361 23:49:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:31.361 23:49:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:31.361 23:49:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@768 -- # rpc_cmd bdev_passthru_delete spare 00:15:31.361 23:49:19 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:31.361 23:49:19 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:31.361 [2024-11-26 23:49:19.289164] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:31.361 [2024-11-26 23:49:19.326401] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:15:31.361 [2024-11-26 23:49:19.326459] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:31.361 [2024-11-26 23:49:19.326472] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:31.361 [2024-11-26 23:49:19.326480] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:15:31.361 23:49:19 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:31.361 23:49:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@769 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:15:31.361 23:49:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:31.361 23:49:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:31.361 23:49:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:31.361 23:49:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:31.361 23:49:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:15:31.361 23:49:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:31.361 23:49:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:31.361 23:49:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:31.361 23:49:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:31.361 23:49:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:31.361 23:49:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:31.361 23:49:19 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:31.361 23:49:19 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:31.361 23:49:19 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:31.361 23:49:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:31.361 "name": "raid_bdev1", 00:15:31.361 "uuid": "1db8de6e-e5f5-48af-8510-2cccfe208f1d", 00:15:31.361 "strip_size_kb": 0, 00:15:31.361 "state": "online", 00:15:31.361 "raid_level": "raid1", 00:15:31.361 "superblock": true, 00:15:31.361 "num_base_bdevs": 2, 00:15:31.361 "num_base_bdevs_discovered": 1, 00:15:31.361 "num_base_bdevs_operational": 1, 00:15:31.361 "base_bdevs_list": [ 00:15:31.361 { 00:15:31.361 "name": null, 00:15:31.361 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:31.361 "is_configured": false, 00:15:31.361 "data_offset": 0, 00:15:31.361 "data_size": 7936 00:15:31.361 }, 00:15:31.361 { 00:15:31.361 "name": "BaseBdev2", 00:15:31.361 "uuid": "2c118720-c56e-58b2-9ce8-0fe1ca19a44e", 00:15:31.361 "is_configured": true, 00:15:31.361 "data_offset": 256, 00:15:31.361 "data_size": 7936 00:15:31.361 } 00:15:31.361 ] 00:15:31.361 }' 00:15:31.361 23:49:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:31.361 23:49:19 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:31.620 23:49:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@770 -- # verify_raid_bdev_process raid_bdev1 none none 00:15:31.620 23:49:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:31.620 23:49:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:15:31.620 23:49:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=none 00:15:31.620 23:49:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:31.620 23:49:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:31.620 23:49:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:31.620 23:49:19 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:31.620 23:49:19 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:31.880 23:49:19 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:31.880 23:49:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:31.880 "name": "raid_bdev1", 00:15:31.880 "uuid": "1db8de6e-e5f5-48af-8510-2cccfe208f1d", 00:15:31.880 "strip_size_kb": 0, 00:15:31.880 "state": "online", 00:15:31.880 "raid_level": "raid1", 00:15:31.880 "superblock": true, 00:15:31.880 "num_base_bdevs": 2, 00:15:31.880 "num_base_bdevs_discovered": 1, 00:15:31.880 "num_base_bdevs_operational": 1, 00:15:31.880 "base_bdevs_list": [ 00:15:31.880 { 00:15:31.880 "name": null, 00:15:31.880 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:31.880 "is_configured": false, 00:15:31.880 "data_offset": 0, 00:15:31.880 "data_size": 7936 00:15:31.880 }, 00:15:31.880 { 00:15:31.881 "name": "BaseBdev2", 00:15:31.881 "uuid": "2c118720-c56e-58b2-9ce8-0fe1ca19a44e", 00:15:31.881 "is_configured": true, 00:15:31.881 "data_offset": 256, 00:15:31.881 "data_size": 7936 00:15:31.881 } 00:15:31.881 ] 00:15:31.881 }' 00:15:31.881 23:49:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:31.881 23:49:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:15:31.881 23:49:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:31.881 23:49:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:15:31.881 23:49:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@773 -- # rpc_cmd bdev_passthru_delete BaseBdev1 00:15:31.881 23:49:19 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:31.881 23:49:19 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:31.881 23:49:19 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:31.881 23:49:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@774 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:15:31.881 23:49:19 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:31.881 23:49:19 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:31.881 [2024-11-26 23:49:19.908567] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:15:31.881 [2024-11-26 23:49:19.908619] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:31.881 [2024-11-26 23:49:19.908639] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a580 00:15:31.881 [2024-11-26 23:49:19.908650] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:31.881 [2024-11-26 23:49:19.908840] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:31.881 [2024-11-26 23:49:19.908862] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:15:31.881 [2024-11-26 23:49:19.908922] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:15:31.881 [2024-11-26 23:49:19.908956] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:15:31.881 [2024-11-26 23:49:19.908964] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:15:31.881 [2024-11-26 23:49:19.908975] bdev_raid.c:3894:raid_bdev_examine_done: *ERROR*: Failed to examine bdev BaseBdev1: Invalid argument 00:15:31.881 BaseBdev1 00:15:31.881 23:49:19 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:31.881 23:49:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@775 -- # sleep 1 00:15:32.819 23:49:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@776 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:15:32.819 23:49:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:32.819 23:49:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:32.819 23:49:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:32.819 23:49:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:32.819 23:49:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:15:32.819 23:49:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:32.819 23:49:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:32.819 23:49:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:32.819 23:49:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:32.819 23:49:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:32.819 23:49:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:32.819 23:49:20 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:32.819 23:49:20 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:32.819 23:49:20 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:33.080 23:49:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:33.080 "name": "raid_bdev1", 00:15:33.080 "uuid": "1db8de6e-e5f5-48af-8510-2cccfe208f1d", 00:15:33.080 "strip_size_kb": 0, 00:15:33.080 "state": "online", 00:15:33.080 "raid_level": "raid1", 00:15:33.080 "superblock": true, 00:15:33.080 "num_base_bdevs": 2, 00:15:33.080 "num_base_bdevs_discovered": 1, 00:15:33.080 "num_base_bdevs_operational": 1, 00:15:33.080 "base_bdevs_list": [ 00:15:33.080 { 00:15:33.080 "name": null, 00:15:33.080 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:33.080 "is_configured": false, 00:15:33.080 "data_offset": 0, 00:15:33.080 "data_size": 7936 00:15:33.080 }, 00:15:33.080 { 00:15:33.080 "name": "BaseBdev2", 00:15:33.080 "uuid": "2c118720-c56e-58b2-9ce8-0fe1ca19a44e", 00:15:33.080 "is_configured": true, 00:15:33.080 "data_offset": 256, 00:15:33.080 "data_size": 7936 00:15:33.080 } 00:15:33.080 ] 00:15:33.080 }' 00:15:33.080 23:49:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:33.080 23:49:20 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:33.339 23:49:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@777 -- # verify_raid_bdev_process raid_bdev1 none none 00:15:33.339 23:49:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:33.339 23:49:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:15:33.339 23:49:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=none 00:15:33.339 23:49:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:33.339 23:49:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:33.340 23:49:21 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:33.340 23:49:21 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:33.340 23:49:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:33.340 23:49:21 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:33.340 23:49:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:33.340 "name": "raid_bdev1", 00:15:33.340 "uuid": "1db8de6e-e5f5-48af-8510-2cccfe208f1d", 00:15:33.340 "strip_size_kb": 0, 00:15:33.340 "state": "online", 00:15:33.340 "raid_level": "raid1", 00:15:33.340 "superblock": true, 00:15:33.340 "num_base_bdevs": 2, 00:15:33.340 "num_base_bdevs_discovered": 1, 00:15:33.340 "num_base_bdevs_operational": 1, 00:15:33.340 "base_bdevs_list": [ 00:15:33.340 { 00:15:33.340 "name": null, 00:15:33.340 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:33.340 "is_configured": false, 00:15:33.340 "data_offset": 0, 00:15:33.340 "data_size": 7936 00:15:33.340 }, 00:15:33.340 { 00:15:33.340 "name": "BaseBdev2", 00:15:33.340 "uuid": "2c118720-c56e-58b2-9ce8-0fe1ca19a44e", 00:15:33.340 "is_configured": true, 00:15:33.340 "data_offset": 256, 00:15:33.340 "data_size": 7936 00:15:33.340 } 00:15:33.340 ] 00:15:33.340 }' 00:15:33.340 23:49:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:33.600 23:49:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:15:33.600 23:49:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:33.600 23:49:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:15:33.600 23:49:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@778 -- # NOT rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:15:33.600 23:49:21 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@652 -- # local es=0 00:15:33.600 23:49:21 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:15:33.600 23:49:21 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:15:33.600 23:49:21 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:15:33.600 23:49:21 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:15:33.600 23:49:21 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:15:33.600 23:49:21 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:15:33.600 23:49:21 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:33.600 23:49:21 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:33.600 [2024-11-26 23:49:21.525806] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:15:33.600 [2024-11-26 23:49:21.525956] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:15:33.600 [2024-11-26 23:49:21.525967] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:15:33.600 request: 00:15:33.600 { 00:15:33.600 "base_bdev": "BaseBdev1", 00:15:33.600 "raid_bdev": "raid_bdev1", 00:15:33.600 "method": "bdev_raid_add_base_bdev", 00:15:33.600 "req_id": 1 00:15:33.600 } 00:15:33.600 Got JSON-RPC error response 00:15:33.600 response: 00:15:33.600 { 00:15:33.600 "code": -22, 00:15:33.600 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:15:33.600 } 00:15:33.600 23:49:21 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:15:33.600 23:49:21 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@655 -- # es=1 00:15:33.600 23:49:21 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:15:33.600 23:49:21 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:15:33.600 23:49:21 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:15:33.600 23:49:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@779 -- # sleep 1 00:15:34.541 23:49:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@780 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:15:34.541 23:49:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:34.541 23:49:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:34.541 23:49:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:34.541 23:49:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:34.541 23:49:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:15:34.541 23:49:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:34.541 23:49:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:34.541 23:49:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:34.541 23:49:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:34.541 23:49:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:34.541 23:49:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:34.541 23:49:22 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:34.541 23:49:22 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:34.541 23:49:22 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:34.541 23:49:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:34.541 "name": "raid_bdev1", 00:15:34.541 "uuid": "1db8de6e-e5f5-48af-8510-2cccfe208f1d", 00:15:34.541 "strip_size_kb": 0, 00:15:34.541 "state": "online", 00:15:34.541 "raid_level": "raid1", 00:15:34.541 "superblock": true, 00:15:34.541 "num_base_bdevs": 2, 00:15:34.541 "num_base_bdevs_discovered": 1, 00:15:34.541 "num_base_bdevs_operational": 1, 00:15:34.541 "base_bdevs_list": [ 00:15:34.541 { 00:15:34.541 "name": null, 00:15:34.541 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:34.541 "is_configured": false, 00:15:34.541 "data_offset": 0, 00:15:34.541 "data_size": 7936 00:15:34.541 }, 00:15:34.541 { 00:15:34.541 "name": "BaseBdev2", 00:15:34.541 "uuid": "2c118720-c56e-58b2-9ce8-0fe1ca19a44e", 00:15:34.541 "is_configured": true, 00:15:34.541 "data_offset": 256, 00:15:34.541 "data_size": 7936 00:15:34.541 } 00:15:34.541 ] 00:15:34.541 }' 00:15:34.541 23:49:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:34.541 23:49:22 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:35.112 23:49:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@781 -- # verify_raid_bdev_process raid_bdev1 none none 00:15:35.112 23:49:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:35.112 23:49:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:15:35.112 23:49:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=none 00:15:35.112 23:49:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:35.112 23:49:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:35.112 23:49:23 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:35.112 23:49:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:35.112 23:49:23 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:35.112 23:49:23 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:35.112 23:49:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:35.112 "name": "raid_bdev1", 00:15:35.112 "uuid": "1db8de6e-e5f5-48af-8510-2cccfe208f1d", 00:15:35.112 "strip_size_kb": 0, 00:15:35.112 "state": "online", 00:15:35.112 "raid_level": "raid1", 00:15:35.112 "superblock": true, 00:15:35.112 "num_base_bdevs": 2, 00:15:35.112 "num_base_bdevs_discovered": 1, 00:15:35.112 "num_base_bdevs_operational": 1, 00:15:35.112 "base_bdevs_list": [ 00:15:35.112 { 00:15:35.112 "name": null, 00:15:35.112 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:35.112 "is_configured": false, 00:15:35.112 "data_offset": 0, 00:15:35.112 "data_size": 7936 00:15:35.112 }, 00:15:35.112 { 00:15:35.112 "name": "BaseBdev2", 00:15:35.112 "uuid": "2c118720-c56e-58b2-9ce8-0fe1ca19a44e", 00:15:35.112 "is_configured": true, 00:15:35.112 "data_offset": 256, 00:15:35.112 "data_size": 7936 00:15:35.112 } 00:15:35.112 ] 00:15:35.112 }' 00:15:35.112 23:49:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:35.112 23:49:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:15:35.112 23:49:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:35.112 23:49:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:15:35.112 23:49:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@784 -- # killprocess 97742 00:15:35.112 23:49:23 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@954 -- # '[' -z 97742 ']' 00:15:35.112 23:49:23 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@958 -- # kill -0 97742 00:15:35.113 23:49:23 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@959 -- # uname 00:15:35.113 23:49:23 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:15:35.113 23:49:23 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 97742 00:15:35.113 23:49:23 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:15:35.113 23:49:23 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:15:35.113 killing process with pid 97742 00:15:35.113 23:49:23 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@972 -- # echo 'killing process with pid 97742' 00:15:35.113 23:49:23 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@973 -- # kill 97742 00:15:35.113 Received shutdown signal, test time was about 60.000000 seconds 00:15:35.113 00:15:35.113 Latency(us) 00:15:35.113 [2024-11-26T23:49:23.245Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:15:35.113 [2024-11-26T23:49:23.245Z] =================================================================================================================== 00:15:35.113 [2024-11-26T23:49:23.245Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:15:35.113 [2024-11-26 23:49:23.207957] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:15:35.113 [2024-11-26 23:49:23.208076] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:35.113 [2024-11-26 23:49:23.208131] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:35.113 [2024-11-26 23:49:23.208141] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state offline 00:15:35.113 23:49:23 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@978 -- # wait 97742 00:15:35.113 [2024-11-26 23:49:23.240657] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:15:35.373 23:49:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@786 -- # return 0 00:15:35.373 00:15:35.373 real 0m18.302s 00:15:35.373 user 0m24.256s 00:15:35.373 sys 0m2.631s 00:15:35.373 23:49:23 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@1130 -- # xtrace_disable 00:15:35.373 23:49:23 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:35.373 ************************************ 00:15:35.373 END TEST raid_rebuild_test_sb_md_separate 00:15:35.373 ************************************ 00:15:35.373 23:49:23 bdev_raid -- bdev/bdev_raid.sh@1010 -- # base_malloc_params='-m 32 -i' 00:15:35.373 23:49:23 bdev_raid -- bdev/bdev_raid.sh@1011 -- # run_test raid_state_function_test_sb_md_interleaved raid_state_function_test raid1 2 true 00:15:35.633 23:49:23 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:15:35.633 23:49:23 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:15:35.633 23:49:23 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:15:35.633 ************************************ 00:15:35.633 START TEST raid_state_function_test_sb_md_interleaved 00:15:35.633 ************************************ 00:15:35.633 23:49:23 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@1129 -- # raid_state_function_test raid1 2 true 00:15:35.633 23:49:23 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@205 -- # local raid_level=raid1 00:15:35.633 23:49:23 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=2 00:15:35.633 23:49:23 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:15:35.633 23:49:23 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:15:35.633 23:49:23 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:15:35.633 23:49:23 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:15:35.633 23:49:23 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:15:35.633 23:49:23 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:15:35.633 23:49:23 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:15:35.633 23:49:23 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:15:35.633 23:49:23 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:15:35.633 23:49:23 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:15:35.633 23:49:23 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:15:35.633 23:49:23 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:15:35.633 23:49:23 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:15:35.633 23:49:23 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@211 -- # local strip_size 00:15:35.633 23:49:23 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:15:35.633 23:49:23 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:15:35.633 23:49:23 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@215 -- # '[' raid1 '!=' raid1 ']' 00:15:35.634 23:49:23 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@219 -- # strip_size=0 00:15:35.634 23:49:23 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:15:35.634 23:49:23 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:15:35.634 23:49:23 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@229 -- # raid_pid=98416 00:15:35.634 23:49:23 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:15:35.634 Process raid pid: 98416 00:15:35.634 23:49:23 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 98416' 00:15:35.634 23:49:23 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@231 -- # waitforlisten 98416 00:15:35.634 23:49:23 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@835 -- # '[' -z 98416 ']' 00:15:35.634 23:49:23 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:15:35.634 23:49:23 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@840 -- # local max_retries=100 00:15:35.634 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:15:35.634 23:49:23 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:15:35.634 23:49:23 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@844 -- # xtrace_disable 00:15:35.634 23:49:23 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:35.634 [2024-11-26 23:49:23.624685] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:15:35.634 [2024-11-26 23:49:23.624815] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:15:35.894 [2024-11-26 23:49:23.782779] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:15:35.894 [2024-11-26 23:49:23.808616] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:15:35.894 [2024-11-26 23:49:23.851703] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:15:35.894 [2024-11-26 23:49:23.851739] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:15:36.463 23:49:24 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:15:36.463 23:49:24 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@868 -- # return 0 00:15:36.463 23:49:24 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:15:36.463 23:49:24 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:36.463 23:49:24 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:36.463 [2024-11-26 23:49:24.454300] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:15:36.463 [2024-11-26 23:49:24.454366] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:15:36.463 [2024-11-26 23:49:24.454398] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:15:36.463 [2024-11-26 23:49:24.454409] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:15:36.463 23:49:24 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:36.463 23:49:24 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:15:36.463 23:49:24 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:36.463 23:49:24 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:36.463 23:49:24 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:36.463 23:49:24 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:36.463 23:49:24 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:36.463 23:49:24 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:36.463 23:49:24 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:36.463 23:49:24 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:36.463 23:49:24 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:36.463 23:49:24 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:36.463 23:49:24 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:36.463 23:49:24 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:36.463 23:49:24 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:36.463 23:49:24 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:36.463 23:49:24 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:36.463 "name": "Existed_Raid", 00:15:36.463 "uuid": "7c93dc67-7d07-41bb-856e-8020d8e9ae4b", 00:15:36.463 "strip_size_kb": 0, 00:15:36.463 "state": "configuring", 00:15:36.463 "raid_level": "raid1", 00:15:36.463 "superblock": true, 00:15:36.463 "num_base_bdevs": 2, 00:15:36.463 "num_base_bdevs_discovered": 0, 00:15:36.463 "num_base_bdevs_operational": 2, 00:15:36.463 "base_bdevs_list": [ 00:15:36.463 { 00:15:36.463 "name": "BaseBdev1", 00:15:36.463 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:36.463 "is_configured": false, 00:15:36.463 "data_offset": 0, 00:15:36.463 "data_size": 0 00:15:36.463 }, 00:15:36.463 { 00:15:36.463 "name": "BaseBdev2", 00:15:36.463 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:36.463 "is_configured": false, 00:15:36.463 "data_offset": 0, 00:15:36.463 "data_size": 0 00:15:36.463 } 00:15:36.463 ] 00:15:36.463 }' 00:15:36.463 23:49:24 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:36.463 23:49:24 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:37.033 23:49:24 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:15:37.033 23:49:24 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:37.033 23:49:24 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:37.033 [2024-11-26 23:49:24.937452] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:15:37.033 [2024-11-26 23:49:24.937492] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:15:37.033 23:49:24 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:37.033 23:49:24 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:15:37.033 23:49:24 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:37.033 23:49:24 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:37.033 [2024-11-26 23:49:24.949464] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:15:37.033 [2024-11-26 23:49:24.949502] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:15:37.033 [2024-11-26 23:49:24.949538] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:15:37.033 [2024-11-26 23:49:24.949561] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:15:37.033 23:49:24 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:37.033 23:49:24 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -i -b BaseBdev1 00:15:37.033 23:49:24 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:37.033 23:49:24 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:37.033 [2024-11-26 23:49:24.970168] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:15:37.033 BaseBdev1 00:15:37.033 23:49:24 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:37.033 23:49:24 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:15:37.033 23:49:24 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:15:37.033 23:49:24 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:15:37.033 23:49:24 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@905 -- # local i 00:15:37.033 23:49:24 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:15:37.033 23:49:24 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:15:37.033 23:49:24 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:15:37.033 23:49:24 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:37.033 23:49:24 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:37.033 23:49:24 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:37.033 23:49:24 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:15:37.033 23:49:24 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:37.033 23:49:24 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:37.033 [ 00:15:37.033 { 00:15:37.033 "name": "BaseBdev1", 00:15:37.033 "aliases": [ 00:15:37.033 "f38cf700-dd54-4e24-bdfe-54bd2a144e23" 00:15:37.033 ], 00:15:37.033 "product_name": "Malloc disk", 00:15:37.033 "block_size": 4128, 00:15:37.033 "num_blocks": 8192, 00:15:37.033 "uuid": "f38cf700-dd54-4e24-bdfe-54bd2a144e23", 00:15:37.033 "md_size": 32, 00:15:37.033 "md_interleave": true, 00:15:37.033 "dif_type": 0, 00:15:37.033 "assigned_rate_limits": { 00:15:37.033 "rw_ios_per_sec": 0, 00:15:37.033 "rw_mbytes_per_sec": 0, 00:15:37.033 "r_mbytes_per_sec": 0, 00:15:37.033 "w_mbytes_per_sec": 0 00:15:37.033 }, 00:15:37.033 "claimed": true, 00:15:37.033 "claim_type": "exclusive_write", 00:15:37.033 "zoned": false, 00:15:37.033 "supported_io_types": { 00:15:37.033 "read": true, 00:15:37.033 "write": true, 00:15:37.033 "unmap": true, 00:15:37.033 "flush": true, 00:15:37.033 "reset": true, 00:15:37.033 "nvme_admin": false, 00:15:37.033 "nvme_io": false, 00:15:37.033 "nvme_io_md": false, 00:15:37.033 "write_zeroes": true, 00:15:37.033 "zcopy": true, 00:15:37.033 "get_zone_info": false, 00:15:37.033 "zone_management": false, 00:15:37.033 "zone_append": false, 00:15:37.033 "compare": false, 00:15:37.033 "compare_and_write": false, 00:15:37.033 "abort": true, 00:15:37.033 "seek_hole": false, 00:15:37.033 "seek_data": false, 00:15:37.033 "copy": true, 00:15:37.033 "nvme_iov_md": false 00:15:37.033 }, 00:15:37.033 "memory_domains": [ 00:15:37.033 { 00:15:37.033 "dma_device_id": "system", 00:15:37.033 "dma_device_type": 1 00:15:37.033 }, 00:15:37.033 { 00:15:37.033 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:37.033 "dma_device_type": 2 00:15:37.033 } 00:15:37.033 ], 00:15:37.033 "driver_specific": {} 00:15:37.033 } 00:15:37.033 ] 00:15:37.033 23:49:25 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:37.033 23:49:25 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@911 -- # return 0 00:15:37.034 23:49:25 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:15:37.034 23:49:25 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:37.034 23:49:25 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:37.034 23:49:25 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:37.034 23:49:25 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:37.034 23:49:25 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:37.034 23:49:25 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:37.034 23:49:25 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:37.034 23:49:25 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:37.034 23:49:25 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:37.034 23:49:25 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:37.034 23:49:25 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:37.034 23:49:25 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:37.034 23:49:25 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:37.034 23:49:25 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:37.034 23:49:25 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:37.034 "name": "Existed_Raid", 00:15:37.034 "uuid": "bd7964cd-7116-487f-9d18-c8b1e89a412e", 00:15:37.034 "strip_size_kb": 0, 00:15:37.034 "state": "configuring", 00:15:37.034 "raid_level": "raid1", 00:15:37.034 "superblock": true, 00:15:37.034 "num_base_bdevs": 2, 00:15:37.034 "num_base_bdevs_discovered": 1, 00:15:37.034 "num_base_bdevs_operational": 2, 00:15:37.034 "base_bdevs_list": [ 00:15:37.034 { 00:15:37.034 "name": "BaseBdev1", 00:15:37.034 "uuid": "f38cf700-dd54-4e24-bdfe-54bd2a144e23", 00:15:37.034 "is_configured": true, 00:15:37.034 "data_offset": 256, 00:15:37.034 "data_size": 7936 00:15:37.034 }, 00:15:37.034 { 00:15:37.034 "name": "BaseBdev2", 00:15:37.034 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:37.034 "is_configured": false, 00:15:37.034 "data_offset": 0, 00:15:37.034 "data_size": 0 00:15:37.034 } 00:15:37.034 ] 00:15:37.034 }' 00:15:37.034 23:49:25 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:37.034 23:49:25 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:37.604 23:49:25 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:15:37.604 23:49:25 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:37.604 23:49:25 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:37.604 [2024-11-26 23:49:25.469352] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:15:37.605 [2024-11-26 23:49:25.469445] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:15:37.605 23:49:25 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:37.605 23:49:25 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:15:37.605 23:49:25 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:37.605 23:49:25 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:37.605 [2024-11-26 23:49:25.481380] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:15:37.605 [2024-11-26 23:49:25.483189] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:15:37.605 [2024-11-26 23:49:25.483262] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:15:37.605 23:49:25 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:37.605 23:49:25 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:15:37.605 23:49:25 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:15:37.605 23:49:25 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:15:37.605 23:49:25 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:37.605 23:49:25 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:37.605 23:49:25 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:37.605 23:49:25 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:37.605 23:49:25 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:37.605 23:49:25 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:37.605 23:49:25 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:37.605 23:49:25 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:37.605 23:49:25 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:37.605 23:49:25 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:37.605 23:49:25 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:37.605 23:49:25 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:37.605 23:49:25 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:37.605 23:49:25 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:37.605 23:49:25 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:37.605 "name": "Existed_Raid", 00:15:37.605 "uuid": "0680b2aa-29c0-46cf-92e6-f079c382ac9b", 00:15:37.605 "strip_size_kb": 0, 00:15:37.605 "state": "configuring", 00:15:37.605 "raid_level": "raid1", 00:15:37.605 "superblock": true, 00:15:37.605 "num_base_bdevs": 2, 00:15:37.605 "num_base_bdevs_discovered": 1, 00:15:37.605 "num_base_bdevs_operational": 2, 00:15:37.605 "base_bdevs_list": [ 00:15:37.605 { 00:15:37.605 "name": "BaseBdev1", 00:15:37.605 "uuid": "f38cf700-dd54-4e24-bdfe-54bd2a144e23", 00:15:37.605 "is_configured": true, 00:15:37.605 "data_offset": 256, 00:15:37.605 "data_size": 7936 00:15:37.605 }, 00:15:37.605 { 00:15:37.605 "name": "BaseBdev2", 00:15:37.605 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:37.605 "is_configured": false, 00:15:37.605 "data_offset": 0, 00:15:37.605 "data_size": 0 00:15:37.605 } 00:15:37.605 ] 00:15:37.605 }' 00:15:37.605 23:49:25 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:37.605 23:49:25 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:37.866 23:49:25 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -i -b BaseBdev2 00:15:37.866 23:49:25 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:37.866 23:49:25 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:37.866 [2024-11-26 23:49:25.919721] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:15:37.866 [2024-11-26 23:49:25.919957] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:15:37.866 [2024-11-26 23:49:25.919993] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4128 00:15:37.866 [2024-11-26 23:49:25.920097] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:15:37.866 [2024-11-26 23:49:25.920177] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:15:37.866 [2024-11-26 23:49:25.920190] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:15:37.866 [2024-11-26 23:49:25.920244] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:37.866 BaseBdev2 00:15:37.866 23:49:25 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:37.866 23:49:25 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:15:37.866 23:49:25 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:15:37.866 23:49:25 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:15:37.866 23:49:25 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@905 -- # local i 00:15:37.866 23:49:25 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:15:37.866 23:49:25 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:15:37.866 23:49:25 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:15:37.866 23:49:25 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:37.866 23:49:25 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:37.866 23:49:25 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:37.866 23:49:25 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:15:37.866 23:49:25 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:37.866 23:49:25 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:37.866 [ 00:15:37.866 { 00:15:37.866 "name": "BaseBdev2", 00:15:37.866 "aliases": [ 00:15:37.866 "d34eeb96-9ebd-405a-b954-9decd6b8c630" 00:15:37.866 ], 00:15:37.866 "product_name": "Malloc disk", 00:15:37.866 "block_size": 4128, 00:15:37.866 "num_blocks": 8192, 00:15:37.866 "uuid": "d34eeb96-9ebd-405a-b954-9decd6b8c630", 00:15:37.866 "md_size": 32, 00:15:37.866 "md_interleave": true, 00:15:37.866 "dif_type": 0, 00:15:37.866 "assigned_rate_limits": { 00:15:37.866 "rw_ios_per_sec": 0, 00:15:37.866 "rw_mbytes_per_sec": 0, 00:15:37.866 "r_mbytes_per_sec": 0, 00:15:37.866 "w_mbytes_per_sec": 0 00:15:37.866 }, 00:15:37.866 "claimed": true, 00:15:37.866 "claim_type": "exclusive_write", 00:15:37.866 "zoned": false, 00:15:37.866 "supported_io_types": { 00:15:37.866 "read": true, 00:15:37.866 "write": true, 00:15:37.866 "unmap": true, 00:15:37.866 "flush": true, 00:15:37.866 "reset": true, 00:15:37.866 "nvme_admin": false, 00:15:37.866 "nvme_io": false, 00:15:37.866 "nvme_io_md": false, 00:15:37.866 "write_zeroes": true, 00:15:37.866 "zcopy": true, 00:15:37.866 "get_zone_info": false, 00:15:37.866 "zone_management": false, 00:15:37.866 "zone_append": false, 00:15:37.866 "compare": false, 00:15:37.866 "compare_and_write": false, 00:15:37.866 "abort": true, 00:15:37.866 "seek_hole": false, 00:15:37.866 "seek_data": false, 00:15:37.866 "copy": true, 00:15:37.866 "nvme_iov_md": false 00:15:37.866 }, 00:15:37.866 "memory_domains": [ 00:15:37.866 { 00:15:37.866 "dma_device_id": "system", 00:15:37.866 "dma_device_type": 1 00:15:37.866 }, 00:15:37.866 { 00:15:37.866 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:37.866 "dma_device_type": 2 00:15:37.866 } 00:15:37.866 ], 00:15:37.866 "driver_specific": {} 00:15:37.866 } 00:15:37.866 ] 00:15:37.866 23:49:25 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:37.866 23:49:25 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@911 -- # return 0 00:15:37.866 23:49:25 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:15:37.866 23:49:25 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:15:37.866 23:49:25 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid1 0 2 00:15:37.866 23:49:25 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:37.866 23:49:25 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:37.866 23:49:25 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:37.866 23:49:25 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:37.866 23:49:25 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:37.866 23:49:25 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:37.866 23:49:25 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:37.866 23:49:25 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:37.866 23:49:25 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:37.866 23:49:25 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:37.866 23:49:25 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:37.866 23:49:25 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:37.866 23:49:25 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:37.866 23:49:25 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:38.126 23:49:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:38.126 "name": "Existed_Raid", 00:15:38.126 "uuid": "0680b2aa-29c0-46cf-92e6-f079c382ac9b", 00:15:38.126 "strip_size_kb": 0, 00:15:38.126 "state": "online", 00:15:38.126 "raid_level": "raid1", 00:15:38.126 "superblock": true, 00:15:38.126 "num_base_bdevs": 2, 00:15:38.126 "num_base_bdevs_discovered": 2, 00:15:38.126 "num_base_bdevs_operational": 2, 00:15:38.126 "base_bdevs_list": [ 00:15:38.126 { 00:15:38.126 "name": "BaseBdev1", 00:15:38.126 "uuid": "f38cf700-dd54-4e24-bdfe-54bd2a144e23", 00:15:38.126 "is_configured": true, 00:15:38.126 "data_offset": 256, 00:15:38.126 "data_size": 7936 00:15:38.126 }, 00:15:38.126 { 00:15:38.126 "name": "BaseBdev2", 00:15:38.126 "uuid": "d34eeb96-9ebd-405a-b954-9decd6b8c630", 00:15:38.126 "is_configured": true, 00:15:38.126 "data_offset": 256, 00:15:38.126 "data_size": 7936 00:15:38.126 } 00:15:38.126 ] 00:15:38.126 }' 00:15:38.126 23:49:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:38.126 23:49:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:38.386 23:49:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:15:38.386 23:49:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:15:38.386 23:49:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:15:38.386 23:49:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:15:38.386 23:49:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@184 -- # local name 00:15:38.386 23:49:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:15:38.386 23:49:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:15:38.386 23:49:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:15:38.386 23:49:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:38.386 23:49:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:38.386 [2024-11-26 23:49:26.411166] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:38.386 23:49:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:38.386 23:49:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:15:38.386 "name": "Existed_Raid", 00:15:38.386 "aliases": [ 00:15:38.386 "0680b2aa-29c0-46cf-92e6-f079c382ac9b" 00:15:38.386 ], 00:15:38.386 "product_name": "Raid Volume", 00:15:38.386 "block_size": 4128, 00:15:38.386 "num_blocks": 7936, 00:15:38.386 "uuid": "0680b2aa-29c0-46cf-92e6-f079c382ac9b", 00:15:38.386 "md_size": 32, 00:15:38.386 "md_interleave": true, 00:15:38.386 "dif_type": 0, 00:15:38.386 "assigned_rate_limits": { 00:15:38.386 "rw_ios_per_sec": 0, 00:15:38.386 "rw_mbytes_per_sec": 0, 00:15:38.386 "r_mbytes_per_sec": 0, 00:15:38.386 "w_mbytes_per_sec": 0 00:15:38.386 }, 00:15:38.386 "claimed": false, 00:15:38.386 "zoned": false, 00:15:38.386 "supported_io_types": { 00:15:38.386 "read": true, 00:15:38.386 "write": true, 00:15:38.386 "unmap": false, 00:15:38.386 "flush": false, 00:15:38.386 "reset": true, 00:15:38.386 "nvme_admin": false, 00:15:38.386 "nvme_io": false, 00:15:38.386 "nvme_io_md": false, 00:15:38.386 "write_zeroes": true, 00:15:38.386 "zcopy": false, 00:15:38.386 "get_zone_info": false, 00:15:38.386 "zone_management": false, 00:15:38.386 "zone_append": false, 00:15:38.386 "compare": false, 00:15:38.386 "compare_and_write": false, 00:15:38.386 "abort": false, 00:15:38.386 "seek_hole": false, 00:15:38.386 "seek_data": false, 00:15:38.386 "copy": false, 00:15:38.386 "nvme_iov_md": false 00:15:38.386 }, 00:15:38.386 "memory_domains": [ 00:15:38.386 { 00:15:38.386 "dma_device_id": "system", 00:15:38.386 "dma_device_type": 1 00:15:38.386 }, 00:15:38.386 { 00:15:38.386 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:38.386 "dma_device_type": 2 00:15:38.386 }, 00:15:38.386 { 00:15:38.386 "dma_device_id": "system", 00:15:38.386 "dma_device_type": 1 00:15:38.386 }, 00:15:38.386 { 00:15:38.386 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:38.386 "dma_device_type": 2 00:15:38.387 } 00:15:38.387 ], 00:15:38.387 "driver_specific": { 00:15:38.387 "raid": { 00:15:38.387 "uuid": "0680b2aa-29c0-46cf-92e6-f079c382ac9b", 00:15:38.387 "strip_size_kb": 0, 00:15:38.387 "state": "online", 00:15:38.387 "raid_level": "raid1", 00:15:38.387 "superblock": true, 00:15:38.387 "num_base_bdevs": 2, 00:15:38.387 "num_base_bdevs_discovered": 2, 00:15:38.387 "num_base_bdevs_operational": 2, 00:15:38.387 "base_bdevs_list": [ 00:15:38.387 { 00:15:38.387 "name": "BaseBdev1", 00:15:38.387 "uuid": "f38cf700-dd54-4e24-bdfe-54bd2a144e23", 00:15:38.387 "is_configured": true, 00:15:38.387 "data_offset": 256, 00:15:38.387 "data_size": 7936 00:15:38.387 }, 00:15:38.387 { 00:15:38.387 "name": "BaseBdev2", 00:15:38.387 "uuid": "d34eeb96-9ebd-405a-b954-9decd6b8c630", 00:15:38.387 "is_configured": true, 00:15:38.387 "data_offset": 256, 00:15:38.387 "data_size": 7936 00:15:38.387 } 00:15:38.387 ] 00:15:38.387 } 00:15:38.387 } 00:15:38.387 }' 00:15:38.387 23:49:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:15:38.387 23:49:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:15:38.387 BaseBdev2' 00:15:38.387 23:49:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:38.647 23:49:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='4128 32 true 0' 00:15:38.647 23:49:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:38.647 23:49:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:38.647 23:49:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:15:38.647 23:49:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:38.647 23:49:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:38.648 23:49:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:38.648 23:49:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4128 32 true 0' 00:15:38.648 23:49:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@193 -- # [[ 4128 32 true 0 == \4\1\2\8\ \3\2\ \t\r\u\e\ \0 ]] 00:15:38.648 23:49:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:38.648 23:49:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:15:38.648 23:49:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:38.648 23:49:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:38.648 23:49:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:38.648 23:49:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:38.648 23:49:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4128 32 true 0' 00:15:38.648 23:49:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@193 -- # [[ 4128 32 true 0 == \4\1\2\8\ \3\2\ \t\r\u\e\ \0 ]] 00:15:38.648 23:49:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:15:38.648 23:49:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:38.648 23:49:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:38.648 [2024-11-26 23:49:26.638726] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:15:38.648 23:49:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:38.648 23:49:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@260 -- # local expected_state 00:15:38.648 23:49:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@261 -- # has_redundancy raid1 00:15:38.648 23:49:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@198 -- # case $1 in 00:15:38.648 23:49:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@199 -- # return 0 00:15:38.648 23:49:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:15:38.648 23:49:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid1 0 1 00:15:38.648 23:49:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:38.648 23:49:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:38.648 23:49:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:38.648 23:49:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:38.648 23:49:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:15:38.648 23:49:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:38.648 23:49:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:38.648 23:49:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:38.648 23:49:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:38.648 23:49:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:38.648 23:49:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:38.648 23:49:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:38.648 23:49:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:38.648 23:49:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:38.648 23:49:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:38.648 "name": "Existed_Raid", 00:15:38.648 "uuid": "0680b2aa-29c0-46cf-92e6-f079c382ac9b", 00:15:38.648 "strip_size_kb": 0, 00:15:38.648 "state": "online", 00:15:38.648 "raid_level": "raid1", 00:15:38.648 "superblock": true, 00:15:38.648 "num_base_bdevs": 2, 00:15:38.648 "num_base_bdevs_discovered": 1, 00:15:38.648 "num_base_bdevs_operational": 1, 00:15:38.648 "base_bdevs_list": [ 00:15:38.648 { 00:15:38.648 "name": null, 00:15:38.648 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:38.648 "is_configured": false, 00:15:38.648 "data_offset": 0, 00:15:38.648 "data_size": 7936 00:15:38.648 }, 00:15:38.648 { 00:15:38.648 "name": "BaseBdev2", 00:15:38.648 "uuid": "d34eeb96-9ebd-405a-b954-9decd6b8c630", 00:15:38.648 "is_configured": true, 00:15:38.648 "data_offset": 256, 00:15:38.648 "data_size": 7936 00:15:38.648 } 00:15:38.648 ] 00:15:38.648 }' 00:15:38.648 23:49:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:38.648 23:49:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:39.227 23:49:27 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:15:39.227 23:49:27 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:15:39.227 23:49:27 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:15:39.227 23:49:27 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:39.227 23:49:27 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:39.227 23:49:27 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:39.227 23:49:27 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:39.227 23:49:27 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:15:39.227 23:49:27 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:15:39.227 23:49:27 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:15:39.227 23:49:27 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:39.227 23:49:27 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:39.227 [2024-11-26 23:49:27.153340] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:15:39.227 [2024-11-26 23:49:27.153443] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:15:39.227 [2024-11-26 23:49:27.165335] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:39.227 [2024-11-26 23:49:27.165403] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:39.227 [2024-11-26 23:49:27.165414] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:15:39.227 23:49:27 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:39.227 23:49:27 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:15:39.227 23:49:27 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:15:39.227 23:49:27 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:39.227 23:49:27 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:39.227 23:49:27 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:39.227 23:49:27 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:15:39.227 23:49:27 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:39.227 23:49:27 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:15:39.227 23:49:27 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:15:39.227 23:49:27 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@284 -- # '[' 2 -gt 2 ']' 00:15:39.227 23:49:27 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@326 -- # killprocess 98416 00:15:39.227 23:49:27 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@954 -- # '[' -z 98416 ']' 00:15:39.227 23:49:27 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@958 -- # kill -0 98416 00:15:39.227 23:49:27 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@959 -- # uname 00:15:39.227 23:49:27 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:15:39.227 23:49:27 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 98416 00:15:39.227 23:49:27 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:15:39.227 23:49:27 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:15:39.227 killing process with pid 98416 00:15:39.227 23:49:27 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@972 -- # echo 'killing process with pid 98416' 00:15:39.227 23:49:27 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@973 -- # kill 98416 00:15:39.227 [2024-11-26 23:49:27.261508] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:15:39.227 23:49:27 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@978 -- # wait 98416 00:15:39.227 [2024-11-26 23:49:27.262464] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:15:39.488 23:49:27 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@328 -- # return 0 00:15:39.488 00:15:39.488 real 0m3.960s 00:15:39.488 user 0m6.269s 00:15:39.488 sys 0m0.845s 00:15:39.488 23:49:27 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@1130 -- # xtrace_disable 00:15:39.488 ************************************ 00:15:39.488 END TEST raid_state_function_test_sb_md_interleaved 00:15:39.488 ************************************ 00:15:39.488 23:49:27 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:39.488 23:49:27 bdev_raid -- bdev/bdev_raid.sh@1012 -- # run_test raid_superblock_test_md_interleaved raid_superblock_test raid1 2 00:15:39.488 23:49:27 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:15:39.488 23:49:27 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:15:39.488 23:49:27 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:15:39.488 ************************************ 00:15:39.488 START TEST raid_superblock_test_md_interleaved 00:15:39.488 ************************************ 00:15:39.488 23:49:27 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@1129 -- # raid_superblock_test raid1 2 00:15:39.488 23:49:27 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@393 -- # local raid_level=raid1 00:15:39.488 23:49:27 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=2 00:15:39.488 23:49:27 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:15:39.488 23:49:27 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:15:39.488 23:49:27 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:15:39.488 23:49:27 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:15:39.488 23:49:27 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:15:39.488 23:49:27 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:15:39.488 23:49:27 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:15:39.488 23:49:27 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@399 -- # local strip_size 00:15:39.488 23:49:27 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:15:39.488 23:49:27 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:15:39.488 23:49:27 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:15:39.488 23:49:27 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@404 -- # '[' raid1 '!=' raid1 ']' 00:15:39.488 23:49:27 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@408 -- # strip_size=0 00:15:39.488 23:49:27 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@412 -- # raid_pid=98657 00:15:39.488 23:49:27 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:15:39.488 23:49:27 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@413 -- # waitforlisten 98657 00:15:39.488 23:49:27 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@835 -- # '[' -z 98657 ']' 00:15:39.488 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:15:39.488 23:49:27 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:15:39.488 23:49:27 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@840 -- # local max_retries=100 00:15:39.488 23:49:27 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:15:39.488 23:49:27 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@844 -- # xtrace_disable 00:15:39.488 23:49:27 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:39.748 [2024-11-26 23:49:27.650301] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:15:39.749 [2024-11-26 23:49:27.650440] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid98657 ] 00:15:39.749 [2024-11-26 23:49:27.804121] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:15:39.749 [2024-11-26 23:49:27.829409] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:15:39.749 [2024-11-26 23:49:27.872024] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:15:39.749 [2024-11-26 23:49:27.872060] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:15:40.368 23:49:28 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:15:40.368 23:49:28 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@868 -- # return 0 00:15:40.368 23:49:28 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:15:40.368 23:49:28 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:15:40.368 23:49:28 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:15:40.368 23:49:28 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:15:40.368 23:49:28 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:15:40.368 23:49:28 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:15:40.368 23:49:28 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:15:40.368 23:49:28 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:15:40.368 23:49:28 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -i -b malloc1 00:15:40.368 23:49:28 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:40.368 23:49:28 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:40.655 malloc1 00:15:40.655 23:49:28 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:40.655 23:49:28 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:15:40.655 23:49:28 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:40.655 23:49:28 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:40.655 [2024-11-26 23:49:28.499228] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:15:40.655 [2024-11-26 23:49:28.499403] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:40.655 [2024-11-26 23:49:28.499446] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:15:40.655 [2024-11-26 23:49:28.499503] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:40.655 [2024-11-26 23:49:28.501335] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:40.655 [2024-11-26 23:49:28.501438] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:15:40.655 pt1 00:15:40.655 23:49:28 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:40.655 23:49:28 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:15:40.655 23:49:28 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:15:40.655 23:49:28 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:15:40.655 23:49:28 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:15:40.655 23:49:28 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:15:40.655 23:49:28 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:15:40.655 23:49:28 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:15:40.655 23:49:28 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:15:40.656 23:49:28 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -i -b malloc2 00:15:40.656 23:49:28 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:40.656 23:49:28 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:40.656 malloc2 00:15:40.656 23:49:28 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:40.656 23:49:28 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:15:40.656 23:49:28 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:40.656 23:49:28 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:40.656 [2024-11-26 23:49:28.531871] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:15:40.656 [2024-11-26 23:49:28.531974] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:40.656 [2024-11-26 23:49:28.532021] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:15:40.656 [2024-11-26 23:49:28.532049] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:40.656 [2024-11-26 23:49:28.533880] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:40.656 [2024-11-26 23:49:28.533917] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:15:40.656 pt2 00:15:40.656 23:49:28 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:40.656 23:49:28 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:15:40.656 23:49:28 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:15:40.656 23:49:28 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''pt1 pt2'\''' -n raid_bdev1 -s 00:15:40.656 23:49:28 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:40.656 23:49:28 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:40.656 [2024-11-26 23:49:28.543874] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:15:40.656 [2024-11-26 23:49:28.545615] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:15:40.656 [2024-11-26 23:49:28.545752] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:15:40.656 [2024-11-26 23:49:28.545773] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4128 00:15:40.656 [2024-11-26 23:49:28.545851] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:15:40.656 [2024-11-26 23:49:28.545917] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:15:40.656 [2024-11-26 23:49:28.545927] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:15:40.656 [2024-11-26 23:49:28.545994] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:40.656 23:49:28 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:40.656 23:49:28 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:15:40.656 23:49:28 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:40.656 23:49:28 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:40.656 23:49:28 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:40.656 23:49:28 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:40.656 23:49:28 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:40.656 23:49:28 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:40.656 23:49:28 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:40.656 23:49:28 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:40.656 23:49:28 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:40.656 23:49:28 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:40.656 23:49:28 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:40.656 23:49:28 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:40.656 23:49:28 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:40.656 23:49:28 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:40.656 23:49:28 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:40.656 "name": "raid_bdev1", 00:15:40.656 "uuid": "1c889e41-1db4-4bd6-a360-0c7a9febaf1f", 00:15:40.656 "strip_size_kb": 0, 00:15:40.656 "state": "online", 00:15:40.656 "raid_level": "raid1", 00:15:40.656 "superblock": true, 00:15:40.656 "num_base_bdevs": 2, 00:15:40.656 "num_base_bdevs_discovered": 2, 00:15:40.656 "num_base_bdevs_operational": 2, 00:15:40.656 "base_bdevs_list": [ 00:15:40.656 { 00:15:40.656 "name": "pt1", 00:15:40.656 "uuid": "00000000-0000-0000-0000-000000000001", 00:15:40.656 "is_configured": true, 00:15:40.656 "data_offset": 256, 00:15:40.656 "data_size": 7936 00:15:40.656 }, 00:15:40.656 { 00:15:40.656 "name": "pt2", 00:15:40.656 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:40.656 "is_configured": true, 00:15:40.656 "data_offset": 256, 00:15:40.656 "data_size": 7936 00:15:40.656 } 00:15:40.656 ] 00:15:40.656 }' 00:15:40.656 23:49:28 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:40.656 23:49:28 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:40.929 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:15:40.929 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:15:40.929 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:15:40.929 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:15:40.930 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@184 -- # local name 00:15:40.930 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:15:40.930 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:15:40.930 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:15:40.930 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:40.930 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:40.930 [2024-11-26 23:49:29.015330] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:40.930 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:40.930 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:15:40.930 "name": "raid_bdev1", 00:15:40.930 "aliases": [ 00:15:40.930 "1c889e41-1db4-4bd6-a360-0c7a9febaf1f" 00:15:40.930 ], 00:15:40.930 "product_name": "Raid Volume", 00:15:40.930 "block_size": 4128, 00:15:40.930 "num_blocks": 7936, 00:15:40.930 "uuid": "1c889e41-1db4-4bd6-a360-0c7a9febaf1f", 00:15:40.930 "md_size": 32, 00:15:40.930 "md_interleave": true, 00:15:40.930 "dif_type": 0, 00:15:40.930 "assigned_rate_limits": { 00:15:40.930 "rw_ios_per_sec": 0, 00:15:40.930 "rw_mbytes_per_sec": 0, 00:15:40.930 "r_mbytes_per_sec": 0, 00:15:40.930 "w_mbytes_per_sec": 0 00:15:40.930 }, 00:15:40.930 "claimed": false, 00:15:40.930 "zoned": false, 00:15:40.930 "supported_io_types": { 00:15:40.930 "read": true, 00:15:40.930 "write": true, 00:15:40.930 "unmap": false, 00:15:40.930 "flush": false, 00:15:40.930 "reset": true, 00:15:40.930 "nvme_admin": false, 00:15:40.930 "nvme_io": false, 00:15:40.930 "nvme_io_md": false, 00:15:40.930 "write_zeroes": true, 00:15:40.930 "zcopy": false, 00:15:40.930 "get_zone_info": false, 00:15:40.930 "zone_management": false, 00:15:40.930 "zone_append": false, 00:15:40.930 "compare": false, 00:15:40.930 "compare_and_write": false, 00:15:40.930 "abort": false, 00:15:40.930 "seek_hole": false, 00:15:40.930 "seek_data": false, 00:15:40.930 "copy": false, 00:15:40.930 "nvme_iov_md": false 00:15:40.930 }, 00:15:40.930 "memory_domains": [ 00:15:40.930 { 00:15:40.930 "dma_device_id": "system", 00:15:40.930 "dma_device_type": 1 00:15:40.930 }, 00:15:40.930 { 00:15:40.930 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:40.930 "dma_device_type": 2 00:15:40.930 }, 00:15:40.930 { 00:15:40.930 "dma_device_id": "system", 00:15:40.930 "dma_device_type": 1 00:15:40.930 }, 00:15:40.930 { 00:15:40.930 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:40.930 "dma_device_type": 2 00:15:40.930 } 00:15:40.930 ], 00:15:40.930 "driver_specific": { 00:15:40.930 "raid": { 00:15:40.930 "uuid": "1c889e41-1db4-4bd6-a360-0c7a9febaf1f", 00:15:40.930 "strip_size_kb": 0, 00:15:40.930 "state": "online", 00:15:40.930 "raid_level": "raid1", 00:15:40.930 "superblock": true, 00:15:40.930 "num_base_bdevs": 2, 00:15:40.930 "num_base_bdevs_discovered": 2, 00:15:40.930 "num_base_bdevs_operational": 2, 00:15:40.930 "base_bdevs_list": [ 00:15:40.930 { 00:15:40.930 "name": "pt1", 00:15:40.930 "uuid": "00000000-0000-0000-0000-000000000001", 00:15:40.930 "is_configured": true, 00:15:40.930 "data_offset": 256, 00:15:40.930 "data_size": 7936 00:15:40.930 }, 00:15:40.930 { 00:15:40.930 "name": "pt2", 00:15:40.930 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:40.930 "is_configured": true, 00:15:40.930 "data_offset": 256, 00:15:40.930 "data_size": 7936 00:15:40.930 } 00:15:40.930 ] 00:15:40.930 } 00:15:40.930 } 00:15:40.930 }' 00:15:40.930 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:15:41.189 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:15:41.189 pt2' 00:15:41.189 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:41.190 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='4128 32 true 0' 00:15:41.190 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:41.190 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:15:41.190 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:41.190 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:41.190 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:41.190 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:41.190 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4128 32 true 0' 00:15:41.190 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@193 -- # [[ 4128 32 true 0 == \4\1\2\8\ \3\2\ \t\r\u\e\ \0 ]] 00:15:41.190 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:41.190 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:15:41.190 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:41.190 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:41.190 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:41.190 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:41.190 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4128 32 true 0' 00:15:41.190 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@193 -- # [[ 4128 32 true 0 == \4\1\2\8\ \3\2\ \t\r\u\e\ \0 ]] 00:15:41.190 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:15:41.190 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:15:41.190 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:41.190 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:41.190 [2024-11-26 23:49:29.230902] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:41.190 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:41.190 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=1c889e41-1db4-4bd6-a360-0c7a9febaf1f 00:15:41.190 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@436 -- # '[' -z 1c889e41-1db4-4bd6-a360-0c7a9febaf1f ']' 00:15:41.190 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:15:41.190 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:41.190 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:41.190 [2024-11-26 23:49:29.270614] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:15:41.190 [2024-11-26 23:49:29.270677] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:15:41.190 [2024-11-26 23:49:29.270786] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:41.190 [2024-11-26 23:49:29.270870] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:41.190 [2024-11-26 23:49:29.270885] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:15:41.190 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:41.190 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:41.190 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:41.190 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:41.190 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:15:41.190 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:41.450 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:15:41.450 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:15:41.450 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:15:41.450 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:15:41.450 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:41.450 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:41.450 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:41.450 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:15:41.450 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:15:41.450 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:41.450 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:41.450 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:41.450 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:15:41.450 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:41.450 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:41.450 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:15:41.450 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:41.450 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:15:41.450 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:15:41.450 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@652 -- # local es=0 00:15:41.450 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:15:41.450 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:15:41.450 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:15:41.450 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:15:41.450 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:15:41.450 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:15:41.450 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:41.450 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:41.450 [2024-11-26 23:49:29.410395] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:15:41.450 [2024-11-26 23:49:29.412076] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:15:41.450 [2024-11-26 23:49:29.412128] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:15:41.450 [2024-11-26 23:49:29.412195] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:15:41.450 [2024-11-26 23:49:29.412213] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:15:41.450 [2024-11-26 23:49:29.412222] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state configuring 00:15:41.450 request: 00:15:41.450 { 00:15:41.450 "name": "raid_bdev1", 00:15:41.450 "raid_level": "raid1", 00:15:41.450 "base_bdevs": [ 00:15:41.450 "malloc1", 00:15:41.450 "malloc2" 00:15:41.450 ], 00:15:41.450 "superblock": false, 00:15:41.450 "method": "bdev_raid_create", 00:15:41.450 "req_id": 1 00:15:41.450 } 00:15:41.450 Got JSON-RPC error response 00:15:41.451 response: 00:15:41.451 { 00:15:41.451 "code": -17, 00:15:41.451 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:15:41.451 } 00:15:41.451 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:15:41.451 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@655 -- # es=1 00:15:41.451 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:15:41.451 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:15:41.451 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:15:41.451 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:41.451 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:15:41.451 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:41.451 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:41.451 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:41.451 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:15:41.451 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:15:41.451 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:15:41.451 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:41.451 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:41.451 [2024-11-26 23:49:29.478237] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:15:41.451 [2024-11-26 23:49:29.478286] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:41.451 [2024-11-26 23:49:29.478318] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:15:41.451 [2024-11-26 23:49:29.478326] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:41.451 [2024-11-26 23:49:29.480165] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:41.451 [2024-11-26 23:49:29.480200] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:15:41.451 [2024-11-26 23:49:29.480240] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:15:41.451 [2024-11-26 23:49:29.480277] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:15:41.451 pt1 00:15:41.451 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:41.451 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 2 00:15:41.451 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:41.451 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:41.451 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:41.451 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:41.451 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:41.451 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:41.451 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:41.451 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:41.451 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:41.451 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:41.451 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:41.451 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:41.451 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:41.451 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:41.451 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:41.451 "name": "raid_bdev1", 00:15:41.451 "uuid": "1c889e41-1db4-4bd6-a360-0c7a9febaf1f", 00:15:41.451 "strip_size_kb": 0, 00:15:41.451 "state": "configuring", 00:15:41.451 "raid_level": "raid1", 00:15:41.451 "superblock": true, 00:15:41.451 "num_base_bdevs": 2, 00:15:41.451 "num_base_bdevs_discovered": 1, 00:15:41.451 "num_base_bdevs_operational": 2, 00:15:41.451 "base_bdevs_list": [ 00:15:41.451 { 00:15:41.451 "name": "pt1", 00:15:41.451 "uuid": "00000000-0000-0000-0000-000000000001", 00:15:41.451 "is_configured": true, 00:15:41.451 "data_offset": 256, 00:15:41.451 "data_size": 7936 00:15:41.451 }, 00:15:41.451 { 00:15:41.451 "name": null, 00:15:41.451 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:41.451 "is_configured": false, 00:15:41.451 "data_offset": 256, 00:15:41.451 "data_size": 7936 00:15:41.451 } 00:15:41.451 ] 00:15:41.451 }' 00:15:41.451 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:41.451 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:42.021 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@470 -- # '[' 2 -gt 2 ']' 00:15:42.021 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:15:42.021 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:15:42.021 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:15:42.021 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:42.021 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:42.021 [2024-11-26 23:49:29.941455] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:15:42.021 [2024-11-26 23:49:29.941557] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:42.021 [2024-11-26 23:49:29.941595] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:15:42.021 [2024-11-26 23:49:29.941621] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:42.021 [2024-11-26 23:49:29.941790] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:42.021 [2024-11-26 23:49:29.941837] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:15:42.021 [2024-11-26 23:49:29.941896] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:15:42.021 [2024-11-26 23:49:29.941936] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:15:42.021 [2024-11-26 23:49:29.942033] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:15:42.021 [2024-11-26 23:49:29.942072] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4128 00:15:42.022 [2024-11-26 23:49:29.942154] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:15:42.022 [2024-11-26 23:49:29.942235] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:15:42.022 [2024-11-26 23:49:29.942273] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:15:42.022 [2024-11-26 23:49:29.942364] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:42.022 pt2 00:15:42.022 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:42.022 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:15:42.022 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:15:42.022 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:15:42.022 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:42.022 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:42.022 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:42.022 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:42.022 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:42.022 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:42.022 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:42.022 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:42.022 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:42.022 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:42.022 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:42.022 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:42.022 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:42.022 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:42.022 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:42.022 "name": "raid_bdev1", 00:15:42.022 "uuid": "1c889e41-1db4-4bd6-a360-0c7a9febaf1f", 00:15:42.022 "strip_size_kb": 0, 00:15:42.022 "state": "online", 00:15:42.022 "raid_level": "raid1", 00:15:42.022 "superblock": true, 00:15:42.022 "num_base_bdevs": 2, 00:15:42.022 "num_base_bdevs_discovered": 2, 00:15:42.022 "num_base_bdevs_operational": 2, 00:15:42.022 "base_bdevs_list": [ 00:15:42.022 { 00:15:42.022 "name": "pt1", 00:15:42.022 "uuid": "00000000-0000-0000-0000-000000000001", 00:15:42.022 "is_configured": true, 00:15:42.022 "data_offset": 256, 00:15:42.022 "data_size": 7936 00:15:42.022 }, 00:15:42.022 { 00:15:42.022 "name": "pt2", 00:15:42.022 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:42.022 "is_configured": true, 00:15:42.022 "data_offset": 256, 00:15:42.022 "data_size": 7936 00:15:42.022 } 00:15:42.022 ] 00:15:42.022 }' 00:15:42.022 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:42.022 23:49:29 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:42.603 23:49:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:15:42.603 23:49:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:15:42.603 23:49:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:15:42.603 23:49:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:15:42.603 23:49:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@184 -- # local name 00:15:42.603 23:49:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:15:42.603 23:49:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:15:42.603 23:49:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:15:42.603 23:49:30 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:42.603 23:49:30 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:42.603 [2024-11-26 23:49:30.436855] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:42.604 23:49:30 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:42.604 23:49:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:15:42.604 "name": "raid_bdev1", 00:15:42.604 "aliases": [ 00:15:42.604 "1c889e41-1db4-4bd6-a360-0c7a9febaf1f" 00:15:42.604 ], 00:15:42.604 "product_name": "Raid Volume", 00:15:42.604 "block_size": 4128, 00:15:42.604 "num_blocks": 7936, 00:15:42.604 "uuid": "1c889e41-1db4-4bd6-a360-0c7a9febaf1f", 00:15:42.604 "md_size": 32, 00:15:42.604 "md_interleave": true, 00:15:42.604 "dif_type": 0, 00:15:42.604 "assigned_rate_limits": { 00:15:42.604 "rw_ios_per_sec": 0, 00:15:42.604 "rw_mbytes_per_sec": 0, 00:15:42.604 "r_mbytes_per_sec": 0, 00:15:42.604 "w_mbytes_per_sec": 0 00:15:42.604 }, 00:15:42.604 "claimed": false, 00:15:42.604 "zoned": false, 00:15:42.604 "supported_io_types": { 00:15:42.604 "read": true, 00:15:42.604 "write": true, 00:15:42.604 "unmap": false, 00:15:42.604 "flush": false, 00:15:42.604 "reset": true, 00:15:42.604 "nvme_admin": false, 00:15:42.604 "nvme_io": false, 00:15:42.604 "nvme_io_md": false, 00:15:42.604 "write_zeroes": true, 00:15:42.604 "zcopy": false, 00:15:42.604 "get_zone_info": false, 00:15:42.604 "zone_management": false, 00:15:42.604 "zone_append": false, 00:15:42.604 "compare": false, 00:15:42.604 "compare_and_write": false, 00:15:42.604 "abort": false, 00:15:42.604 "seek_hole": false, 00:15:42.604 "seek_data": false, 00:15:42.604 "copy": false, 00:15:42.604 "nvme_iov_md": false 00:15:42.604 }, 00:15:42.604 "memory_domains": [ 00:15:42.604 { 00:15:42.604 "dma_device_id": "system", 00:15:42.604 "dma_device_type": 1 00:15:42.604 }, 00:15:42.604 { 00:15:42.604 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:42.604 "dma_device_type": 2 00:15:42.604 }, 00:15:42.604 { 00:15:42.604 "dma_device_id": "system", 00:15:42.604 "dma_device_type": 1 00:15:42.604 }, 00:15:42.604 { 00:15:42.604 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:42.604 "dma_device_type": 2 00:15:42.604 } 00:15:42.604 ], 00:15:42.604 "driver_specific": { 00:15:42.604 "raid": { 00:15:42.604 "uuid": "1c889e41-1db4-4bd6-a360-0c7a9febaf1f", 00:15:42.604 "strip_size_kb": 0, 00:15:42.604 "state": "online", 00:15:42.604 "raid_level": "raid1", 00:15:42.604 "superblock": true, 00:15:42.604 "num_base_bdevs": 2, 00:15:42.604 "num_base_bdevs_discovered": 2, 00:15:42.604 "num_base_bdevs_operational": 2, 00:15:42.604 "base_bdevs_list": [ 00:15:42.604 { 00:15:42.604 "name": "pt1", 00:15:42.604 "uuid": "00000000-0000-0000-0000-000000000001", 00:15:42.604 "is_configured": true, 00:15:42.604 "data_offset": 256, 00:15:42.604 "data_size": 7936 00:15:42.604 }, 00:15:42.605 { 00:15:42.605 "name": "pt2", 00:15:42.605 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:42.605 "is_configured": true, 00:15:42.605 "data_offset": 256, 00:15:42.605 "data_size": 7936 00:15:42.605 } 00:15:42.605 ] 00:15:42.605 } 00:15:42.605 } 00:15:42.605 }' 00:15:42.605 23:49:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:15:42.605 23:49:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:15:42.605 pt2' 00:15:42.605 23:49:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:42.605 23:49:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='4128 32 true 0' 00:15:42.605 23:49:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:42.605 23:49:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:15:42.605 23:49:30 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:42.605 23:49:30 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:42.605 23:49:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:42.605 23:49:30 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:42.605 23:49:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4128 32 true 0' 00:15:42.605 23:49:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@193 -- # [[ 4128 32 true 0 == \4\1\2\8\ \3\2\ \t\r\u\e\ \0 ]] 00:15:42.605 23:49:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:42.605 23:49:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:15:42.605 23:49:30 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:42.605 23:49:30 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:42.605 23:49:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:42.605 23:49:30 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:42.605 23:49:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4128 32 true 0' 00:15:42.605 23:49:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@193 -- # [[ 4128 32 true 0 == \4\1\2\8\ \3\2\ \t\r\u\e\ \0 ]] 00:15:42.605 23:49:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:15:42.605 23:49:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:15:42.605 23:49:30 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:42.605 23:49:30 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:42.605 [2024-11-26 23:49:30.684405] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:42.606 23:49:30 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:42.606 23:49:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@487 -- # '[' 1c889e41-1db4-4bd6-a360-0c7a9febaf1f '!=' 1c889e41-1db4-4bd6-a360-0c7a9febaf1f ']' 00:15:42.606 23:49:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@491 -- # has_redundancy raid1 00:15:42.606 23:49:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@198 -- # case $1 in 00:15:42.606 23:49:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@199 -- # return 0 00:15:42.606 23:49:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@493 -- # rpc_cmd bdev_passthru_delete pt1 00:15:42.606 23:49:30 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:42.606 23:49:30 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:42.606 [2024-11-26 23:49:30.712149] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt1 00:15:42.606 23:49:30 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:42.606 23:49:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@496 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:15:42.606 23:49:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:42.606 23:49:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:42.606 23:49:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:42.606 23:49:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:42.606 23:49:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:15:42.606 23:49:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:42.606 23:49:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:42.606 23:49:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:42.606 23:49:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:42.606 23:49:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:42.606 23:49:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:42.606 23:49:30 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:42.606 23:49:30 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:42.867 23:49:30 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:42.867 23:49:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:42.867 "name": "raid_bdev1", 00:15:42.867 "uuid": "1c889e41-1db4-4bd6-a360-0c7a9febaf1f", 00:15:42.867 "strip_size_kb": 0, 00:15:42.867 "state": "online", 00:15:42.867 "raid_level": "raid1", 00:15:42.867 "superblock": true, 00:15:42.867 "num_base_bdevs": 2, 00:15:42.867 "num_base_bdevs_discovered": 1, 00:15:42.867 "num_base_bdevs_operational": 1, 00:15:42.867 "base_bdevs_list": [ 00:15:42.867 { 00:15:42.867 "name": null, 00:15:42.867 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:42.867 "is_configured": false, 00:15:42.867 "data_offset": 0, 00:15:42.867 "data_size": 7936 00:15:42.867 }, 00:15:42.867 { 00:15:42.867 "name": "pt2", 00:15:42.867 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:42.867 "is_configured": true, 00:15:42.867 "data_offset": 256, 00:15:42.867 "data_size": 7936 00:15:42.867 } 00:15:42.867 ] 00:15:42.867 }' 00:15:42.867 23:49:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:42.867 23:49:30 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:43.127 23:49:31 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@499 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:15:43.127 23:49:31 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:43.127 23:49:31 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:43.127 [2024-11-26 23:49:31.211248] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:15:43.127 [2024-11-26 23:49:31.211317] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:15:43.127 [2024-11-26 23:49:31.211395] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:43.127 [2024-11-26 23:49:31.211464] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:43.127 [2024-11-26 23:49:31.211475] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:15:43.127 23:49:31 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:43.127 23:49:31 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@500 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:43.127 23:49:31 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:43.127 23:49:31 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@500 -- # jq -r '.[]' 00:15:43.127 23:49:31 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:43.127 23:49:31 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:43.388 23:49:31 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@500 -- # raid_bdev= 00:15:43.388 23:49:31 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@501 -- # '[' -n '' ']' 00:15:43.388 23:49:31 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@506 -- # (( i = 1 )) 00:15:43.388 23:49:31 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:15:43.388 23:49:31 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt2 00:15:43.388 23:49:31 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:43.388 23:49:31 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:43.388 23:49:31 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:43.388 23:49:31 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:15:43.388 23:49:31 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:15:43.388 23:49:31 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@511 -- # (( i = 1 )) 00:15:43.388 23:49:31 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:15:43.388 23:49:31 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@519 -- # i=1 00:15:43.388 23:49:31 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@520 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:15:43.388 23:49:31 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:43.388 23:49:31 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:43.388 [2024-11-26 23:49:31.287115] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:15:43.388 [2024-11-26 23:49:31.287160] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:43.388 [2024-11-26 23:49:31.287192] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008780 00:15:43.388 [2024-11-26 23:49:31.287200] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:43.388 [2024-11-26 23:49:31.289052] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:43.388 [2024-11-26 23:49:31.289086] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:15:43.388 [2024-11-26 23:49:31.289130] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:15:43.388 [2024-11-26 23:49:31.289163] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:15:43.388 [2024-11-26 23:49:31.289222] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:15:43.388 [2024-11-26 23:49:31.289229] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4128 00:15:43.388 [2024-11-26 23:49:31.289290] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:15:43.388 [2024-11-26 23:49:31.289337] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:15:43.388 [2024-11-26 23:49:31.289369] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001c80 00:15:43.388 [2024-11-26 23:49:31.289417] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:43.388 pt2 00:15:43.388 23:49:31 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:43.388 23:49:31 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@523 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:15:43.388 23:49:31 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:43.388 23:49:31 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:43.388 23:49:31 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:43.388 23:49:31 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:43.388 23:49:31 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:15:43.388 23:49:31 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:43.388 23:49:31 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:43.388 23:49:31 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:43.388 23:49:31 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:43.388 23:49:31 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:43.389 23:49:31 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:43.389 23:49:31 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:43.389 23:49:31 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:43.389 23:49:31 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:43.389 23:49:31 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:43.389 "name": "raid_bdev1", 00:15:43.389 "uuid": "1c889e41-1db4-4bd6-a360-0c7a9febaf1f", 00:15:43.389 "strip_size_kb": 0, 00:15:43.389 "state": "online", 00:15:43.389 "raid_level": "raid1", 00:15:43.389 "superblock": true, 00:15:43.389 "num_base_bdevs": 2, 00:15:43.389 "num_base_bdevs_discovered": 1, 00:15:43.389 "num_base_bdevs_operational": 1, 00:15:43.389 "base_bdevs_list": [ 00:15:43.389 { 00:15:43.389 "name": null, 00:15:43.389 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:43.389 "is_configured": false, 00:15:43.389 "data_offset": 256, 00:15:43.389 "data_size": 7936 00:15:43.389 }, 00:15:43.389 { 00:15:43.389 "name": "pt2", 00:15:43.389 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:43.389 "is_configured": true, 00:15:43.389 "data_offset": 256, 00:15:43.389 "data_size": 7936 00:15:43.389 } 00:15:43.389 ] 00:15:43.389 }' 00:15:43.389 23:49:31 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:43.389 23:49:31 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:43.650 23:49:31 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@526 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:15:43.650 23:49:31 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:43.650 23:49:31 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:43.650 [2024-11-26 23:49:31.694422] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:15:43.650 [2024-11-26 23:49:31.694494] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:15:43.650 [2024-11-26 23:49:31.694582] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:43.650 [2024-11-26 23:49:31.694624] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:43.650 [2024-11-26 23:49:31.694637] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name raid_bdev1, state offline 00:15:43.650 23:49:31 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:43.650 23:49:31 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@527 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:43.650 23:49:31 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@527 -- # jq -r '.[]' 00:15:43.650 23:49:31 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:43.650 23:49:31 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:43.650 23:49:31 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:43.650 23:49:31 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@527 -- # raid_bdev= 00:15:43.650 23:49:31 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@528 -- # '[' -n '' ']' 00:15:43.650 23:49:31 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@532 -- # '[' 2 -gt 2 ']' 00:15:43.650 23:49:31 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@540 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:15:43.650 23:49:31 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:43.650 23:49:31 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:43.650 [2024-11-26 23:49:31.754315] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:15:43.650 [2024-11-26 23:49:31.754442] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:43.650 [2024-11-26 23:49:31.754461] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008d80 00:15:43.650 [2024-11-26 23:49:31.754473] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:43.650 [2024-11-26 23:49:31.756313] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:43.650 [2024-11-26 23:49:31.756363] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:15:43.650 [2024-11-26 23:49:31.756409] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:15:43.650 [2024-11-26 23:49:31.756441] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:15:43.650 [2024-11-26 23:49:31.756511] bdev_raid.c:3685:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev pt2 (4) greater than existing raid bdev raid_bdev1 (2) 00:15:43.650 [2024-11-26 23:49:31.756522] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:15:43.650 [2024-11-26 23:49:31.756539] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002000 name raid_bdev1, state configuring 00:15:43.650 [2024-11-26 23:49:31.756584] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:15:43.650 [2024-11-26 23:49:31.756646] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000002380 00:15:43.650 [2024-11-26 23:49:31.756657] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4128 00:15:43.650 [2024-11-26 23:49:31.756733] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:15:43.650 [2024-11-26 23:49:31.756794] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000002380 00:15:43.650 [2024-11-26 23:49:31.756801] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000002380 00:15:43.650 [2024-11-26 23:49:31.756876] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:43.650 pt1 00:15:43.650 23:49:31 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:43.650 23:49:31 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@542 -- # '[' 2 -gt 2 ']' 00:15:43.650 23:49:31 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@554 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:15:43.650 23:49:31 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:43.650 23:49:31 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:43.650 23:49:31 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:43.650 23:49:31 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:43.650 23:49:31 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:15:43.650 23:49:31 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:43.650 23:49:31 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:43.650 23:49:31 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:43.650 23:49:31 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:43.650 23:49:31 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:43.650 23:49:31 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:43.651 23:49:31 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:43.651 23:49:31 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:43.912 23:49:31 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:43.912 23:49:31 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:43.912 "name": "raid_bdev1", 00:15:43.912 "uuid": "1c889e41-1db4-4bd6-a360-0c7a9febaf1f", 00:15:43.912 "strip_size_kb": 0, 00:15:43.912 "state": "online", 00:15:43.912 "raid_level": "raid1", 00:15:43.912 "superblock": true, 00:15:43.912 "num_base_bdevs": 2, 00:15:43.912 "num_base_bdevs_discovered": 1, 00:15:43.912 "num_base_bdevs_operational": 1, 00:15:43.912 "base_bdevs_list": [ 00:15:43.912 { 00:15:43.912 "name": null, 00:15:43.912 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:43.912 "is_configured": false, 00:15:43.912 "data_offset": 256, 00:15:43.912 "data_size": 7936 00:15:43.912 }, 00:15:43.912 { 00:15:43.912 "name": "pt2", 00:15:43.912 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:43.912 "is_configured": true, 00:15:43.912 "data_offset": 256, 00:15:43.912 "data_size": 7936 00:15:43.912 } 00:15:43.912 ] 00:15:43.912 }' 00:15:43.912 23:49:31 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:43.912 23:49:31 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:44.170 23:49:32 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@555 -- # rpc_cmd bdev_raid_get_bdevs online 00:15:44.171 23:49:32 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@555 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:15:44.171 23:49:32 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:44.171 23:49:32 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:44.171 23:49:32 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:44.171 23:49:32 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@555 -- # [[ false == \f\a\l\s\e ]] 00:15:44.171 23:49:32 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@558 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:15:44.171 23:49:32 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@558 -- # jq -r '.[] | .uuid' 00:15:44.171 23:49:32 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:44.171 23:49:32 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:44.171 [2024-11-26 23:49:32.269620] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:44.171 23:49:32 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:44.431 23:49:32 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@558 -- # '[' 1c889e41-1db4-4bd6-a360-0c7a9febaf1f '!=' 1c889e41-1db4-4bd6-a360-0c7a9febaf1f ']' 00:15:44.431 23:49:32 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@563 -- # killprocess 98657 00:15:44.431 23:49:32 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@954 -- # '[' -z 98657 ']' 00:15:44.431 23:49:32 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@958 -- # kill -0 98657 00:15:44.431 23:49:32 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@959 -- # uname 00:15:44.431 23:49:32 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:15:44.431 23:49:32 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 98657 00:15:44.431 killing process with pid 98657 00:15:44.431 23:49:32 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:15:44.431 23:49:32 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:15:44.431 23:49:32 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@972 -- # echo 'killing process with pid 98657' 00:15:44.431 23:49:32 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@973 -- # kill 98657 00:15:44.431 [2024-11-26 23:49:32.357211] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:15:44.431 [2024-11-26 23:49:32.357286] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:44.431 [2024-11-26 23:49:32.357322] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:44.431 [2024-11-26 23:49:32.357330] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002380 name raid_bdev1, state offline 00:15:44.431 23:49:32 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@978 -- # wait 98657 00:15:44.431 [2024-11-26 23:49:32.380111] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:15:44.691 23:49:32 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@565 -- # return 0 00:15:44.692 ************************************ 00:15:44.692 END TEST raid_superblock_test_md_interleaved 00:15:44.692 ************************************ 00:15:44.692 00:15:44.692 real 0m5.034s 00:15:44.692 user 0m8.244s 00:15:44.692 sys 0m1.102s 00:15:44.692 23:49:32 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@1130 -- # xtrace_disable 00:15:44.692 23:49:32 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:44.692 23:49:32 bdev_raid -- bdev/bdev_raid.sh@1013 -- # run_test raid_rebuild_test_sb_md_interleaved raid_rebuild_test raid1 2 true false false 00:15:44.692 23:49:32 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:15:44.692 23:49:32 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:15:44.692 23:49:32 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:15:44.692 ************************************ 00:15:44.692 START TEST raid_rebuild_test_sb_md_interleaved 00:15:44.692 ************************************ 00:15:44.692 23:49:32 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@1129 -- # raid_rebuild_test raid1 2 true false false 00:15:44.692 23:49:32 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:15:44.692 23:49:32 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=2 00:15:44.692 23:49:32 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@571 -- # local superblock=true 00:15:44.692 23:49:32 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:15:44.692 23:49:32 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@573 -- # local verify=false 00:15:44.692 23:49:32 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:15:44.692 23:49:32 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:15:44.692 23:49:32 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:15:44.692 23:49:32 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:15:44.692 23:49:32 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:15:44.692 23:49:32 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:15:44.692 23:49:32 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:15:44.692 23:49:32 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:15:44.692 23:49:32 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:15:44.692 23:49:32 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:15:44.692 23:49:32 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:15:44.692 23:49:32 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@576 -- # local strip_size 00:15:44.692 23:49:32 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@577 -- # local create_arg 00:15:44.692 23:49:32 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:15:44.692 23:49:32 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@579 -- # local data_offset 00:15:44.692 23:49:32 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:15:44.692 23:49:32 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:15:44.692 23:49:32 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@592 -- # '[' true = true ']' 00:15:44.692 23:49:32 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@593 -- # create_arg+=' -s' 00:15:44.692 23:49:32 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@597 -- # raid_pid=98974 00:15:44.692 23:49:32 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@598 -- # waitforlisten 98974 00:15:44.692 23:49:32 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:15:44.692 23:49:32 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@835 -- # '[' -z 98974 ']' 00:15:44.692 23:49:32 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:15:44.692 23:49:32 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@840 -- # local max_retries=100 00:15:44.692 23:49:32 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:15:44.692 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:15:44.692 23:49:32 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@844 -- # xtrace_disable 00:15:44.692 23:49:32 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:44.692 [2024-11-26 23:49:32.762104] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:15:44.692 [2024-11-26 23:49:32.762336] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.ealI/O size of 3145728 is greater than zero copy threshold (65536). 00:15:44.692 Zero copy mechanism will not be used. 00:15:44.692 :6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid98974 ] 00:15:44.952 [2024-11-26 23:49:32.906269] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:15:44.952 [2024-11-26 23:49:32.930618] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:15:44.952 [2024-11-26 23:49:32.972973] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:15:44.952 [2024-11-26 23:49:32.973086] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:15:45.523 23:49:33 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:15:45.523 23:49:33 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@868 -- # return 0 00:15:45.523 23:49:33 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:15:45.523 23:49:33 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -i -b BaseBdev1_malloc 00:15:45.523 23:49:33 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:45.523 23:49:33 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:45.523 BaseBdev1_malloc 00:15:45.523 23:49:33 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:45.523 23:49:33 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:15:45.523 23:49:33 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:45.523 23:49:33 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:45.523 [2024-11-26 23:49:33.588263] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:15:45.523 [2024-11-26 23:49:33.588337] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:45.523 [2024-11-26 23:49:33.588373] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:15:45.523 [2024-11-26 23:49:33.588382] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:45.523 [2024-11-26 23:49:33.590250] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:45.523 [2024-11-26 23:49:33.590287] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:15:45.523 BaseBdev1 00:15:45.523 23:49:33 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:45.523 23:49:33 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:15:45.523 23:49:33 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -i -b BaseBdev2_malloc 00:15:45.523 23:49:33 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:45.523 23:49:33 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:45.523 BaseBdev2_malloc 00:15:45.523 23:49:33 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:45.523 23:49:33 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:15:45.523 23:49:33 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:45.523 23:49:33 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:45.523 [2024-11-26 23:49:33.616809] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:15:45.523 [2024-11-26 23:49:33.616860] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:45.523 [2024-11-26 23:49:33.616880] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:15:45.523 [2024-11-26 23:49:33.616890] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:45.523 [2024-11-26 23:49:33.618730] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:45.523 [2024-11-26 23:49:33.618771] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:15:45.523 BaseBdev2 00:15:45.523 23:49:33 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:45.523 23:49:33 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -i -b spare_malloc 00:15:45.523 23:49:33 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:45.523 23:49:33 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:45.783 spare_malloc 00:15:45.783 23:49:33 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:45.783 23:49:33 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:15:45.783 23:49:33 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:45.783 23:49:33 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:45.783 spare_delay 00:15:45.783 23:49:33 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:45.783 23:49:33 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:15:45.783 23:49:33 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:45.783 23:49:33 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:45.783 [2024-11-26 23:49:33.674869] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:15:45.783 [2024-11-26 23:49:33.674941] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:45.783 [2024-11-26 23:49:33.674972] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:15:45.783 [2024-11-26 23:49:33.674986] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:45.783 [2024-11-26 23:49:33.677894] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:45.783 [2024-11-26 23:49:33.677944] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:15:45.783 spare 00:15:45.783 23:49:33 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:45.783 23:49:33 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 00:15:45.783 23:49:33 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:45.783 23:49:33 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:45.783 [2024-11-26 23:49:33.686909] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:15:45.783 [2024-11-26 23:49:33.688905] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:15:45.783 [2024-11-26 23:49:33.689143] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:15:45.783 [2024-11-26 23:49:33.689163] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4128 00:15:45.783 [2024-11-26 23:49:33.689274] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:15:45.783 [2024-11-26 23:49:33.689377] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:15:45.783 [2024-11-26 23:49:33.689392] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:15:45.783 [2024-11-26 23:49:33.689465] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:45.783 23:49:33 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:45.783 23:49:33 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:15:45.783 23:49:33 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:45.783 23:49:33 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:45.783 23:49:33 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:45.783 23:49:33 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:45.783 23:49:33 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:45.783 23:49:33 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:45.783 23:49:33 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:45.783 23:49:33 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:45.783 23:49:33 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:45.783 23:49:33 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:45.783 23:49:33 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:45.783 23:49:33 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:45.784 23:49:33 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:45.784 23:49:33 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:45.784 23:49:33 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:45.784 "name": "raid_bdev1", 00:15:45.784 "uuid": "75398e08-848d-4ed8-8997-7b85200fc38b", 00:15:45.784 "strip_size_kb": 0, 00:15:45.784 "state": "online", 00:15:45.784 "raid_level": "raid1", 00:15:45.784 "superblock": true, 00:15:45.784 "num_base_bdevs": 2, 00:15:45.784 "num_base_bdevs_discovered": 2, 00:15:45.784 "num_base_bdevs_operational": 2, 00:15:45.784 "base_bdevs_list": [ 00:15:45.784 { 00:15:45.784 "name": "BaseBdev1", 00:15:45.784 "uuid": "ca77bd2c-bb03-5fb8-94c5-f6fa9d8f4391", 00:15:45.784 "is_configured": true, 00:15:45.784 "data_offset": 256, 00:15:45.784 "data_size": 7936 00:15:45.784 }, 00:15:45.784 { 00:15:45.784 "name": "BaseBdev2", 00:15:45.784 "uuid": "4aef7dd1-2ba9-5077-94c0-1d9f69fa03c3", 00:15:45.784 "is_configured": true, 00:15:45.784 "data_offset": 256, 00:15:45.784 "data_size": 7936 00:15:45.784 } 00:15:45.784 ] 00:15:45.784 }' 00:15:45.784 23:49:33 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:45.784 23:49:33 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:46.043 23:49:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:15:46.044 23:49:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:15:46.044 23:49:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:46.044 23:49:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:46.304 [2024-11-26 23:49:34.174259] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:46.304 23:49:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:46.304 23:49:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=7936 00:15:46.304 23:49:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:46.304 23:49:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:46.304 23:49:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:46.304 23:49:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:15:46.304 23:49:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:46.304 23:49:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@619 -- # data_offset=256 00:15:46.304 23:49:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:15:46.304 23:49:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@624 -- # '[' false = true ']' 00:15:46.304 23:49:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:15:46.304 23:49:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:46.304 23:49:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:46.304 [2024-11-26 23:49:34.269831] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:15:46.304 23:49:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:46.304 23:49:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:15:46.304 23:49:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:46.304 23:49:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:46.304 23:49:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:46.304 23:49:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:46.304 23:49:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:15:46.304 23:49:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:46.304 23:49:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:46.304 23:49:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:46.304 23:49:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:46.304 23:49:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:46.304 23:49:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:46.304 23:49:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:46.304 23:49:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:46.304 23:49:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:46.304 23:49:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:46.304 "name": "raid_bdev1", 00:15:46.304 "uuid": "75398e08-848d-4ed8-8997-7b85200fc38b", 00:15:46.304 "strip_size_kb": 0, 00:15:46.304 "state": "online", 00:15:46.304 "raid_level": "raid1", 00:15:46.304 "superblock": true, 00:15:46.304 "num_base_bdevs": 2, 00:15:46.305 "num_base_bdevs_discovered": 1, 00:15:46.305 "num_base_bdevs_operational": 1, 00:15:46.305 "base_bdevs_list": [ 00:15:46.305 { 00:15:46.305 "name": null, 00:15:46.305 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:46.305 "is_configured": false, 00:15:46.305 "data_offset": 0, 00:15:46.305 "data_size": 7936 00:15:46.305 }, 00:15:46.305 { 00:15:46.305 "name": "BaseBdev2", 00:15:46.305 "uuid": "4aef7dd1-2ba9-5077-94c0-1d9f69fa03c3", 00:15:46.305 "is_configured": true, 00:15:46.305 "data_offset": 256, 00:15:46.305 "data_size": 7936 00:15:46.305 } 00:15:46.305 ] 00:15:46.305 }' 00:15:46.305 23:49:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:46.305 23:49:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:46.876 23:49:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:15:46.876 23:49:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:46.876 23:49:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:46.876 [2024-11-26 23:49:34.753013] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:15:46.876 [2024-11-26 23:49:34.756704] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:15:46.876 [2024-11-26 23:49:34.758596] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:15:46.876 23:49:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:46.876 23:49:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@647 -- # sleep 1 00:15:47.815 23:49:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:47.815 23:49:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:47.815 23:49:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:47.815 23:49:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:47.815 23:49:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:47.815 23:49:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:47.815 23:49:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:47.815 23:49:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:47.815 23:49:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:47.815 23:49:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:47.815 23:49:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:47.815 "name": "raid_bdev1", 00:15:47.815 "uuid": "75398e08-848d-4ed8-8997-7b85200fc38b", 00:15:47.815 "strip_size_kb": 0, 00:15:47.815 "state": "online", 00:15:47.815 "raid_level": "raid1", 00:15:47.815 "superblock": true, 00:15:47.815 "num_base_bdevs": 2, 00:15:47.815 "num_base_bdevs_discovered": 2, 00:15:47.815 "num_base_bdevs_operational": 2, 00:15:47.815 "process": { 00:15:47.815 "type": "rebuild", 00:15:47.815 "target": "spare", 00:15:47.815 "progress": { 00:15:47.815 "blocks": 2560, 00:15:47.815 "percent": 32 00:15:47.815 } 00:15:47.815 }, 00:15:47.815 "base_bdevs_list": [ 00:15:47.815 { 00:15:47.815 "name": "spare", 00:15:47.815 "uuid": "12a2047b-fa25-50a7-904a-f6f0ffc5a284", 00:15:47.815 "is_configured": true, 00:15:47.815 "data_offset": 256, 00:15:47.815 "data_size": 7936 00:15:47.815 }, 00:15:47.815 { 00:15:47.815 "name": "BaseBdev2", 00:15:47.815 "uuid": "4aef7dd1-2ba9-5077-94c0-1d9f69fa03c3", 00:15:47.815 "is_configured": true, 00:15:47.815 "data_offset": 256, 00:15:47.815 "data_size": 7936 00:15:47.815 } 00:15:47.815 ] 00:15:47.815 }' 00:15:47.815 23:49:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:47.815 23:49:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:47.815 23:49:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:47.815 23:49:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:47.815 23:49:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:15:47.815 23:49:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:47.815 23:49:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:47.815 [2024-11-26 23:49:35.917328] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:48.075 [2024-11-26 23:49:35.963238] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:15:48.075 [2024-11-26 23:49:35.963365] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:48.075 [2024-11-26 23:49:35.963405] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:48.075 [2024-11-26 23:49:35.963446] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:15:48.075 23:49:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:48.075 23:49:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:15:48.075 23:49:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:48.075 23:49:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:48.075 23:49:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:48.075 23:49:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:48.075 23:49:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:15:48.075 23:49:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:48.075 23:49:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:48.075 23:49:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:48.075 23:49:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:48.075 23:49:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:48.075 23:49:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:48.075 23:49:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:48.076 23:49:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:48.076 23:49:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:48.076 23:49:36 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:48.076 "name": "raid_bdev1", 00:15:48.076 "uuid": "75398e08-848d-4ed8-8997-7b85200fc38b", 00:15:48.076 "strip_size_kb": 0, 00:15:48.076 "state": "online", 00:15:48.076 "raid_level": "raid1", 00:15:48.076 "superblock": true, 00:15:48.076 "num_base_bdevs": 2, 00:15:48.076 "num_base_bdevs_discovered": 1, 00:15:48.076 "num_base_bdevs_operational": 1, 00:15:48.076 "base_bdevs_list": [ 00:15:48.076 { 00:15:48.076 "name": null, 00:15:48.076 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:48.076 "is_configured": false, 00:15:48.076 "data_offset": 0, 00:15:48.076 "data_size": 7936 00:15:48.076 }, 00:15:48.076 { 00:15:48.076 "name": "BaseBdev2", 00:15:48.076 "uuid": "4aef7dd1-2ba9-5077-94c0-1d9f69fa03c3", 00:15:48.076 "is_configured": true, 00:15:48.076 "data_offset": 256, 00:15:48.076 "data_size": 7936 00:15:48.076 } 00:15:48.076 ] 00:15:48.076 }' 00:15:48.076 23:49:36 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:48.076 23:49:36 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:48.335 23:49:36 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:15:48.335 23:49:36 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:48.335 23:49:36 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:15:48.335 23:49:36 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=none 00:15:48.335 23:49:36 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:48.335 23:49:36 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:48.335 23:49:36 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:48.335 23:49:36 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:48.335 23:49:36 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:48.335 23:49:36 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:48.595 23:49:36 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:48.595 "name": "raid_bdev1", 00:15:48.595 "uuid": "75398e08-848d-4ed8-8997-7b85200fc38b", 00:15:48.595 "strip_size_kb": 0, 00:15:48.595 "state": "online", 00:15:48.595 "raid_level": "raid1", 00:15:48.595 "superblock": true, 00:15:48.595 "num_base_bdevs": 2, 00:15:48.595 "num_base_bdevs_discovered": 1, 00:15:48.595 "num_base_bdevs_operational": 1, 00:15:48.595 "base_bdevs_list": [ 00:15:48.595 { 00:15:48.595 "name": null, 00:15:48.595 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:48.595 "is_configured": false, 00:15:48.595 "data_offset": 0, 00:15:48.595 "data_size": 7936 00:15:48.595 }, 00:15:48.595 { 00:15:48.595 "name": "BaseBdev2", 00:15:48.595 "uuid": "4aef7dd1-2ba9-5077-94c0-1d9f69fa03c3", 00:15:48.595 "is_configured": true, 00:15:48.595 "data_offset": 256, 00:15:48.595 "data_size": 7936 00:15:48.595 } 00:15:48.595 ] 00:15:48.595 }' 00:15:48.595 23:49:36 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:48.595 23:49:36 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:15:48.595 23:49:36 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:48.595 23:49:36 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:15:48.595 23:49:36 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:15:48.595 23:49:36 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:48.595 23:49:36 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:48.595 [2024-11-26 23:49:36.570490] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:15:48.595 [2024-11-26 23:49:36.573669] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:15:48.595 [2024-11-26 23:49:36.575413] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:15:48.595 23:49:36 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:48.595 23:49:36 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@663 -- # sleep 1 00:15:49.535 23:49:37 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:49.535 23:49:37 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:49.535 23:49:37 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:49.536 23:49:37 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:49.536 23:49:37 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:49.536 23:49:37 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:49.536 23:49:37 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:49.536 23:49:37 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:49.536 23:49:37 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:49.536 23:49:37 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:49.536 23:49:37 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:49.536 "name": "raid_bdev1", 00:15:49.536 "uuid": "75398e08-848d-4ed8-8997-7b85200fc38b", 00:15:49.536 "strip_size_kb": 0, 00:15:49.536 "state": "online", 00:15:49.536 "raid_level": "raid1", 00:15:49.536 "superblock": true, 00:15:49.536 "num_base_bdevs": 2, 00:15:49.536 "num_base_bdevs_discovered": 2, 00:15:49.536 "num_base_bdevs_operational": 2, 00:15:49.536 "process": { 00:15:49.536 "type": "rebuild", 00:15:49.536 "target": "spare", 00:15:49.536 "progress": { 00:15:49.536 "blocks": 2560, 00:15:49.536 "percent": 32 00:15:49.536 } 00:15:49.536 }, 00:15:49.536 "base_bdevs_list": [ 00:15:49.536 { 00:15:49.536 "name": "spare", 00:15:49.536 "uuid": "12a2047b-fa25-50a7-904a-f6f0ffc5a284", 00:15:49.536 "is_configured": true, 00:15:49.536 "data_offset": 256, 00:15:49.536 "data_size": 7936 00:15:49.536 }, 00:15:49.536 { 00:15:49.536 "name": "BaseBdev2", 00:15:49.536 "uuid": "4aef7dd1-2ba9-5077-94c0-1d9f69fa03c3", 00:15:49.536 "is_configured": true, 00:15:49.536 "data_offset": 256, 00:15:49.536 "data_size": 7936 00:15:49.536 } 00:15:49.536 ] 00:15:49.536 }' 00:15:49.536 23:49:37 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:49.796 23:49:37 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:49.796 23:49:37 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:49.796 23:49:37 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:49.796 23:49:37 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@666 -- # '[' true = true ']' 00:15:49.796 23:49:37 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@666 -- # '[' = false ']' 00:15:49.796 /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh: line 666: [: =: unary operator expected 00:15:49.796 23:49:37 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=2 00:15:49.796 23:49:37 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:15:49.796 23:49:37 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@693 -- # '[' 2 -gt 2 ']' 00:15:49.796 23:49:37 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@706 -- # local timeout=609 00:15:49.796 23:49:37 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:49.796 23:49:37 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:49.796 23:49:37 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:49.796 23:49:37 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:49.796 23:49:37 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:49.796 23:49:37 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:49.796 23:49:37 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:49.796 23:49:37 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:49.796 23:49:37 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:49.796 23:49:37 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:49.796 23:49:37 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:49.796 23:49:37 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:49.796 "name": "raid_bdev1", 00:15:49.796 "uuid": "75398e08-848d-4ed8-8997-7b85200fc38b", 00:15:49.796 "strip_size_kb": 0, 00:15:49.796 "state": "online", 00:15:49.796 "raid_level": "raid1", 00:15:49.796 "superblock": true, 00:15:49.796 "num_base_bdevs": 2, 00:15:49.796 "num_base_bdevs_discovered": 2, 00:15:49.796 "num_base_bdevs_operational": 2, 00:15:49.796 "process": { 00:15:49.796 "type": "rebuild", 00:15:49.796 "target": "spare", 00:15:49.796 "progress": { 00:15:49.796 "blocks": 2816, 00:15:49.796 "percent": 35 00:15:49.796 } 00:15:49.796 }, 00:15:49.796 "base_bdevs_list": [ 00:15:49.796 { 00:15:49.796 "name": "spare", 00:15:49.796 "uuid": "12a2047b-fa25-50a7-904a-f6f0ffc5a284", 00:15:49.796 "is_configured": true, 00:15:49.796 "data_offset": 256, 00:15:49.796 "data_size": 7936 00:15:49.796 }, 00:15:49.796 { 00:15:49.796 "name": "BaseBdev2", 00:15:49.796 "uuid": "4aef7dd1-2ba9-5077-94c0-1d9f69fa03c3", 00:15:49.796 "is_configured": true, 00:15:49.796 "data_offset": 256, 00:15:49.796 "data_size": 7936 00:15:49.796 } 00:15:49.796 ] 00:15:49.796 }' 00:15:49.796 23:49:37 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:49.796 23:49:37 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:49.796 23:49:37 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:49.796 23:49:37 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:49.796 23:49:37 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:50.760 23:49:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:50.760 23:49:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:50.760 23:49:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:50.760 23:49:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:50.760 23:49:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:50.760 23:49:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:51.030 23:49:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:51.031 23:49:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:51.031 23:49:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:51.031 23:49:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:51.031 23:49:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:51.031 23:49:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:51.031 "name": "raid_bdev1", 00:15:51.031 "uuid": "75398e08-848d-4ed8-8997-7b85200fc38b", 00:15:51.031 "strip_size_kb": 0, 00:15:51.031 "state": "online", 00:15:51.031 "raid_level": "raid1", 00:15:51.031 "superblock": true, 00:15:51.031 "num_base_bdevs": 2, 00:15:51.031 "num_base_bdevs_discovered": 2, 00:15:51.031 "num_base_bdevs_operational": 2, 00:15:51.031 "process": { 00:15:51.031 "type": "rebuild", 00:15:51.031 "target": "spare", 00:15:51.031 "progress": { 00:15:51.031 "blocks": 5888, 00:15:51.031 "percent": 74 00:15:51.031 } 00:15:51.031 }, 00:15:51.031 "base_bdevs_list": [ 00:15:51.031 { 00:15:51.031 "name": "spare", 00:15:51.031 "uuid": "12a2047b-fa25-50a7-904a-f6f0ffc5a284", 00:15:51.031 "is_configured": true, 00:15:51.031 "data_offset": 256, 00:15:51.031 "data_size": 7936 00:15:51.031 }, 00:15:51.031 { 00:15:51.031 "name": "BaseBdev2", 00:15:51.031 "uuid": "4aef7dd1-2ba9-5077-94c0-1d9f69fa03c3", 00:15:51.031 "is_configured": true, 00:15:51.031 "data_offset": 256, 00:15:51.031 "data_size": 7936 00:15:51.031 } 00:15:51.031 ] 00:15:51.031 }' 00:15:51.031 23:49:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:51.031 23:49:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:51.031 23:49:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:51.031 23:49:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:51.031 23:49:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:51.598 [2024-11-26 23:49:39.685770] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:15:51.598 [2024-11-26 23:49:39.685839] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:15:51.599 [2024-11-26 23:49:39.685946] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:52.169 23:49:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:52.169 23:49:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:52.169 23:49:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:52.169 23:49:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:52.169 23:49:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:52.169 23:49:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:52.169 23:49:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:52.169 23:49:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:52.169 23:49:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:52.169 23:49:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:52.169 23:49:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:52.169 23:49:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:52.169 "name": "raid_bdev1", 00:15:52.169 "uuid": "75398e08-848d-4ed8-8997-7b85200fc38b", 00:15:52.169 "strip_size_kb": 0, 00:15:52.169 "state": "online", 00:15:52.169 "raid_level": "raid1", 00:15:52.169 "superblock": true, 00:15:52.169 "num_base_bdevs": 2, 00:15:52.169 "num_base_bdevs_discovered": 2, 00:15:52.169 "num_base_bdevs_operational": 2, 00:15:52.169 "base_bdevs_list": [ 00:15:52.169 { 00:15:52.169 "name": "spare", 00:15:52.169 "uuid": "12a2047b-fa25-50a7-904a-f6f0ffc5a284", 00:15:52.169 "is_configured": true, 00:15:52.169 "data_offset": 256, 00:15:52.169 "data_size": 7936 00:15:52.169 }, 00:15:52.169 { 00:15:52.169 "name": "BaseBdev2", 00:15:52.169 "uuid": "4aef7dd1-2ba9-5077-94c0-1d9f69fa03c3", 00:15:52.169 "is_configured": true, 00:15:52.169 "data_offset": 256, 00:15:52.169 "data_size": 7936 00:15:52.169 } 00:15:52.169 ] 00:15:52.169 }' 00:15:52.169 23:49:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:52.169 23:49:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:15:52.169 23:49:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:52.169 23:49:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:15:52.169 23:49:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@709 -- # break 00:15:52.169 23:49:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:15:52.169 23:49:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:52.169 23:49:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:15:52.169 23:49:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=none 00:15:52.169 23:49:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:52.169 23:49:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:52.169 23:49:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:52.169 23:49:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:52.169 23:49:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:52.169 23:49:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:52.169 23:49:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:52.169 "name": "raid_bdev1", 00:15:52.169 "uuid": "75398e08-848d-4ed8-8997-7b85200fc38b", 00:15:52.169 "strip_size_kb": 0, 00:15:52.169 "state": "online", 00:15:52.169 "raid_level": "raid1", 00:15:52.169 "superblock": true, 00:15:52.169 "num_base_bdevs": 2, 00:15:52.169 "num_base_bdevs_discovered": 2, 00:15:52.169 "num_base_bdevs_operational": 2, 00:15:52.169 "base_bdevs_list": [ 00:15:52.169 { 00:15:52.169 "name": "spare", 00:15:52.169 "uuid": "12a2047b-fa25-50a7-904a-f6f0ffc5a284", 00:15:52.169 "is_configured": true, 00:15:52.169 "data_offset": 256, 00:15:52.169 "data_size": 7936 00:15:52.169 }, 00:15:52.169 { 00:15:52.169 "name": "BaseBdev2", 00:15:52.169 "uuid": "4aef7dd1-2ba9-5077-94c0-1d9f69fa03c3", 00:15:52.169 "is_configured": true, 00:15:52.169 "data_offset": 256, 00:15:52.169 "data_size": 7936 00:15:52.169 } 00:15:52.169 ] 00:15:52.169 }' 00:15:52.169 23:49:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:52.169 23:49:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:15:52.169 23:49:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:52.169 23:49:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:15:52.169 23:49:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:15:52.169 23:49:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:52.169 23:49:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:52.169 23:49:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:52.169 23:49:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:52.169 23:49:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:52.169 23:49:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:52.169 23:49:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:52.169 23:49:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:52.169 23:49:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:52.428 23:49:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:52.428 23:49:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:52.428 23:49:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:52.428 23:49:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:52.428 23:49:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:52.428 23:49:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:52.428 "name": "raid_bdev1", 00:15:52.428 "uuid": "75398e08-848d-4ed8-8997-7b85200fc38b", 00:15:52.428 "strip_size_kb": 0, 00:15:52.428 "state": "online", 00:15:52.428 "raid_level": "raid1", 00:15:52.428 "superblock": true, 00:15:52.428 "num_base_bdevs": 2, 00:15:52.428 "num_base_bdevs_discovered": 2, 00:15:52.428 "num_base_bdevs_operational": 2, 00:15:52.428 "base_bdevs_list": [ 00:15:52.428 { 00:15:52.428 "name": "spare", 00:15:52.428 "uuid": "12a2047b-fa25-50a7-904a-f6f0ffc5a284", 00:15:52.428 "is_configured": true, 00:15:52.428 "data_offset": 256, 00:15:52.428 "data_size": 7936 00:15:52.428 }, 00:15:52.428 { 00:15:52.428 "name": "BaseBdev2", 00:15:52.428 "uuid": "4aef7dd1-2ba9-5077-94c0-1d9f69fa03c3", 00:15:52.428 "is_configured": true, 00:15:52.428 "data_offset": 256, 00:15:52.428 "data_size": 7936 00:15:52.428 } 00:15:52.428 ] 00:15:52.428 }' 00:15:52.428 23:49:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:52.428 23:49:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:52.688 23:49:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:15:52.688 23:49:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:52.688 23:49:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:52.688 [2024-11-26 23:49:40.711766] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:15:52.688 [2024-11-26 23:49:40.711846] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:15:52.688 [2024-11-26 23:49:40.711957] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:52.688 [2024-11-26 23:49:40.712075] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:52.688 [2024-11-26 23:49:40.712164] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:15:52.688 23:49:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:52.688 23:49:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:52.688 23:49:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@720 -- # jq length 00:15:52.688 23:49:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:52.688 23:49:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:52.688 23:49:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:52.688 23:49:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:15:52.688 23:49:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@722 -- # '[' false = true ']' 00:15:52.688 23:49:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@743 -- # '[' true = true ']' 00:15:52.688 23:49:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@745 -- # rpc_cmd bdev_passthru_delete spare 00:15:52.688 23:49:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:52.688 23:49:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:52.688 23:49:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:52.688 23:49:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@746 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:15:52.688 23:49:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:52.688 23:49:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:52.688 [2024-11-26 23:49:40.783656] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:15:52.688 [2024-11-26 23:49:40.783708] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:52.688 [2024-11-26 23:49:40.783725] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:15:52.688 [2024-11-26 23:49:40.783735] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:52.688 [2024-11-26 23:49:40.785621] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:52.688 [2024-11-26 23:49:40.785662] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:15:52.688 [2024-11-26 23:49:40.785710] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:15:52.688 [2024-11-26 23:49:40.785755] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:15:52.688 [2024-11-26 23:49:40.785839] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:15:52.688 spare 00:15:52.688 23:49:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:52.688 23:49:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@747 -- # rpc_cmd bdev_wait_for_examine 00:15:52.688 23:49:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:52.688 23:49:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:52.948 [2024-11-26 23:49:40.885726] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001580 00:15:52.948 [2024-11-26 23:49:40.885749] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4128 00:15:52.948 [2024-11-26 23:49:40.885830] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000026d0 00:15:52.948 [2024-11-26 23:49:40.885907] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001580 00:15:52.948 [2024-11-26 23:49:40.885919] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001580 00:15:52.948 [2024-11-26 23:49:40.885987] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:52.948 23:49:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:52.948 23:49:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@749 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:15:52.948 23:49:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:52.948 23:49:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:52.948 23:49:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:52.948 23:49:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:52.948 23:49:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:52.948 23:49:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:52.948 23:49:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:52.948 23:49:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:52.948 23:49:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:52.948 23:49:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:52.948 23:49:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:52.948 23:49:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:52.948 23:49:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:52.948 23:49:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:52.948 23:49:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:52.948 "name": "raid_bdev1", 00:15:52.948 "uuid": "75398e08-848d-4ed8-8997-7b85200fc38b", 00:15:52.948 "strip_size_kb": 0, 00:15:52.948 "state": "online", 00:15:52.948 "raid_level": "raid1", 00:15:52.948 "superblock": true, 00:15:52.948 "num_base_bdevs": 2, 00:15:52.948 "num_base_bdevs_discovered": 2, 00:15:52.948 "num_base_bdevs_operational": 2, 00:15:52.948 "base_bdevs_list": [ 00:15:52.948 { 00:15:52.948 "name": "spare", 00:15:52.948 "uuid": "12a2047b-fa25-50a7-904a-f6f0ffc5a284", 00:15:52.948 "is_configured": true, 00:15:52.948 "data_offset": 256, 00:15:52.948 "data_size": 7936 00:15:52.948 }, 00:15:52.948 { 00:15:52.948 "name": "BaseBdev2", 00:15:52.948 "uuid": "4aef7dd1-2ba9-5077-94c0-1d9f69fa03c3", 00:15:52.948 "is_configured": true, 00:15:52.948 "data_offset": 256, 00:15:52.948 "data_size": 7936 00:15:52.948 } 00:15:52.948 ] 00:15:52.948 }' 00:15:52.948 23:49:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:52.948 23:49:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:53.526 23:49:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@750 -- # verify_raid_bdev_process raid_bdev1 none none 00:15:53.526 23:49:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:53.526 23:49:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:15:53.526 23:49:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=none 00:15:53.526 23:49:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:53.526 23:49:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:53.526 23:49:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:53.526 23:49:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:53.526 23:49:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:53.526 23:49:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:53.526 23:49:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:53.526 "name": "raid_bdev1", 00:15:53.526 "uuid": "75398e08-848d-4ed8-8997-7b85200fc38b", 00:15:53.526 "strip_size_kb": 0, 00:15:53.526 "state": "online", 00:15:53.527 "raid_level": "raid1", 00:15:53.527 "superblock": true, 00:15:53.527 "num_base_bdevs": 2, 00:15:53.527 "num_base_bdevs_discovered": 2, 00:15:53.527 "num_base_bdevs_operational": 2, 00:15:53.527 "base_bdevs_list": [ 00:15:53.527 { 00:15:53.527 "name": "spare", 00:15:53.527 "uuid": "12a2047b-fa25-50a7-904a-f6f0ffc5a284", 00:15:53.527 "is_configured": true, 00:15:53.527 "data_offset": 256, 00:15:53.527 "data_size": 7936 00:15:53.527 }, 00:15:53.527 { 00:15:53.527 "name": "BaseBdev2", 00:15:53.527 "uuid": "4aef7dd1-2ba9-5077-94c0-1d9f69fa03c3", 00:15:53.527 "is_configured": true, 00:15:53.527 "data_offset": 256, 00:15:53.527 "data_size": 7936 00:15:53.527 } 00:15:53.527 ] 00:15:53.527 }' 00:15:53.527 23:49:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:53.527 23:49:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:15:53.527 23:49:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:53.527 23:49:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:15:53.527 23:49:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@751 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:53.527 23:49:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@751 -- # jq -r '.[].base_bdevs_list[0].name' 00:15:53.527 23:49:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:53.527 23:49:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:53.527 23:49:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:53.527 23:49:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@751 -- # [[ spare == \s\p\a\r\e ]] 00:15:53.527 23:49:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@754 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:15:53.527 23:49:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:53.527 23:49:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:53.527 [2024-11-26 23:49:41.562472] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:53.527 23:49:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:53.527 23:49:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@755 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:15:53.527 23:49:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:53.527 23:49:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:53.527 23:49:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:53.527 23:49:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:53.527 23:49:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:15:53.527 23:49:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:53.527 23:49:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:53.527 23:49:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:53.527 23:49:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:53.527 23:49:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:53.527 23:49:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:53.527 23:49:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:53.527 23:49:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:53.527 23:49:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:53.527 23:49:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:53.527 "name": "raid_bdev1", 00:15:53.527 "uuid": "75398e08-848d-4ed8-8997-7b85200fc38b", 00:15:53.527 "strip_size_kb": 0, 00:15:53.527 "state": "online", 00:15:53.527 "raid_level": "raid1", 00:15:53.527 "superblock": true, 00:15:53.527 "num_base_bdevs": 2, 00:15:53.527 "num_base_bdevs_discovered": 1, 00:15:53.527 "num_base_bdevs_operational": 1, 00:15:53.527 "base_bdevs_list": [ 00:15:53.527 { 00:15:53.527 "name": null, 00:15:53.527 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:53.527 "is_configured": false, 00:15:53.527 "data_offset": 0, 00:15:53.527 "data_size": 7936 00:15:53.527 }, 00:15:53.527 { 00:15:53.527 "name": "BaseBdev2", 00:15:53.527 "uuid": "4aef7dd1-2ba9-5077-94c0-1d9f69fa03c3", 00:15:53.527 "is_configured": true, 00:15:53.527 "data_offset": 256, 00:15:53.527 "data_size": 7936 00:15:53.527 } 00:15:53.527 ] 00:15:53.527 }' 00:15:53.527 23:49:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:53.527 23:49:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:54.099 23:49:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@756 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:15:54.099 23:49:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:54.099 23:49:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:54.099 [2024-11-26 23:49:42.001727] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:15:54.099 [2024-11-26 23:49:42.001870] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:15:54.099 [2024-11-26 23:49:42.001887] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:15:54.099 [2024-11-26 23:49:42.001921] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:15:54.099 [2024-11-26 23:49:42.005310] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000027a0 00:15:54.099 [2024-11-26 23:49:42.007220] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:15:54.099 23:49:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:54.099 23:49:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@757 -- # sleep 1 00:15:55.039 23:49:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@758 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:55.039 23:49:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:55.039 23:49:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:55.039 23:49:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:55.039 23:49:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:55.039 23:49:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:55.039 23:49:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:55.039 23:49:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:55.039 23:49:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:55.039 23:49:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:55.039 23:49:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:55.039 "name": "raid_bdev1", 00:15:55.039 "uuid": "75398e08-848d-4ed8-8997-7b85200fc38b", 00:15:55.039 "strip_size_kb": 0, 00:15:55.039 "state": "online", 00:15:55.039 "raid_level": "raid1", 00:15:55.039 "superblock": true, 00:15:55.039 "num_base_bdevs": 2, 00:15:55.039 "num_base_bdevs_discovered": 2, 00:15:55.039 "num_base_bdevs_operational": 2, 00:15:55.039 "process": { 00:15:55.039 "type": "rebuild", 00:15:55.039 "target": "spare", 00:15:55.039 "progress": { 00:15:55.039 "blocks": 2560, 00:15:55.039 "percent": 32 00:15:55.039 } 00:15:55.039 }, 00:15:55.039 "base_bdevs_list": [ 00:15:55.039 { 00:15:55.039 "name": "spare", 00:15:55.039 "uuid": "12a2047b-fa25-50a7-904a-f6f0ffc5a284", 00:15:55.039 "is_configured": true, 00:15:55.039 "data_offset": 256, 00:15:55.039 "data_size": 7936 00:15:55.039 }, 00:15:55.039 { 00:15:55.039 "name": "BaseBdev2", 00:15:55.039 "uuid": "4aef7dd1-2ba9-5077-94c0-1d9f69fa03c3", 00:15:55.039 "is_configured": true, 00:15:55.039 "data_offset": 256, 00:15:55.039 "data_size": 7936 00:15:55.039 } 00:15:55.039 ] 00:15:55.039 }' 00:15:55.039 23:49:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:55.039 23:49:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:55.039 23:49:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:55.040 23:49:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:55.040 23:49:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@761 -- # rpc_cmd bdev_passthru_delete spare 00:15:55.040 23:49:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:55.040 23:49:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:55.040 [2024-11-26 23:49:43.145879] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:55.300 [2024-11-26 23:49:43.211072] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:15:55.300 [2024-11-26 23:49:43.211124] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:55.300 [2024-11-26 23:49:43.211139] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:55.300 [2024-11-26 23:49:43.211146] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:15:55.300 23:49:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:55.300 23:49:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@762 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:15:55.300 23:49:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:55.300 23:49:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:55.300 23:49:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:55.300 23:49:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:55.300 23:49:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:15:55.300 23:49:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:55.300 23:49:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:55.300 23:49:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:55.300 23:49:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:55.300 23:49:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:55.300 23:49:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:55.300 23:49:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:55.300 23:49:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:55.300 23:49:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:55.300 23:49:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:55.300 "name": "raid_bdev1", 00:15:55.300 "uuid": "75398e08-848d-4ed8-8997-7b85200fc38b", 00:15:55.300 "strip_size_kb": 0, 00:15:55.300 "state": "online", 00:15:55.300 "raid_level": "raid1", 00:15:55.300 "superblock": true, 00:15:55.300 "num_base_bdevs": 2, 00:15:55.300 "num_base_bdevs_discovered": 1, 00:15:55.300 "num_base_bdevs_operational": 1, 00:15:55.300 "base_bdevs_list": [ 00:15:55.300 { 00:15:55.300 "name": null, 00:15:55.300 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:55.300 "is_configured": false, 00:15:55.300 "data_offset": 0, 00:15:55.300 "data_size": 7936 00:15:55.300 }, 00:15:55.300 { 00:15:55.300 "name": "BaseBdev2", 00:15:55.300 "uuid": "4aef7dd1-2ba9-5077-94c0-1d9f69fa03c3", 00:15:55.300 "is_configured": true, 00:15:55.300 "data_offset": 256, 00:15:55.300 "data_size": 7936 00:15:55.300 } 00:15:55.300 ] 00:15:55.300 }' 00:15:55.300 23:49:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:55.300 23:49:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:55.561 23:49:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@763 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:15:55.561 23:49:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:55.561 23:49:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:55.561 [2024-11-26 23:49:43.670140] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:15:55.561 [2024-11-26 23:49:43.670250] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:55.561 [2024-11-26 23:49:43.670291] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009c80 00:15:55.561 [2024-11-26 23:49:43.670330] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:55.561 [2024-11-26 23:49:43.670533] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:55.561 [2024-11-26 23:49:43.670577] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:15:55.561 [2024-11-26 23:49:43.670649] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:15:55.561 [2024-11-26 23:49:43.670682] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:15:55.561 [2024-11-26 23:49:43.670729] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:15:55.561 [2024-11-26 23:49:43.670819] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:15:55.561 [2024-11-26 23:49:43.673685] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002870 00:15:55.561 [2024-11-26 23:49:43.675567] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:15:55.561 spare 00:15:55.561 23:49:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:55.561 23:49:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@764 -- # sleep 1 00:15:56.945 23:49:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@765 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:56.945 23:49:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:56.945 23:49:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:56.945 23:49:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:56.945 23:49:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:56.945 23:49:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:56.945 23:49:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:56.945 23:49:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:56.945 23:49:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:56.945 23:49:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:56.945 23:49:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:56.945 "name": "raid_bdev1", 00:15:56.945 "uuid": "75398e08-848d-4ed8-8997-7b85200fc38b", 00:15:56.945 "strip_size_kb": 0, 00:15:56.945 "state": "online", 00:15:56.945 "raid_level": "raid1", 00:15:56.945 "superblock": true, 00:15:56.945 "num_base_bdevs": 2, 00:15:56.945 "num_base_bdevs_discovered": 2, 00:15:56.945 "num_base_bdevs_operational": 2, 00:15:56.945 "process": { 00:15:56.945 "type": "rebuild", 00:15:56.945 "target": "spare", 00:15:56.945 "progress": { 00:15:56.945 "blocks": 2560, 00:15:56.945 "percent": 32 00:15:56.945 } 00:15:56.945 }, 00:15:56.945 "base_bdevs_list": [ 00:15:56.945 { 00:15:56.945 "name": "spare", 00:15:56.945 "uuid": "12a2047b-fa25-50a7-904a-f6f0ffc5a284", 00:15:56.945 "is_configured": true, 00:15:56.945 "data_offset": 256, 00:15:56.945 "data_size": 7936 00:15:56.945 }, 00:15:56.945 { 00:15:56.945 "name": "BaseBdev2", 00:15:56.945 "uuid": "4aef7dd1-2ba9-5077-94c0-1d9f69fa03c3", 00:15:56.945 "is_configured": true, 00:15:56.945 "data_offset": 256, 00:15:56.945 "data_size": 7936 00:15:56.945 } 00:15:56.945 ] 00:15:56.945 }' 00:15:56.945 23:49:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:56.945 23:49:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:56.945 23:49:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:56.945 23:49:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:56.945 23:49:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@768 -- # rpc_cmd bdev_passthru_delete spare 00:15:56.945 23:49:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:56.945 23:49:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:56.945 [2024-11-26 23:49:44.814205] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:56.945 [2024-11-26 23:49:44.879403] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:15:56.945 [2024-11-26 23:49:44.879479] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:56.945 [2024-11-26 23:49:44.879493] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:56.945 [2024-11-26 23:49:44.879502] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:15:56.945 23:49:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:56.945 23:49:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@769 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:15:56.945 23:49:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:56.945 23:49:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:56.945 23:49:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:56.945 23:49:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:56.945 23:49:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:15:56.945 23:49:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:56.945 23:49:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:56.945 23:49:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:56.945 23:49:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:56.945 23:49:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:56.945 23:49:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:56.945 23:49:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:56.945 23:49:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:56.945 23:49:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:56.945 23:49:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:56.945 "name": "raid_bdev1", 00:15:56.945 "uuid": "75398e08-848d-4ed8-8997-7b85200fc38b", 00:15:56.945 "strip_size_kb": 0, 00:15:56.945 "state": "online", 00:15:56.945 "raid_level": "raid1", 00:15:56.945 "superblock": true, 00:15:56.945 "num_base_bdevs": 2, 00:15:56.945 "num_base_bdevs_discovered": 1, 00:15:56.945 "num_base_bdevs_operational": 1, 00:15:56.945 "base_bdevs_list": [ 00:15:56.945 { 00:15:56.945 "name": null, 00:15:56.945 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:56.945 "is_configured": false, 00:15:56.945 "data_offset": 0, 00:15:56.945 "data_size": 7936 00:15:56.945 }, 00:15:56.945 { 00:15:56.945 "name": "BaseBdev2", 00:15:56.945 "uuid": "4aef7dd1-2ba9-5077-94c0-1d9f69fa03c3", 00:15:56.945 "is_configured": true, 00:15:56.945 "data_offset": 256, 00:15:56.945 "data_size": 7936 00:15:56.945 } 00:15:56.945 ] 00:15:56.945 }' 00:15:56.945 23:49:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:56.945 23:49:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:57.515 23:49:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@770 -- # verify_raid_bdev_process raid_bdev1 none none 00:15:57.515 23:49:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:57.515 23:49:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:15:57.515 23:49:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=none 00:15:57.515 23:49:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:57.515 23:49:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:57.515 23:49:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:57.515 23:49:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:57.515 23:49:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:57.515 23:49:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:57.515 23:49:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:57.515 "name": "raid_bdev1", 00:15:57.515 "uuid": "75398e08-848d-4ed8-8997-7b85200fc38b", 00:15:57.515 "strip_size_kb": 0, 00:15:57.515 "state": "online", 00:15:57.515 "raid_level": "raid1", 00:15:57.515 "superblock": true, 00:15:57.515 "num_base_bdevs": 2, 00:15:57.515 "num_base_bdevs_discovered": 1, 00:15:57.515 "num_base_bdevs_operational": 1, 00:15:57.515 "base_bdevs_list": [ 00:15:57.515 { 00:15:57.515 "name": null, 00:15:57.515 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:57.515 "is_configured": false, 00:15:57.515 "data_offset": 0, 00:15:57.515 "data_size": 7936 00:15:57.515 }, 00:15:57.515 { 00:15:57.515 "name": "BaseBdev2", 00:15:57.515 "uuid": "4aef7dd1-2ba9-5077-94c0-1d9f69fa03c3", 00:15:57.515 "is_configured": true, 00:15:57.516 "data_offset": 256, 00:15:57.516 "data_size": 7936 00:15:57.516 } 00:15:57.516 ] 00:15:57.516 }' 00:15:57.516 23:49:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:57.516 23:49:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:15:57.516 23:49:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:57.516 23:49:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:15:57.516 23:49:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@773 -- # rpc_cmd bdev_passthru_delete BaseBdev1 00:15:57.516 23:49:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:57.516 23:49:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:57.516 23:49:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:57.516 23:49:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@774 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:15:57.516 23:49:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:57.516 23:49:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:57.516 [2024-11-26 23:49:45.506144] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:15:57.516 [2024-11-26 23:49:45.506197] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:57.516 [2024-11-26 23:49:45.506217] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:15:57.516 [2024-11-26 23:49:45.506227] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:57.516 [2024-11-26 23:49:45.506389] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:57.516 [2024-11-26 23:49:45.506406] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:15:57.516 [2024-11-26 23:49:45.506446] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:15:57.516 [2024-11-26 23:49:45.506458] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:15:57.516 [2024-11-26 23:49:45.506465] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:15:57.516 [2024-11-26 23:49:45.506477] bdev_raid.c:3894:raid_bdev_examine_done: *ERROR*: Failed to examine bdev BaseBdev1: Invalid argument 00:15:57.516 BaseBdev1 00:15:57.516 23:49:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:57.516 23:49:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@775 -- # sleep 1 00:15:58.455 23:49:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@776 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:15:58.455 23:49:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:58.455 23:49:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:58.455 23:49:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:58.455 23:49:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:58.455 23:49:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:15:58.455 23:49:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:58.455 23:49:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:58.455 23:49:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:58.455 23:49:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:58.455 23:49:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:58.455 23:49:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:58.455 23:49:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:58.455 23:49:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:58.455 23:49:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:58.455 23:49:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:58.455 "name": "raid_bdev1", 00:15:58.455 "uuid": "75398e08-848d-4ed8-8997-7b85200fc38b", 00:15:58.455 "strip_size_kb": 0, 00:15:58.455 "state": "online", 00:15:58.455 "raid_level": "raid1", 00:15:58.455 "superblock": true, 00:15:58.455 "num_base_bdevs": 2, 00:15:58.455 "num_base_bdevs_discovered": 1, 00:15:58.455 "num_base_bdevs_operational": 1, 00:15:58.455 "base_bdevs_list": [ 00:15:58.455 { 00:15:58.455 "name": null, 00:15:58.455 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:58.455 "is_configured": false, 00:15:58.455 "data_offset": 0, 00:15:58.455 "data_size": 7936 00:15:58.455 }, 00:15:58.455 { 00:15:58.455 "name": "BaseBdev2", 00:15:58.455 "uuid": "4aef7dd1-2ba9-5077-94c0-1d9f69fa03c3", 00:15:58.455 "is_configured": true, 00:15:58.455 "data_offset": 256, 00:15:58.455 "data_size": 7936 00:15:58.456 } 00:15:58.456 ] 00:15:58.456 }' 00:15:58.456 23:49:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:58.456 23:49:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:59.025 23:49:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@777 -- # verify_raid_bdev_process raid_bdev1 none none 00:15:59.025 23:49:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:59.025 23:49:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:15:59.025 23:49:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=none 00:15:59.025 23:49:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:59.025 23:49:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:59.025 23:49:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:59.025 23:49:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:59.025 23:49:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:59.025 23:49:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:59.025 23:49:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:59.025 "name": "raid_bdev1", 00:15:59.025 "uuid": "75398e08-848d-4ed8-8997-7b85200fc38b", 00:15:59.025 "strip_size_kb": 0, 00:15:59.025 "state": "online", 00:15:59.025 "raid_level": "raid1", 00:15:59.025 "superblock": true, 00:15:59.025 "num_base_bdevs": 2, 00:15:59.025 "num_base_bdevs_discovered": 1, 00:15:59.025 "num_base_bdevs_operational": 1, 00:15:59.025 "base_bdevs_list": [ 00:15:59.025 { 00:15:59.025 "name": null, 00:15:59.025 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:59.025 "is_configured": false, 00:15:59.025 "data_offset": 0, 00:15:59.025 "data_size": 7936 00:15:59.025 }, 00:15:59.025 { 00:15:59.025 "name": "BaseBdev2", 00:15:59.025 "uuid": "4aef7dd1-2ba9-5077-94c0-1d9f69fa03c3", 00:15:59.025 "is_configured": true, 00:15:59.025 "data_offset": 256, 00:15:59.025 "data_size": 7936 00:15:59.025 } 00:15:59.025 ] 00:15:59.025 }' 00:15:59.025 23:49:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:59.025 23:49:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:15:59.025 23:49:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:59.025 23:49:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:15:59.025 23:49:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@778 -- # NOT rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:15:59.025 23:49:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@652 -- # local es=0 00:15:59.025 23:49:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:15:59.025 23:49:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:15:59.025 23:49:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:15:59.025 23:49:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:15:59.025 23:49:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:15:59.025 23:49:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:15:59.025 23:49:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:59.025 23:49:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:15:59.025 [2024-11-26 23:49:47.139353] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:15:59.025 [2024-11-26 23:49:47.139563] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:15:59.025 [2024-11-26 23:49:47.139578] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:15:59.025 request: 00:15:59.025 { 00:15:59.025 "base_bdev": "BaseBdev1", 00:15:59.025 "raid_bdev": "raid_bdev1", 00:15:59.025 "method": "bdev_raid_add_base_bdev", 00:15:59.025 "req_id": 1 00:15:59.025 } 00:15:59.025 Got JSON-RPC error response 00:15:59.025 response: 00:15:59.025 { 00:15:59.025 "code": -22, 00:15:59.025 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:15:59.025 } 00:15:59.025 23:49:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:15:59.026 23:49:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@655 -- # es=1 00:15:59.026 23:49:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:15:59.026 23:49:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:15:59.026 23:49:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:15:59.026 23:49:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@779 -- # sleep 1 00:16:00.407 23:49:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@780 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:00.407 23:49:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:00.407 23:49:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:00.407 23:49:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:00.407 23:49:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:00.407 23:49:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:00.407 23:49:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:00.407 23:49:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:00.407 23:49:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:00.407 23:49:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:00.407 23:49:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:00.407 23:49:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:00.407 23:49:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:00.407 23:49:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:00.407 23:49:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:00.407 23:49:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:00.407 "name": "raid_bdev1", 00:16:00.407 "uuid": "75398e08-848d-4ed8-8997-7b85200fc38b", 00:16:00.407 "strip_size_kb": 0, 00:16:00.407 "state": "online", 00:16:00.407 "raid_level": "raid1", 00:16:00.407 "superblock": true, 00:16:00.407 "num_base_bdevs": 2, 00:16:00.407 "num_base_bdevs_discovered": 1, 00:16:00.407 "num_base_bdevs_operational": 1, 00:16:00.407 "base_bdevs_list": [ 00:16:00.407 { 00:16:00.407 "name": null, 00:16:00.407 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:00.407 "is_configured": false, 00:16:00.407 "data_offset": 0, 00:16:00.407 "data_size": 7936 00:16:00.407 }, 00:16:00.407 { 00:16:00.407 "name": "BaseBdev2", 00:16:00.407 "uuid": "4aef7dd1-2ba9-5077-94c0-1d9f69fa03c3", 00:16:00.407 "is_configured": true, 00:16:00.407 "data_offset": 256, 00:16:00.407 "data_size": 7936 00:16:00.407 } 00:16:00.407 ] 00:16:00.407 }' 00:16:00.407 23:49:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:00.407 23:49:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:00.668 23:49:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@781 -- # verify_raid_bdev_process raid_bdev1 none none 00:16:00.668 23:49:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:00.668 23:49:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:16:00.668 23:49:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=none 00:16:00.668 23:49:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:00.668 23:49:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:00.668 23:49:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:00.668 23:49:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:00.668 23:49:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:00.668 23:49:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:00.668 23:49:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:00.668 "name": "raid_bdev1", 00:16:00.668 "uuid": "75398e08-848d-4ed8-8997-7b85200fc38b", 00:16:00.668 "strip_size_kb": 0, 00:16:00.668 "state": "online", 00:16:00.668 "raid_level": "raid1", 00:16:00.668 "superblock": true, 00:16:00.668 "num_base_bdevs": 2, 00:16:00.668 "num_base_bdevs_discovered": 1, 00:16:00.668 "num_base_bdevs_operational": 1, 00:16:00.668 "base_bdevs_list": [ 00:16:00.668 { 00:16:00.668 "name": null, 00:16:00.668 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:00.668 "is_configured": false, 00:16:00.668 "data_offset": 0, 00:16:00.668 "data_size": 7936 00:16:00.668 }, 00:16:00.668 { 00:16:00.668 "name": "BaseBdev2", 00:16:00.668 "uuid": "4aef7dd1-2ba9-5077-94c0-1d9f69fa03c3", 00:16:00.668 "is_configured": true, 00:16:00.668 "data_offset": 256, 00:16:00.668 "data_size": 7936 00:16:00.668 } 00:16:00.668 ] 00:16:00.668 }' 00:16:00.668 23:49:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:00.668 23:49:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:16:00.668 23:49:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:00.668 23:49:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:16:00.668 23:49:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@784 -- # killprocess 98974 00:16:00.668 23:49:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@954 -- # '[' -z 98974 ']' 00:16:00.668 23:49:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@958 -- # kill -0 98974 00:16:00.668 23:49:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@959 -- # uname 00:16:00.668 23:49:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:16:00.668 23:49:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 98974 00:16:00.668 23:49:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:16:00.668 killing process with pid 98974 00:16:00.668 Received shutdown signal, test time was about 60.000000 seconds 00:16:00.668 00:16:00.668 Latency(us) 00:16:00.668 [2024-11-26T23:49:48.800Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:16:00.668 [2024-11-26T23:49:48.800Z] =================================================================================================================== 00:16:00.668 [2024-11-26T23:49:48.800Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:16:00.668 23:49:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:16:00.668 23:49:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@972 -- # echo 'killing process with pid 98974' 00:16:00.668 23:49:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@973 -- # kill 98974 00:16:00.668 [2024-11-26 23:49:48.795396] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:16:00.668 [2024-11-26 23:49:48.795501] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:00.668 [2024-11-26 23:49:48.795544] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:00.668 [2024-11-26 23:49:48.795566] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state offline 00:16:00.668 23:49:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@978 -- # wait 98974 00:16:00.929 [2024-11-26 23:49:48.827542] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:16:00.929 23:49:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@786 -- # return 0 00:16:00.929 00:16:00.929 real 0m16.351s 00:16:00.929 user 0m22.020s 00:16:00.929 sys 0m1.622s 00:16:00.929 ************************************ 00:16:00.929 END TEST raid_rebuild_test_sb_md_interleaved 00:16:00.929 ************************************ 00:16:00.929 23:49:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@1130 -- # xtrace_disable 00:16:00.929 23:49:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:01.198 23:49:49 bdev_raid -- bdev/bdev_raid.sh@1015 -- # trap - EXIT 00:16:01.198 23:49:49 bdev_raid -- bdev/bdev_raid.sh@1016 -- # cleanup 00:16:01.198 23:49:49 bdev_raid -- bdev/bdev_raid.sh@56 -- # '[' -n 98974 ']' 00:16:01.198 23:49:49 bdev_raid -- bdev/bdev_raid.sh@56 -- # ps -p 98974 00:16:01.198 23:49:49 bdev_raid -- bdev/bdev_raid.sh@60 -- # rm -rf /raidtest 00:16:01.198 00:16:01.198 real 9m50.535s 00:16:01.198 user 14m4.279s 00:16:01.198 sys 1m43.869s 00:16:01.198 ************************************ 00:16:01.198 END TEST bdev_raid 00:16:01.198 ************************************ 00:16:01.198 23:49:49 bdev_raid -- common/autotest_common.sh@1130 -- # xtrace_disable 00:16:01.198 23:49:49 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:16:01.198 23:49:49 -- spdk/autotest.sh@190 -- # run_test spdkcli_raid /home/vagrant/spdk_repo/spdk/test/spdkcli/raid.sh 00:16:01.198 23:49:49 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:16:01.198 23:49:49 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:16:01.198 23:49:49 -- common/autotest_common.sh@10 -- # set +x 00:16:01.198 ************************************ 00:16:01.198 START TEST spdkcli_raid 00:16:01.198 ************************************ 00:16:01.198 23:49:49 spdkcli_raid -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/spdkcli/raid.sh 00:16:01.198 * Looking for test storage... 00:16:01.462 * Found test storage at /home/vagrant/spdk_repo/spdk/test/spdkcli 00:16:01.462 23:49:49 spdkcli_raid -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:16:01.462 23:49:49 spdkcli_raid -- common/autotest_common.sh@1693 -- # lcov --version 00:16:01.462 23:49:49 spdkcli_raid -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:16:01.462 23:49:49 spdkcli_raid -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:16:01.462 23:49:49 spdkcli_raid -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:16:01.462 23:49:49 spdkcli_raid -- scripts/common.sh@333 -- # local ver1 ver1_l 00:16:01.462 23:49:49 spdkcli_raid -- scripts/common.sh@334 -- # local ver2 ver2_l 00:16:01.462 23:49:49 spdkcli_raid -- scripts/common.sh@336 -- # IFS=.-: 00:16:01.462 23:49:49 spdkcli_raid -- scripts/common.sh@336 -- # read -ra ver1 00:16:01.462 23:49:49 spdkcli_raid -- scripts/common.sh@337 -- # IFS=.-: 00:16:01.462 23:49:49 spdkcli_raid -- scripts/common.sh@337 -- # read -ra ver2 00:16:01.462 23:49:49 spdkcli_raid -- scripts/common.sh@338 -- # local 'op=<' 00:16:01.462 23:49:49 spdkcli_raid -- scripts/common.sh@340 -- # ver1_l=2 00:16:01.462 23:49:49 spdkcli_raid -- scripts/common.sh@341 -- # ver2_l=1 00:16:01.462 23:49:49 spdkcli_raid -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:16:01.462 23:49:49 spdkcli_raid -- scripts/common.sh@344 -- # case "$op" in 00:16:01.462 23:49:49 spdkcli_raid -- scripts/common.sh@345 -- # : 1 00:16:01.462 23:49:49 spdkcli_raid -- scripts/common.sh@364 -- # (( v = 0 )) 00:16:01.462 23:49:49 spdkcli_raid -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:16:01.462 23:49:49 spdkcli_raid -- scripts/common.sh@365 -- # decimal 1 00:16:01.462 23:49:49 spdkcli_raid -- scripts/common.sh@353 -- # local d=1 00:16:01.462 23:49:49 spdkcli_raid -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:16:01.462 23:49:49 spdkcli_raid -- scripts/common.sh@355 -- # echo 1 00:16:01.462 23:49:49 spdkcli_raid -- scripts/common.sh@365 -- # ver1[v]=1 00:16:01.462 23:49:49 spdkcli_raid -- scripts/common.sh@366 -- # decimal 2 00:16:01.462 23:49:49 spdkcli_raid -- scripts/common.sh@353 -- # local d=2 00:16:01.462 23:49:49 spdkcli_raid -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:16:01.462 23:49:49 spdkcli_raid -- scripts/common.sh@355 -- # echo 2 00:16:01.462 23:49:49 spdkcli_raid -- scripts/common.sh@366 -- # ver2[v]=2 00:16:01.462 23:49:49 spdkcli_raid -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:16:01.462 23:49:49 spdkcli_raid -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:16:01.462 23:49:49 spdkcli_raid -- scripts/common.sh@368 -- # return 0 00:16:01.462 23:49:49 spdkcli_raid -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:16:01.462 23:49:49 spdkcli_raid -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:16:01.462 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:16:01.462 --rc genhtml_branch_coverage=1 00:16:01.462 --rc genhtml_function_coverage=1 00:16:01.462 --rc genhtml_legend=1 00:16:01.462 --rc geninfo_all_blocks=1 00:16:01.462 --rc geninfo_unexecuted_blocks=1 00:16:01.462 00:16:01.462 ' 00:16:01.462 23:49:49 spdkcli_raid -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:16:01.462 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:16:01.462 --rc genhtml_branch_coverage=1 00:16:01.462 --rc genhtml_function_coverage=1 00:16:01.462 --rc genhtml_legend=1 00:16:01.462 --rc geninfo_all_blocks=1 00:16:01.462 --rc geninfo_unexecuted_blocks=1 00:16:01.462 00:16:01.462 ' 00:16:01.462 23:49:49 spdkcli_raid -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:16:01.462 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:16:01.462 --rc genhtml_branch_coverage=1 00:16:01.462 --rc genhtml_function_coverage=1 00:16:01.462 --rc genhtml_legend=1 00:16:01.462 --rc geninfo_all_blocks=1 00:16:01.462 --rc geninfo_unexecuted_blocks=1 00:16:01.462 00:16:01.462 ' 00:16:01.462 23:49:49 spdkcli_raid -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:16:01.462 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:16:01.462 --rc genhtml_branch_coverage=1 00:16:01.462 --rc genhtml_function_coverage=1 00:16:01.462 --rc genhtml_legend=1 00:16:01.462 --rc geninfo_all_blocks=1 00:16:01.462 --rc geninfo_unexecuted_blocks=1 00:16:01.462 00:16:01.462 ' 00:16:01.462 23:49:49 spdkcli_raid -- spdkcli/raid.sh@9 -- # source /home/vagrant/spdk_repo/spdk/test/spdkcli/common.sh 00:16:01.462 23:49:49 spdkcli_raid -- spdkcli/common.sh@6 -- # spdkcli_job=/home/vagrant/spdk_repo/spdk/test/spdkcli/spdkcli_job.py 00:16:01.462 23:49:49 spdkcli_raid -- spdkcli/common.sh@7 -- # spdk_clear_config_py=/home/vagrant/spdk_repo/spdk/test/json_config/clear_config.py 00:16:01.462 23:49:49 spdkcli_raid -- spdkcli/raid.sh@10 -- # source /home/vagrant/spdk_repo/spdk/test/iscsi_tgt/common.sh 00:16:01.462 23:49:49 spdkcli_raid -- iscsi_tgt/common.sh@9 -- # ISCSI_BRIDGE=iscsi_br 00:16:01.462 23:49:49 spdkcli_raid -- iscsi_tgt/common.sh@10 -- # INITIATOR_INTERFACE=spdk_init_int 00:16:01.462 23:49:49 spdkcli_raid -- iscsi_tgt/common.sh@11 -- # INITIATOR_BRIDGE=init_br 00:16:01.462 23:49:49 spdkcli_raid -- iscsi_tgt/common.sh@12 -- # TARGET_NAMESPACE=spdk_iscsi_ns 00:16:01.462 23:49:49 spdkcli_raid -- iscsi_tgt/common.sh@13 -- # TARGET_NS_CMD=(ip netns exec "$TARGET_NAMESPACE") 00:16:01.462 23:49:49 spdkcli_raid -- iscsi_tgt/common.sh@14 -- # TARGET_INTERFACE=spdk_tgt_int 00:16:01.462 23:49:49 spdkcli_raid -- iscsi_tgt/common.sh@15 -- # TARGET_INTERFACE2=spdk_tgt_int2 00:16:01.462 23:49:49 spdkcli_raid -- iscsi_tgt/common.sh@16 -- # TARGET_BRIDGE=tgt_br 00:16:01.462 23:49:49 spdkcli_raid -- iscsi_tgt/common.sh@17 -- # TARGET_BRIDGE2=tgt_br2 00:16:01.462 23:49:49 spdkcli_raid -- iscsi_tgt/common.sh@20 -- # TARGET_IP=10.0.0.1 00:16:01.462 23:49:49 spdkcli_raid -- iscsi_tgt/common.sh@21 -- # TARGET_IP2=10.0.0.3 00:16:01.462 23:49:49 spdkcli_raid -- iscsi_tgt/common.sh@22 -- # INITIATOR_IP=10.0.0.2 00:16:01.462 23:49:49 spdkcli_raid -- iscsi_tgt/common.sh@23 -- # ISCSI_PORT=3260 00:16:01.462 23:49:49 spdkcli_raid -- iscsi_tgt/common.sh@24 -- # NETMASK=10.0.0.2/32 00:16:01.462 23:49:49 spdkcli_raid -- iscsi_tgt/common.sh@25 -- # INITIATOR_TAG=2 00:16:01.462 23:49:49 spdkcli_raid -- iscsi_tgt/common.sh@26 -- # INITIATOR_NAME=ANY 00:16:01.462 23:49:49 spdkcli_raid -- iscsi_tgt/common.sh@27 -- # PORTAL_TAG=1 00:16:01.462 23:49:49 spdkcli_raid -- iscsi_tgt/common.sh@28 -- # ISCSI_APP=("${TARGET_NS_CMD[@]}" "${ISCSI_APP[@]}") 00:16:01.462 23:49:49 spdkcli_raid -- iscsi_tgt/common.sh@29 -- # ISCSI_TEST_CORE_MASK=0xF 00:16:01.462 23:49:49 spdkcli_raid -- spdkcli/raid.sh@12 -- # MATCH_FILE=spdkcli_raid.test 00:16:01.462 23:49:49 spdkcli_raid -- spdkcli/raid.sh@13 -- # SPDKCLI_BRANCH=/bdevs 00:16:01.462 23:49:49 spdkcli_raid -- spdkcli/raid.sh@14 -- # dirname /home/vagrant/spdk_repo/spdk/test/spdkcli/raid.sh 00:16:01.462 23:49:49 spdkcli_raid -- spdkcli/raid.sh@14 -- # readlink -f /home/vagrant/spdk_repo/spdk/test/spdkcli 00:16:01.462 23:49:49 spdkcli_raid -- spdkcli/raid.sh@14 -- # testdir=/home/vagrant/spdk_repo/spdk/test/spdkcli 00:16:01.462 23:49:49 spdkcli_raid -- spdkcli/raid.sh@15 -- # . /home/vagrant/spdk_repo/spdk/test/spdkcli/common.sh 00:16:01.462 23:49:49 spdkcli_raid -- spdkcli/common.sh@6 -- # spdkcli_job=/home/vagrant/spdk_repo/spdk/test/spdkcli/spdkcli_job.py 00:16:01.462 23:49:49 spdkcli_raid -- spdkcli/common.sh@7 -- # spdk_clear_config_py=/home/vagrant/spdk_repo/spdk/test/json_config/clear_config.py 00:16:01.462 23:49:49 spdkcli_raid -- spdkcli/raid.sh@17 -- # trap cleanup EXIT 00:16:01.462 23:49:49 spdkcli_raid -- spdkcli/raid.sh@19 -- # timing_enter run_spdk_tgt 00:16:01.462 23:49:49 spdkcli_raid -- common/autotest_common.sh@726 -- # xtrace_disable 00:16:01.462 23:49:49 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:16:01.462 23:49:49 spdkcli_raid -- spdkcli/raid.sh@20 -- # run_spdk_tgt 00:16:01.462 23:49:49 spdkcli_raid -- spdkcli/common.sh@27 -- # spdk_tgt_pid=99645 00:16:01.462 23:49:49 spdkcli_raid -- spdkcli/common.sh@26 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x3 -p 0 00:16:01.462 23:49:49 spdkcli_raid -- spdkcli/common.sh@28 -- # waitforlisten 99645 00:16:01.462 23:49:49 spdkcli_raid -- common/autotest_common.sh@835 -- # '[' -z 99645 ']' 00:16:01.462 23:49:49 spdkcli_raid -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:16:01.462 23:49:49 spdkcli_raid -- common/autotest_common.sh@840 -- # local max_retries=100 00:16:01.462 23:49:49 spdkcli_raid -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:16:01.462 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:16:01.462 23:49:49 spdkcli_raid -- common/autotest_common.sh@844 -- # xtrace_disable 00:16:01.462 23:49:49 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:16:01.462 [2024-11-26 23:49:49.574036] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:16:01.462 [2024-11-26 23:49:49.574240] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x3 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid99645 ] 00:16:01.723 [2024-11-26 23:49:49.729210] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 2 00:16:01.723 [2024-11-26 23:49:49.756172] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:16:01.723 [2024-11-26 23:49:49.756263] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:16:02.292 23:49:50 spdkcli_raid -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:16:02.292 23:49:50 spdkcli_raid -- common/autotest_common.sh@868 -- # return 0 00:16:02.292 23:49:50 spdkcli_raid -- spdkcli/raid.sh@21 -- # timing_exit run_spdk_tgt 00:16:02.292 23:49:50 spdkcli_raid -- common/autotest_common.sh@732 -- # xtrace_disable 00:16:02.292 23:49:50 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:16:02.292 23:49:50 spdkcli_raid -- spdkcli/raid.sh@23 -- # timing_enter spdkcli_create_malloc 00:16:02.292 23:49:50 spdkcli_raid -- common/autotest_common.sh@726 -- # xtrace_disable 00:16:02.292 23:49:50 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:16:02.292 23:49:50 spdkcli_raid -- spdkcli/raid.sh@26 -- # /home/vagrant/spdk_repo/spdk/test/spdkcli/spdkcli_job.py ''\''/bdevs/malloc create 8 512 Malloc1'\'' '\''Malloc1'\'' True 00:16:02.292 '\''/bdevs/malloc create 8 512 Malloc2'\'' '\''Malloc2'\'' True 00:16:02.292 ' 00:16:04.196 Executing command: ['/bdevs/malloc create 8 512 Malloc1', 'Malloc1', True] 00:16:04.196 Executing command: ['/bdevs/malloc create 8 512 Malloc2', 'Malloc2', True] 00:16:04.196 23:49:52 spdkcli_raid -- spdkcli/raid.sh@27 -- # timing_exit spdkcli_create_malloc 00:16:04.196 23:49:52 spdkcli_raid -- common/autotest_common.sh@732 -- # xtrace_disable 00:16:04.196 23:49:52 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:16:04.196 23:49:52 spdkcli_raid -- spdkcli/raid.sh@29 -- # timing_enter spdkcli_create_raid 00:16:04.196 23:49:52 spdkcli_raid -- common/autotest_common.sh@726 -- # xtrace_disable 00:16:04.196 23:49:52 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:16:04.196 23:49:52 spdkcli_raid -- spdkcli/raid.sh@31 -- # /home/vagrant/spdk_repo/spdk/test/spdkcli/spdkcli_job.py ''\''/bdevs/raid_volume create testraid 0 "Malloc1 Malloc2" 4'\'' '\''testraid'\'' True 00:16:04.196 ' 00:16:05.131 Executing command: ['/bdevs/raid_volume create testraid 0 "Malloc1 Malloc2" 4', 'testraid', True] 00:16:05.131 23:49:53 spdkcli_raid -- spdkcli/raid.sh@32 -- # timing_exit spdkcli_create_raid 00:16:05.131 23:49:53 spdkcli_raid -- common/autotest_common.sh@732 -- # xtrace_disable 00:16:05.131 23:49:53 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:16:05.389 23:49:53 spdkcli_raid -- spdkcli/raid.sh@34 -- # timing_enter spdkcli_check_match 00:16:05.389 23:49:53 spdkcli_raid -- common/autotest_common.sh@726 -- # xtrace_disable 00:16:05.389 23:49:53 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:16:05.389 23:49:53 spdkcli_raid -- spdkcli/raid.sh@35 -- # check_match 00:16:05.389 23:49:53 spdkcli_raid -- spdkcli/common.sh@44 -- # /home/vagrant/spdk_repo/spdk/scripts/spdkcli.py ll /bdevs 00:16:05.647 23:49:53 spdkcli_raid -- spdkcli/common.sh@45 -- # /home/vagrant/spdk_repo/spdk/test/app/match/match /home/vagrant/spdk_repo/spdk/test/spdkcli/match_files/spdkcli_raid.test.match 00:16:05.904 23:49:53 spdkcli_raid -- spdkcli/common.sh@46 -- # rm -f /home/vagrant/spdk_repo/spdk/test/spdkcli/match_files/spdkcli_raid.test 00:16:05.904 23:49:53 spdkcli_raid -- spdkcli/raid.sh@36 -- # timing_exit spdkcli_check_match 00:16:05.904 23:49:53 spdkcli_raid -- common/autotest_common.sh@732 -- # xtrace_disable 00:16:05.904 23:49:53 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:16:05.904 23:49:53 spdkcli_raid -- spdkcli/raid.sh@38 -- # timing_enter spdkcli_delete_raid 00:16:05.904 23:49:53 spdkcli_raid -- common/autotest_common.sh@726 -- # xtrace_disable 00:16:05.904 23:49:53 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:16:05.904 23:49:53 spdkcli_raid -- spdkcli/raid.sh@40 -- # /home/vagrant/spdk_repo/spdk/test/spdkcli/spdkcli_job.py ''\''/bdevs/raid_volume delete testraid'\'' '\'''\'' True 00:16:05.904 ' 00:16:06.839 Executing command: ['/bdevs/raid_volume delete testraid', '', True] 00:16:06.839 23:49:54 spdkcli_raid -- spdkcli/raid.sh@41 -- # timing_exit spdkcli_delete_raid 00:16:06.839 23:49:54 spdkcli_raid -- common/autotest_common.sh@732 -- # xtrace_disable 00:16:06.839 23:49:54 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:16:07.103 23:49:55 spdkcli_raid -- spdkcli/raid.sh@43 -- # timing_enter spdkcli_delete_malloc 00:16:07.103 23:49:55 spdkcli_raid -- common/autotest_common.sh@726 -- # xtrace_disable 00:16:07.103 23:49:55 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:16:07.103 23:49:55 spdkcli_raid -- spdkcli/raid.sh@46 -- # /home/vagrant/spdk_repo/spdk/test/spdkcli/spdkcli_job.py ''\''/bdevs/malloc delete Malloc1'\'' '\'''\'' True 00:16:07.103 '\''/bdevs/malloc delete Malloc2'\'' '\'''\'' True 00:16:07.103 ' 00:16:08.546 Executing command: ['/bdevs/malloc delete Malloc1', '', True] 00:16:08.546 Executing command: ['/bdevs/malloc delete Malloc2', '', True] 00:16:08.546 23:49:56 spdkcli_raid -- spdkcli/raid.sh@47 -- # timing_exit spdkcli_delete_malloc 00:16:08.546 23:49:56 spdkcli_raid -- common/autotest_common.sh@732 -- # xtrace_disable 00:16:08.546 23:49:56 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:16:08.546 23:49:56 spdkcli_raid -- spdkcli/raid.sh@49 -- # killprocess 99645 00:16:08.546 23:49:56 spdkcli_raid -- common/autotest_common.sh@954 -- # '[' -z 99645 ']' 00:16:08.546 23:49:56 spdkcli_raid -- common/autotest_common.sh@958 -- # kill -0 99645 00:16:08.546 23:49:56 spdkcli_raid -- common/autotest_common.sh@959 -- # uname 00:16:08.546 23:49:56 spdkcli_raid -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:16:08.546 23:49:56 spdkcli_raid -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 99645 00:16:08.546 23:49:56 spdkcli_raid -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:16:08.546 23:49:56 spdkcli_raid -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:16:08.546 killing process with pid 99645 00:16:08.546 23:49:56 spdkcli_raid -- common/autotest_common.sh@972 -- # echo 'killing process with pid 99645' 00:16:08.546 23:49:56 spdkcli_raid -- common/autotest_common.sh@973 -- # kill 99645 00:16:08.546 23:49:56 spdkcli_raid -- common/autotest_common.sh@978 -- # wait 99645 00:16:09.117 23:49:57 spdkcli_raid -- spdkcli/raid.sh@1 -- # cleanup 00:16:09.117 23:49:57 spdkcli_raid -- spdkcli/common.sh@10 -- # '[' -n 99645 ']' 00:16:09.117 23:49:57 spdkcli_raid -- spdkcli/common.sh@11 -- # killprocess 99645 00:16:09.117 Process with pid 99645 is not found 00:16:09.117 23:49:57 spdkcli_raid -- common/autotest_common.sh@954 -- # '[' -z 99645 ']' 00:16:09.117 23:49:57 spdkcli_raid -- common/autotest_common.sh@958 -- # kill -0 99645 00:16:09.117 /home/vagrant/spdk_repo/spdk/test/common/autotest_common.sh: line 958: kill: (99645) - No such process 00:16:09.117 23:49:57 spdkcli_raid -- common/autotest_common.sh@981 -- # echo 'Process with pid 99645 is not found' 00:16:09.117 23:49:57 spdkcli_raid -- spdkcli/common.sh@13 -- # '[' -n '' ']' 00:16:09.117 23:49:57 spdkcli_raid -- spdkcli/common.sh@16 -- # '[' -n '' ']' 00:16:09.117 23:49:57 spdkcli_raid -- spdkcli/common.sh@19 -- # '[' -n '' ']' 00:16:09.117 23:49:57 spdkcli_raid -- spdkcli/common.sh@22 -- # rm -f /home/vagrant/spdk_repo/spdk/test/spdkcli/spdkcli_raid.test /home/vagrant/spdk_repo/spdk/test/spdkcli/match_files/spdkcli_details_vhost.test /tmp/sample_aio 00:16:09.117 00:16:09.117 real 0m7.962s 00:16:09.117 user 0m16.743s 00:16:09.117 sys 0m1.116s 00:16:09.117 23:49:57 spdkcli_raid -- common/autotest_common.sh@1130 -- # xtrace_disable 00:16:09.117 23:49:57 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:16:09.117 ************************************ 00:16:09.117 END TEST spdkcli_raid 00:16:09.117 ************************************ 00:16:09.117 23:49:57 -- spdk/autotest.sh@191 -- # run_test blockdev_raid5f /home/vagrant/spdk_repo/spdk/test/bdev/blockdev.sh raid5f 00:16:09.117 23:49:57 -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:16:09.117 23:49:57 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:16:09.117 23:49:57 -- common/autotest_common.sh@10 -- # set +x 00:16:09.117 ************************************ 00:16:09.117 START TEST blockdev_raid5f 00:16:09.117 ************************************ 00:16:09.117 23:49:57 blockdev_raid5f -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/bdev/blockdev.sh raid5f 00:16:09.377 * Looking for test storage... 00:16:09.377 * Found test storage at /home/vagrant/spdk_repo/spdk/test/bdev 00:16:09.377 23:49:57 blockdev_raid5f -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:16:09.377 23:49:57 blockdev_raid5f -- common/autotest_common.sh@1693 -- # lcov --version 00:16:09.377 23:49:57 blockdev_raid5f -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:16:09.377 23:49:57 blockdev_raid5f -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:16:09.377 23:49:57 blockdev_raid5f -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:16:09.377 23:49:57 blockdev_raid5f -- scripts/common.sh@333 -- # local ver1 ver1_l 00:16:09.377 23:49:57 blockdev_raid5f -- scripts/common.sh@334 -- # local ver2 ver2_l 00:16:09.377 23:49:57 blockdev_raid5f -- scripts/common.sh@336 -- # IFS=.-: 00:16:09.377 23:49:57 blockdev_raid5f -- scripts/common.sh@336 -- # read -ra ver1 00:16:09.377 23:49:57 blockdev_raid5f -- scripts/common.sh@337 -- # IFS=.-: 00:16:09.377 23:49:57 blockdev_raid5f -- scripts/common.sh@337 -- # read -ra ver2 00:16:09.377 23:49:57 blockdev_raid5f -- scripts/common.sh@338 -- # local 'op=<' 00:16:09.377 23:49:57 blockdev_raid5f -- scripts/common.sh@340 -- # ver1_l=2 00:16:09.377 23:49:57 blockdev_raid5f -- scripts/common.sh@341 -- # ver2_l=1 00:16:09.377 23:49:57 blockdev_raid5f -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:16:09.377 23:49:57 blockdev_raid5f -- scripts/common.sh@344 -- # case "$op" in 00:16:09.377 23:49:57 blockdev_raid5f -- scripts/common.sh@345 -- # : 1 00:16:09.377 23:49:57 blockdev_raid5f -- scripts/common.sh@364 -- # (( v = 0 )) 00:16:09.377 23:49:57 blockdev_raid5f -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:16:09.377 23:49:57 blockdev_raid5f -- scripts/common.sh@365 -- # decimal 1 00:16:09.377 23:49:57 blockdev_raid5f -- scripts/common.sh@353 -- # local d=1 00:16:09.377 23:49:57 blockdev_raid5f -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:16:09.377 23:49:57 blockdev_raid5f -- scripts/common.sh@355 -- # echo 1 00:16:09.377 23:49:57 blockdev_raid5f -- scripts/common.sh@365 -- # ver1[v]=1 00:16:09.377 23:49:57 blockdev_raid5f -- scripts/common.sh@366 -- # decimal 2 00:16:09.377 23:49:57 blockdev_raid5f -- scripts/common.sh@353 -- # local d=2 00:16:09.377 23:49:57 blockdev_raid5f -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:16:09.377 23:49:57 blockdev_raid5f -- scripts/common.sh@355 -- # echo 2 00:16:09.377 23:49:57 blockdev_raid5f -- scripts/common.sh@366 -- # ver2[v]=2 00:16:09.377 23:49:57 blockdev_raid5f -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:16:09.377 23:49:57 blockdev_raid5f -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:16:09.377 23:49:57 blockdev_raid5f -- scripts/common.sh@368 -- # return 0 00:16:09.377 23:49:57 blockdev_raid5f -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:16:09.377 23:49:57 blockdev_raid5f -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:16:09.377 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:16:09.377 --rc genhtml_branch_coverage=1 00:16:09.377 --rc genhtml_function_coverage=1 00:16:09.377 --rc genhtml_legend=1 00:16:09.377 --rc geninfo_all_blocks=1 00:16:09.377 --rc geninfo_unexecuted_blocks=1 00:16:09.377 00:16:09.377 ' 00:16:09.377 23:49:57 blockdev_raid5f -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:16:09.377 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:16:09.377 --rc genhtml_branch_coverage=1 00:16:09.377 --rc genhtml_function_coverage=1 00:16:09.377 --rc genhtml_legend=1 00:16:09.377 --rc geninfo_all_blocks=1 00:16:09.377 --rc geninfo_unexecuted_blocks=1 00:16:09.377 00:16:09.377 ' 00:16:09.377 23:49:57 blockdev_raid5f -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:16:09.377 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:16:09.377 --rc genhtml_branch_coverage=1 00:16:09.377 --rc genhtml_function_coverage=1 00:16:09.377 --rc genhtml_legend=1 00:16:09.377 --rc geninfo_all_blocks=1 00:16:09.377 --rc geninfo_unexecuted_blocks=1 00:16:09.377 00:16:09.377 ' 00:16:09.377 23:49:57 blockdev_raid5f -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:16:09.377 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:16:09.377 --rc genhtml_branch_coverage=1 00:16:09.377 --rc genhtml_function_coverage=1 00:16:09.377 --rc genhtml_legend=1 00:16:09.377 --rc geninfo_all_blocks=1 00:16:09.377 --rc geninfo_unexecuted_blocks=1 00:16:09.377 00:16:09.377 ' 00:16:09.377 23:49:57 blockdev_raid5f -- bdev/blockdev.sh@10 -- # source /home/vagrant/spdk_repo/spdk/test/bdev/nbd_common.sh 00:16:09.377 23:49:57 blockdev_raid5f -- bdev/nbd_common.sh@6 -- # set -e 00:16:09.377 23:49:57 blockdev_raid5f -- bdev/blockdev.sh@12 -- # rpc_py=rpc_cmd 00:16:09.377 23:49:57 blockdev_raid5f -- bdev/blockdev.sh@13 -- # conf_file=/home/vagrant/spdk_repo/spdk/test/bdev/bdev.json 00:16:09.377 23:49:57 blockdev_raid5f -- bdev/blockdev.sh@14 -- # nonenclosed_conf_file=/home/vagrant/spdk_repo/spdk/test/bdev/nonenclosed.json 00:16:09.377 23:49:57 blockdev_raid5f -- bdev/blockdev.sh@15 -- # nonarray_conf_file=/home/vagrant/spdk_repo/spdk/test/bdev/nonarray.json 00:16:09.377 23:49:57 blockdev_raid5f -- bdev/blockdev.sh@17 -- # export RPC_PIPE_TIMEOUT=30 00:16:09.377 23:49:57 blockdev_raid5f -- bdev/blockdev.sh@17 -- # RPC_PIPE_TIMEOUT=30 00:16:09.377 23:49:57 blockdev_raid5f -- bdev/blockdev.sh@20 -- # : 00:16:09.377 23:49:57 blockdev_raid5f -- bdev/blockdev.sh@707 -- # QOS_DEV_1=Malloc_0 00:16:09.377 23:49:57 blockdev_raid5f -- bdev/blockdev.sh@708 -- # QOS_DEV_2=Null_1 00:16:09.377 23:49:57 blockdev_raid5f -- bdev/blockdev.sh@709 -- # QOS_RUN_TIME=5 00:16:09.377 23:49:57 blockdev_raid5f -- bdev/blockdev.sh@711 -- # uname -s 00:16:09.377 23:49:57 blockdev_raid5f -- bdev/blockdev.sh@711 -- # '[' Linux = Linux ']' 00:16:09.377 23:49:57 blockdev_raid5f -- bdev/blockdev.sh@713 -- # PRE_RESERVED_MEM=0 00:16:09.377 23:49:57 blockdev_raid5f -- bdev/blockdev.sh@719 -- # test_type=raid5f 00:16:09.377 23:49:57 blockdev_raid5f -- bdev/blockdev.sh@720 -- # crypto_device= 00:16:09.377 23:49:57 blockdev_raid5f -- bdev/blockdev.sh@721 -- # dek= 00:16:09.377 23:49:57 blockdev_raid5f -- bdev/blockdev.sh@722 -- # env_ctx= 00:16:09.377 23:49:57 blockdev_raid5f -- bdev/blockdev.sh@723 -- # wait_for_rpc= 00:16:09.377 23:49:57 blockdev_raid5f -- bdev/blockdev.sh@724 -- # '[' -n '' ']' 00:16:09.377 23:49:57 blockdev_raid5f -- bdev/blockdev.sh@727 -- # [[ raid5f == bdev ]] 00:16:09.377 23:49:57 blockdev_raid5f -- bdev/blockdev.sh@727 -- # [[ raid5f == crypto_* ]] 00:16:09.377 23:49:57 blockdev_raid5f -- bdev/blockdev.sh@730 -- # start_spdk_tgt 00:16:09.377 23:49:57 blockdev_raid5f -- bdev/blockdev.sh@47 -- # spdk_tgt_pid=99903 00:16:09.377 23:49:57 blockdev_raid5f -- bdev/blockdev.sh@46 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt '' '' 00:16:09.377 23:49:57 blockdev_raid5f -- bdev/blockdev.sh@48 -- # trap 'killprocess "$spdk_tgt_pid"; exit 1' SIGINT SIGTERM EXIT 00:16:09.377 23:49:57 blockdev_raid5f -- bdev/blockdev.sh@49 -- # waitforlisten 99903 00:16:09.377 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:16:09.378 23:49:57 blockdev_raid5f -- common/autotest_common.sh@835 -- # '[' -z 99903 ']' 00:16:09.378 23:49:57 blockdev_raid5f -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:16:09.378 23:49:57 blockdev_raid5f -- common/autotest_common.sh@840 -- # local max_retries=100 00:16:09.378 23:49:57 blockdev_raid5f -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:16:09.378 23:49:57 blockdev_raid5f -- common/autotest_common.sh@844 -- # xtrace_disable 00:16:09.378 23:49:57 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:16:09.636 [2024-11-26 23:49:57.596378] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:16:09.636 [2024-11-26 23:49:57.596537] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid99903 ] 00:16:09.636 [2024-11-26 23:49:57.752118] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:16:09.896 [2024-11-26 23:49:57.794560] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:16:10.466 23:49:58 blockdev_raid5f -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:16:10.466 23:49:58 blockdev_raid5f -- common/autotest_common.sh@868 -- # return 0 00:16:10.466 23:49:58 blockdev_raid5f -- bdev/blockdev.sh@731 -- # case "$test_type" in 00:16:10.466 23:49:58 blockdev_raid5f -- bdev/blockdev.sh@763 -- # setup_raid5f_conf 00:16:10.466 23:49:58 blockdev_raid5f -- bdev/blockdev.sh@279 -- # rpc_cmd 00:16:10.466 23:49:58 blockdev_raid5f -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:10.466 23:49:58 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:16:10.466 Malloc0 00:16:10.466 Malloc1 00:16:10.466 Malloc2 00:16:10.466 23:49:58 blockdev_raid5f -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:10.466 23:49:58 blockdev_raid5f -- bdev/blockdev.sh@774 -- # rpc_cmd bdev_wait_for_examine 00:16:10.467 23:49:58 blockdev_raid5f -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:10.467 23:49:58 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:16:10.467 23:49:58 blockdev_raid5f -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:10.467 23:49:58 blockdev_raid5f -- bdev/blockdev.sh@777 -- # cat 00:16:10.467 23:49:58 blockdev_raid5f -- bdev/blockdev.sh@777 -- # rpc_cmd save_subsystem_config -n accel 00:16:10.467 23:49:58 blockdev_raid5f -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:10.467 23:49:58 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:16:10.467 23:49:58 blockdev_raid5f -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:10.467 23:49:58 blockdev_raid5f -- bdev/blockdev.sh@777 -- # rpc_cmd save_subsystem_config -n bdev 00:16:10.467 23:49:58 blockdev_raid5f -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:10.467 23:49:58 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:16:10.467 23:49:58 blockdev_raid5f -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:10.467 23:49:58 blockdev_raid5f -- bdev/blockdev.sh@777 -- # rpc_cmd save_subsystem_config -n iobuf 00:16:10.467 23:49:58 blockdev_raid5f -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:10.467 23:49:58 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:16:10.467 23:49:58 blockdev_raid5f -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:10.467 23:49:58 blockdev_raid5f -- bdev/blockdev.sh@785 -- # mapfile -t bdevs 00:16:10.467 23:49:58 blockdev_raid5f -- bdev/blockdev.sh@785 -- # rpc_cmd bdev_get_bdevs 00:16:10.467 23:49:58 blockdev_raid5f -- bdev/blockdev.sh@785 -- # jq -r '.[] | select(.claimed == false)' 00:16:10.467 23:49:58 blockdev_raid5f -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:10.467 23:49:58 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:16:10.467 23:49:58 blockdev_raid5f -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:10.467 23:49:58 blockdev_raid5f -- bdev/blockdev.sh@786 -- # mapfile -t bdevs_name 00:16:10.727 23:49:58 blockdev_raid5f -- bdev/blockdev.sh@786 -- # printf '%s\n' '{' ' "name": "raid5f",' ' "aliases": [' ' "ba0693a7-890f-4e27-87d5-edfe708395e9"' ' ],' ' "product_name": "Raid Volume",' ' "block_size": 512,' ' "num_blocks": 131072,' ' "uuid": "ba0693a7-890f-4e27-87d5-edfe708395e9",' ' "assigned_rate_limits": {' ' "rw_ios_per_sec": 0,' ' "rw_mbytes_per_sec": 0,' ' "r_mbytes_per_sec": 0,' ' "w_mbytes_per_sec": 0' ' },' ' "claimed": false,' ' "zoned": false,' ' "supported_io_types": {' ' "read": true,' ' "write": true,' ' "unmap": false,' ' "flush": false,' ' "reset": true,' ' "nvme_admin": false,' ' "nvme_io": false,' ' "nvme_io_md": false,' ' "write_zeroes": true,' ' "zcopy": false,' ' "get_zone_info": false,' ' "zone_management": false,' ' "zone_append": false,' ' "compare": false,' ' "compare_and_write": false,' ' "abort": false,' ' "seek_hole": false,' ' "seek_data": false,' ' "copy": false,' ' "nvme_iov_md": false' ' },' ' "driver_specific": {' ' "raid": {' ' "uuid": "ba0693a7-890f-4e27-87d5-edfe708395e9",' ' "strip_size_kb": 2,' ' "state": "online",' ' "raid_level": "raid5f",' ' "superblock": false,' ' "num_base_bdevs": 3,' ' "num_base_bdevs_discovered": 3,' ' "num_base_bdevs_operational": 3,' ' "base_bdevs_list": [' ' {' ' "name": "Malloc0",' ' "uuid": "7123a51f-862c-49a9-af5a-0bf8b025a1e3",' ' "is_configured": true,' ' "data_offset": 0,' ' "data_size": 65536' ' },' ' {' ' "name": "Malloc1",' ' "uuid": "b8891a8f-118b-4b0f-86f1-1be5692fb209",' ' "is_configured": true,' ' "data_offset": 0,' ' "data_size": 65536' ' },' ' {' ' "name": "Malloc2",' ' "uuid": "8c3c9d09-7a50-4d4d-a9cd-5c176f376d6f",' ' "is_configured": true,' ' "data_offset": 0,' ' "data_size": 65536' ' }' ' ]' ' }' ' }' '}' 00:16:10.727 23:49:58 blockdev_raid5f -- bdev/blockdev.sh@786 -- # jq -r .name 00:16:10.727 23:49:58 blockdev_raid5f -- bdev/blockdev.sh@787 -- # bdev_list=("${bdevs_name[@]}") 00:16:10.727 23:49:58 blockdev_raid5f -- bdev/blockdev.sh@789 -- # hello_world_bdev=raid5f 00:16:10.728 23:49:58 blockdev_raid5f -- bdev/blockdev.sh@790 -- # trap - SIGINT SIGTERM EXIT 00:16:10.728 23:49:58 blockdev_raid5f -- bdev/blockdev.sh@791 -- # killprocess 99903 00:16:10.728 23:49:58 blockdev_raid5f -- common/autotest_common.sh@954 -- # '[' -z 99903 ']' 00:16:10.728 23:49:58 blockdev_raid5f -- common/autotest_common.sh@958 -- # kill -0 99903 00:16:10.728 23:49:58 blockdev_raid5f -- common/autotest_common.sh@959 -- # uname 00:16:10.728 23:49:58 blockdev_raid5f -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:16:10.728 23:49:58 blockdev_raid5f -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 99903 00:16:10.728 23:49:58 blockdev_raid5f -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:16:10.728 23:49:58 blockdev_raid5f -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:16:10.728 23:49:58 blockdev_raid5f -- common/autotest_common.sh@972 -- # echo 'killing process with pid 99903' 00:16:10.728 killing process with pid 99903 00:16:10.728 23:49:58 blockdev_raid5f -- common/autotest_common.sh@973 -- # kill 99903 00:16:10.728 23:49:58 blockdev_raid5f -- common/autotest_common.sh@978 -- # wait 99903 00:16:11.308 23:49:59 blockdev_raid5f -- bdev/blockdev.sh@795 -- # trap cleanup SIGINT SIGTERM EXIT 00:16:11.308 23:49:59 blockdev_raid5f -- bdev/blockdev.sh@797 -- # run_test bdev_hello_world /home/vagrant/spdk_repo/spdk/build/examples/hello_bdev --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json -b raid5f '' 00:16:11.308 23:49:59 blockdev_raid5f -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:16:11.308 23:49:59 blockdev_raid5f -- common/autotest_common.sh@1111 -- # xtrace_disable 00:16:11.308 23:49:59 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:16:11.308 ************************************ 00:16:11.308 START TEST bdev_hello_world 00:16:11.309 ************************************ 00:16:11.309 23:49:59 blockdev_raid5f.bdev_hello_world -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/build/examples/hello_bdev --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json -b raid5f '' 00:16:11.569 [2024-11-26 23:49:59.464198] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:16:11.569 [2024-11-26 23:49:59.464389] [ DPDK EAL parameters: hello_bdev --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid99943 ] 00:16:11.569 [2024-11-26 23:49:59.620002] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:16:11.569 [2024-11-26 23:49:59.661404] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:16:11.829 [2024-11-26 23:49:59.912712] hello_bdev.c: 222:hello_start: *NOTICE*: Successfully started the application 00:16:11.829 [2024-11-26 23:49:59.912759] hello_bdev.c: 231:hello_start: *NOTICE*: Opening the bdev raid5f 00:16:11.829 [2024-11-26 23:49:59.912776] hello_bdev.c: 244:hello_start: *NOTICE*: Opening io channel 00:16:11.829 [2024-11-26 23:49:59.913065] hello_bdev.c: 138:hello_write: *NOTICE*: Writing to the bdev 00:16:11.829 [2024-11-26 23:49:59.913207] hello_bdev.c: 117:write_complete: *NOTICE*: bdev io write completed successfully 00:16:11.829 [2024-11-26 23:49:59.913228] hello_bdev.c: 84:hello_read: *NOTICE*: Reading io 00:16:11.829 [2024-11-26 23:49:59.913288] hello_bdev.c: 65:read_complete: *NOTICE*: Read string from bdev : Hello World! 00:16:11.829 00:16:11.829 [2024-11-26 23:49:59.913304] hello_bdev.c: 74:read_complete: *NOTICE*: Stopping app 00:16:12.398 00:16:12.398 real 0m0.901s 00:16:12.398 user 0m0.501s 00:16:12.398 sys 0m0.294s 00:16:12.398 ************************************ 00:16:12.398 END TEST bdev_hello_world 00:16:12.398 ************************************ 00:16:12.398 23:50:00 blockdev_raid5f.bdev_hello_world -- common/autotest_common.sh@1130 -- # xtrace_disable 00:16:12.398 23:50:00 blockdev_raid5f.bdev_hello_world -- common/autotest_common.sh@10 -- # set +x 00:16:12.398 23:50:00 blockdev_raid5f -- bdev/blockdev.sh@798 -- # run_test bdev_bounds bdev_bounds '' 00:16:12.398 23:50:00 blockdev_raid5f -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:16:12.398 23:50:00 blockdev_raid5f -- common/autotest_common.sh@1111 -- # xtrace_disable 00:16:12.398 23:50:00 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:16:12.398 ************************************ 00:16:12.398 START TEST bdev_bounds 00:16:12.398 ************************************ 00:16:12.398 Process bdevio pid: 99974 00:16:12.398 23:50:00 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@1129 -- # bdev_bounds '' 00:16:12.398 23:50:00 blockdev_raid5f.bdev_bounds -- bdev/blockdev.sh@289 -- # bdevio_pid=99974 00:16:12.398 23:50:00 blockdev_raid5f.bdev_bounds -- bdev/blockdev.sh@288 -- # /home/vagrant/spdk_repo/spdk/test/bdev/bdevio/bdevio -w -s 0 --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json '' 00:16:12.399 23:50:00 blockdev_raid5f.bdev_bounds -- bdev/blockdev.sh@290 -- # trap 'cleanup; killprocess $bdevio_pid; exit 1' SIGINT SIGTERM EXIT 00:16:12.399 23:50:00 blockdev_raid5f.bdev_bounds -- bdev/blockdev.sh@291 -- # echo 'Process bdevio pid: 99974' 00:16:12.399 23:50:00 blockdev_raid5f.bdev_bounds -- bdev/blockdev.sh@292 -- # waitforlisten 99974 00:16:12.399 23:50:00 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@835 -- # '[' -z 99974 ']' 00:16:12.399 23:50:00 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:16:12.399 23:50:00 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@840 -- # local max_retries=100 00:16:12.399 23:50:00 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:16:12.399 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:16:12.399 23:50:00 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@844 -- # xtrace_disable 00:16:12.399 23:50:00 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@10 -- # set +x 00:16:12.399 [2024-11-26 23:50:00.441732] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:16:12.399 [2024-11-26 23:50:00.441984] [ DPDK EAL parameters: bdevio --no-shconf -c 0x7 -m 0 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid99974 ] 00:16:12.659 [2024-11-26 23:50:00.600562] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 3 00:16:12.659 [2024-11-26 23:50:00.644289] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:16:12.659 [2024-11-26 23:50:00.647407] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:16:12.659 [2024-11-26 23:50:00.647462] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:16:13.235 23:50:01 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:16:13.235 23:50:01 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@868 -- # return 0 00:16:13.235 23:50:01 blockdev_raid5f.bdev_bounds -- bdev/blockdev.sh@293 -- # /home/vagrant/spdk_repo/spdk/test/bdev/bdevio/tests.py perform_tests 00:16:13.235 I/O targets: 00:16:13.235 raid5f: 131072 blocks of 512 bytes (64 MiB) 00:16:13.235 00:16:13.235 00:16:13.235 CUnit - A unit testing framework for C - Version 2.1-3 00:16:13.235 http://cunit.sourceforge.net/ 00:16:13.235 00:16:13.235 00:16:13.235 Suite: bdevio tests on: raid5f 00:16:13.235 Test: blockdev write read block ...passed 00:16:13.235 Test: blockdev write zeroes read block ...passed 00:16:13.235 Test: blockdev write zeroes read no split ...passed 00:16:13.495 Test: blockdev write zeroes read split ...passed 00:16:13.495 Test: blockdev write zeroes read split partial ...passed 00:16:13.495 Test: blockdev reset ...passed 00:16:13.495 Test: blockdev write read 8 blocks ...passed 00:16:13.495 Test: blockdev write read size > 128k ...passed 00:16:13.495 Test: blockdev write read invalid size ...passed 00:16:13.496 Test: blockdev write read offset + nbytes == size of blockdev ...passed 00:16:13.496 Test: blockdev write read offset + nbytes > size of blockdev ...passed 00:16:13.496 Test: blockdev write read max offset ...passed 00:16:13.496 Test: blockdev write read 2 blocks on overlapped address offset ...passed 00:16:13.496 Test: blockdev writev readv 8 blocks ...passed 00:16:13.496 Test: blockdev writev readv 30 x 1block ...passed 00:16:13.496 Test: blockdev writev readv block ...passed 00:16:13.496 Test: blockdev writev readv size > 128k ...passed 00:16:13.496 Test: blockdev writev readv size > 128k in two iovs ...passed 00:16:13.496 Test: blockdev comparev and writev ...passed 00:16:13.496 Test: blockdev nvme passthru rw ...passed 00:16:13.496 Test: blockdev nvme passthru vendor specific ...passed 00:16:13.496 Test: blockdev nvme admin passthru ...passed 00:16:13.496 Test: blockdev copy ...passed 00:16:13.496 00:16:13.496 Run Summary: Type Total Ran Passed Failed Inactive 00:16:13.496 suites 1 1 n/a 0 0 00:16:13.496 tests 23 23 23 0 0 00:16:13.496 asserts 130 130 130 0 n/a 00:16:13.496 00:16:13.496 Elapsed time = 0.356 seconds 00:16:13.496 0 00:16:13.496 23:50:01 blockdev_raid5f.bdev_bounds -- bdev/blockdev.sh@294 -- # killprocess 99974 00:16:13.496 23:50:01 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@954 -- # '[' -z 99974 ']' 00:16:13.496 23:50:01 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@958 -- # kill -0 99974 00:16:13.496 23:50:01 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@959 -- # uname 00:16:13.496 23:50:01 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:16:13.496 23:50:01 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 99974 00:16:13.496 23:50:01 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:16:13.496 23:50:01 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:16:13.496 23:50:01 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@972 -- # echo 'killing process with pid 99974' 00:16:13.496 killing process with pid 99974 00:16:13.496 23:50:01 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@973 -- # kill 99974 00:16:13.496 23:50:01 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@978 -- # wait 99974 00:16:14.067 23:50:01 blockdev_raid5f.bdev_bounds -- bdev/blockdev.sh@295 -- # trap - SIGINT SIGTERM EXIT 00:16:14.067 00:16:14.067 real 0m1.586s 00:16:14.067 user 0m3.725s 00:16:14.067 sys 0m0.422s 00:16:14.067 23:50:01 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@1130 -- # xtrace_disable 00:16:14.067 23:50:01 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@10 -- # set +x 00:16:14.067 ************************************ 00:16:14.067 END TEST bdev_bounds 00:16:14.067 ************************************ 00:16:14.067 23:50:01 blockdev_raid5f -- bdev/blockdev.sh@799 -- # run_test bdev_nbd nbd_function_test /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json raid5f '' 00:16:14.067 23:50:01 blockdev_raid5f -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:16:14.067 23:50:01 blockdev_raid5f -- common/autotest_common.sh@1111 -- # xtrace_disable 00:16:14.067 23:50:01 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:16:14.067 ************************************ 00:16:14.067 START TEST bdev_nbd 00:16:14.067 ************************************ 00:16:14.067 23:50:02 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@1129 -- # nbd_function_test /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json raid5f '' 00:16:14.067 23:50:02 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@299 -- # uname -s 00:16:14.067 23:50:02 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@299 -- # [[ Linux == Linux ]] 00:16:14.067 23:50:02 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@301 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:16:14.067 23:50:02 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@302 -- # local conf=/home/vagrant/spdk_repo/spdk/test/bdev/bdev.json 00:16:14.067 23:50:02 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@303 -- # bdev_all=('raid5f') 00:16:14.067 23:50:02 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@303 -- # local bdev_all 00:16:14.067 23:50:02 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@304 -- # local bdev_num=1 00:16:14.067 23:50:02 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@308 -- # [[ -e /sys/module/nbd ]] 00:16:14.067 23:50:02 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@310 -- # nbd_all=('/dev/nbd0' '/dev/nbd1' '/dev/nbd10' '/dev/nbd11' '/dev/nbd12' '/dev/nbd13' '/dev/nbd14' '/dev/nbd15' '/dev/nbd2' '/dev/nbd3' '/dev/nbd4' '/dev/nbd5' '/dev/nbd6' '/dev/nbd7' '/dev/nbd8' '/dev/nbd9') 00:16:14.067 23:50:02 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@310 -- # local nbd_all 00:16:14.067 23:50:02 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@311 -- # bdev_num=1 00:16:14.067 23:50:02 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@313 -- # nbd_list=('/dev/nbd0') 00:16:14.067 23:50:02 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@313 -- # local nbd_list 00:16:14.067 23:50:02 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@314 -- # bdev_list=('raid5f') 00:16:14.067 23:50:02 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@314 -- # local bdev_list 00:16:14.067 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock... 00:16:14.067 23:50:02 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@317 -- # nbd_pid=100028 00:16:14.067 23:50:02 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@318 -- # trap 'cleanup; killprocess $nbd_pid' SIGINT SIGTERM EXIT 00:16:14.067 23:50:02 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@316 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -r /var/tmp/spdk-nbd.sock -i 0 --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json '' 00:16:14.067 23:50:02 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@319 -- # waitforlisten 100028 /var/tmp/spdk-nbd.sock 00:16:14.067 23:50:02 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@835 -- # '[' -z 100028 ']' 00:16:14.067 23:50:02 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk-nbd.sock 00:16:14.067 23:50:02 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@840 -- # local max_retries=100 00:16:14.067 23:50:02 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock...' 00:16:14.067 23:50:02 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@844 -- # xtrace_disable 00:16:14.067 23:50:02 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@10 -- # set +x 00:16:14.067 [2024-11-26 23:50:02.117198] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:16:14.067 [2024-11-26 23:50:02.117325] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:16:14.328 [2024-11-26 23:50:02.276624] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:16:14.328 [2024-11-26 23:50:02.318500] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:16:14.899 23:50:02 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:16:14.899 23:50:02 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@868 -- # return 0 00:16:14.899 23:50:02 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@321 -- # nbd_rpc_start_stop_verify /var/tmp/spdk-nbd.sock raid5f 00:16:14.899 23:50:02 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@113 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:16:14.899 23:50:02 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@114 -- # bdev_list=('raid5f') 00:16:14.899 23:50:02 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@114 -- # local bdev_list 00:16:14.899 23:50:02 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@116 -- # nbd_start_disks_without_nbd_idx /var/tmp/spdk-nbd.sock raid5f 00:16:14.899 23:50:02 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@22 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:16:14.899 23:50:02 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@23 -- # bdev_list=('raid5f') 00:16:14.899 23:50:02 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@23 -- # local bdev_list 00:16:14.899 23:50:02 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@24 -- # local i 00:16:14.899 23:50:02 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@25 -- # local nbd_device 00:16:14.899 23:50:02 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@27 -- # (( i = 0 )) 00:16:14.899 23:50:02 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@27 -- # (( i < 1 )) 00:16:14.899 23:50:02 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@28 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk raid5f 00:16:15.177 23:50:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@28 -- # nbd_device=/dev/nbd0 00:16:15.177 23:50:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@30 -- # basename /dev/nbd0 00:16:15.177 23:50:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@30 -- # waitfornbd nbd0 00:16:15.177 23:50:03 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:16:15.177 23:50:03 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@873 -- # local i 00:16:15.177 23:50:03 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:16:15.177 23:50:03 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:16:15.177 23:50:03 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:16:15.177 23:50:03 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@877 -- # break 00:16:15.177 23:50:03 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:16:15.177 23:50:03 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:16:15.177 23:50:03 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:16:15.177 1+0 records in 00:16:15.177 1+0 records out 00:16:15.177 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000425555 s, 9.6 MB/s 00:16:15.177 23:50:03 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:16:15.177 23:50:03 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@890 -- # size=4096 00:16:15.177 23:50:03 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:16:15.177 23:50:03 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:16:15.177 23:50:03 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@893 -- # return 0 00:16:15.177 23:50:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@27 -- # (( i++ )) 00:16:15.177 23:50:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@27 -- # (( i < 1 )) 00:16:15.177 23:50:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@118 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:16:15.437 23:50:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@118 -- # nbd_disks_json='[ 00:16:15.437 { 00:16:15.437 "nbd_device": "/dev/nbd0", 00:16:15.437 "bdev_name": "raid5f" 00:16:15.437 } 00:16:15.437 ]' 00:16:15.437 23:50:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@119 -- # nbd_disks_name=($(echo "${nbd_disks_json}" | jq -r '.[] | .nbd_device')) 00:16:15.437 23:50:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@119 -- # echo '[ 00:16:15.437 { 00:16:15.437 "nbd_device": "/dev/nbd0", 00:16:15.437 "bdev_name": "raid5f" 00:16:15.437 } 00:16:15.437 ]' 00:16:15.437 23:50:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@119 -- # jq -r '.[] | .nbd_device' 00:16:15.437 23:50:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@120 -- # nbd_stop_disks /var/tmp/spdk-nbd.sock /dev/nbd0 00:16:15.437 23:50:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:16:15.437 23:50:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:16:15.437 23:50:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@50 -- # local nbd_list 00:16:15.437 23:50:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@51 -- # local i 00:16:15.437 23:50:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:16:15.437 23:50:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd0 00:16:15.697 23:50:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:16:15.697 23:50:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:16:15.698 23:50:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:16:15.698 23:50:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:16:15.698 23:50:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:16:15.698 23:50:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:16:15.698 23:50:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@41 -- # break 00:16:15.698 23:50:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@45 -- # return 0 00:16:15.698 23:50:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@122 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:16:15.698 23:50:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:16:15.698 23:50:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:16:15.958 23:50:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:16:15.958 23:50:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@64 -- # echo '[]' 00:16:15.958 23:50:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:16:15.958 23:50:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:16:15.958 23:50:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # echo '' 00:16:15.958 23:50:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:16:15.958 23:50:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # true 00:16:15.958 23:50:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # count=0 00:16:15.958 23:50:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@66 -- # echo 0 00:16:15.958 23:50:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@122 -- # count=0 00:16:15.958 23:50:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@123 -- # '[' 0 -ne 0 ']' 00:16:15.958 23:50:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@127 -- # return 0 00:16:15.958 23:50:03 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@322 -- # nbd_rpc_data_verify /var/tmp/spdk-nbd.sock raid5f /dev/nbd0 00:16:15.958 23:50:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@90 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:16:15.958 23:50:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@91 -- # bdev_list=('raid5f') 00:16:15.958 23:50:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@91 -- # local bdev_list 00:16:15.958 23:50:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@92 -- # nbd_list=('/dev/nbd0') 00:16:15.958 23:50:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@92 -- # local nbd_list 00:16:15.958 23:50:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@94 -- # nbd_start_disks /var/tmp/spdk-nbd.sock raid5f /dev/nbd0 00:16:15.958 23:50:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:16:15.958 23:50:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@10 -- # bdev_list=('raid5f') 00:16:15.958 23:50:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@10 -- # local bdev_list 00:16:15.958 23:50:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:16:15.958 23:50:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@11 -- # local nbd_list 00:16:15.958 23:50:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@12 -- # local i 00:16:15.958 23:50:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:16:15.958 23:50:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:16:15.958 23:50:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk raid5f /dev/nbd0 00:16:16.218 /dev/nbd0 00:16:16.218 23:50:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:16:16.218 23:50:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:16:16.218 23:50:04 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:16:16.218 23:50:04 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@873 -- # local i 00:16:16.218 23:50:04 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:16:16.218 23:50:04 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:16:16.218 23:50:04 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:16:16.218 23:50:04 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@877 -- # break 00:16:16.218 23:50:04 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:16:16.218 23:50:04 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:16:16.218 23:50:04 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:16:16.218 1+0 records in 00:16:16.218 1+0 records out 00:16:16.218 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000519983 s, 7.9 MB/s 00:16:16.218 23:50:04 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:16:16.218 23:50:04 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@890 -- # size=4096 00:16:16.218 23:50:04 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:16:16.218 23:50:04 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:16:16.218 23:50:04 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@893 -- # return 0 00:16:16.218 23:50:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:16:16.218 23:50:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:16:16.218 23:50:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@95 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:16:16.218 23:50:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:16:16.218 23:50:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:16:16.477 23:50:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[ 00:16:16.477 { 00:16:16.477 "nbd_device": "/dev/nbd0", 00:16:16.477 "bdev_name": "raid5f" 00:16:16.477 } 00:16:16.477 ]' 00:16:16.477 23:50:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@64 -- # echo '[ 00:16:16.477 { 00:16:16.477 "nbd_device": "/dev/nbd0", 00:16:16.477 "bdev_name": "raid5f" 00:16:16.477 } 00:16:16.477 ]' 00:16:16.477 23:50:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:16:16.477 23:50:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@64 -- # nbd_disks_name=/dev/nbd0 00:16:16.477 23:50:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # echo /dev/nbd0 00:16:16.477 23:50:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:16:16.477 23:50:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # count=1 00:16:16.477 23:50:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@66 -- # echo 1 00:16:16.477 23:50:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@95 -- # count=1 00:16:16.477 23:50:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@96 -- # '[' 1 -ne 1 ']' 00:16:16.477 23:50:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@100 -- # nbd_dd_data_verify /dev/nbd0 write 00:16:16.477 23:50:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0') 00:16:16.477 23:50:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@70 -- # local nbd_list 00:16:16.477 23:50:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@71 -- # local operation=write 00:16:16.477 23:50:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@72 -- # local tmp_file=/home/vagrant/spdk_repo/spdk/test/bdev/nbdrandtest 00:16:16.477 23:50:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@74 -- # '[' write = write ']' 00:16:16.477 23:50:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@76 -- # dd if=/dev/urandom of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdrandtest bs=4096 count=256 00:16:16.477 256+0 records in 00:16:16.477 256+0 records out 00:16:16.477 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0128684 s, 81.5 MB/s 00:16:16.477 23:50:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:16:16.477 23:50:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@78 -- # dd if=/home/vagrant/spdk_repo/spdk/test/bdev/nbdrandtest of=/dev/nbd0 bs=4096 count=256 oflag=direct 00:16:16.477 256+0 records in 00:16:16.477 256+0 records out 00:16:16.477 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0307705 s, 34.1 MB/s 00:16:16.477 23:50:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@101 -- # nbd_dd_data_verify /dev/nbd0 verify 00:16:16.477 23:50:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0') 00:16:16.477 23:50:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@70 -- # local nbd_list 00:16:16.477 23:50:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@71 -- # local operation=verify 00:16:16.477 23:50:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@72 -- # local tmp_file=/home/vagrant/spdk_repo/spdk/test/bdev/nbdrandtest 00:16:16.477 23:50:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@74 -- # '[' verify = write ']' 00:16:16.477 23:50:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@80 -- # '[' verify = verify ']' 00:16:16.477 23:50:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:16:16.477 23:50:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /home/vagrant/spdk_repo/spdk/test/bdev/nbdrandtest /dev/nbd0 00:16:16.477 23:50:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@85 -- # rm /home/vagrant/spdk_repo/spdk/test/bdev/nbdrandtest 00:16:16.477 23:50:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@103 -- # nbd_stop_disks /var/tmp/spdk-nbd.sock /dev/nbd0 00:16:16.477 23:50:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:16:16.477 23:50:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:16:16.477 23:50:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@50 -- # local nbd_list 00:16:16.477 23:50:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@51 -- # local i 00:16:16.477 23:50:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:16:16.478 23:50:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd0 00:16:16.737 23:50:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:16:16.737 23:50:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:16:16.737 23:50:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:16:16.737 23:50:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:16:16.737 23:50:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:16:16.737 23:50:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:16:16.737 23:50:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@41 -- # break 00:16:16.737 23:50:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@45 -- # return 0 00:16:16.737 23:50:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@104 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:16:16.737 23:50:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:16:16.737 23:50:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:16:16.997 23:50:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:16:16.997 23:50:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:16:16.997 23:50:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@64 -- # echo '[]' 00:16:16.997 23:50:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:16:16.997 23:50:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # echo '' 00:16:16.997 23:50:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:16:16.997 23:50:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # true 00:16:16.997 23:50:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # count=0 00:16:16.997 23:50:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@66 -- # echo 0 00:16:16.997 23:50:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@104 -- # count=0 00:16:16.997 23:50:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@105 -- # '[' 0 -ne 0 ']' 00:16:16.997 23:50:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@109 -- # return 0 00:16:16.997 23:50:04 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@323 -- # nbd_with_lvol_verify /var/tmp/spdk-nbd.sock /dev/nbd0 00:16:16.997 23:50:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@131 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:16:16.997 23:50:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@132 -- # local nbd=/dev/nbd0 00:16:16.997 23:50:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@134 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create -b malloc_lvol_verify 16 512 00:16:17.256 malloc_lvol_verify 00:16:17.256 23:50:05 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@135 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_lvol_create_lvstore malloc_lvol_verify lvs 00:16:17.256 3602de05-f6b9-4c5e-967c-0913c53b1d40 00:16:17.514 23:50:05 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@136 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_lvol_create lvol 4 -l lvs 00:16:17.514 df790e04-0bf0-40be-a7c2-ecd7f67ce067 00:16:17.514 23:50:05 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@137 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk lvs/lvol /dev/nbd0 00:16:17.773 /dev/nbd0 00:16:17.773 23:50:05 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@139 -- # wait_for_nbd_set_capacity /dev/nbd0 00:16:17.773 23:50:05 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@146 -- # local nbd=nbd0 00:16:17.773 23:50:05 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@148 -- # [[ -e /sys/block/nbd0/size ]] 00:16:17.773 23:50:05 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@150 -- # (( 8192 == 0 )) 00:16:17.773 23:50:05 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@141 -- # mkfs.ext4 /dev/nbd0 00:16:17.773 mke2fs 1.47.0 (5-Feb-2023) 00:16:17.773 Discarding device blocks: 0/4096 done 00:16:17.773 Creating filesystem with 4096 1k blocks and 1024 inodes 00:16:17.773 00:16:17.773 Allocating group tables: 0/1 done 00:16:17.773 Writing inode tables: 0/1 done 00:16:17.773 Creating journal (1024 blocks): done 00:16:17.773 Writing superblocks and filesystem accounting information: 0/1 done 00:16:17.773 00:16:17.773 23:50:05 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@142 -- # nbd_stop_disks /var/tmp/spdk-nbd.sock /dev/nbd0 00:16:17.773 23:50:05 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:16:17.773 23:50:05 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:16:17.773 23:50:05 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@50 -- # local nbd_list 00:16:17.773 23:50:05 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@51 -- # local i 00:16:17.773 23:50:05 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:16:17.773 23:50:05 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd0 00:16:18.033 23:50:05 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:16:18.033 23:50:05 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:16:18.033 23:50:05 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:16:18.033 23:50:05 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:16:18.033 23:50:05 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:16:18.033 23:50:05 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:16:18.033 23:50:05 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@41 -- # break 00:16:18.033 23:50:05 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@45 -- # return 0 00:16:18.033 23:50:05 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@325 -- # killprocess 100028 00:16:18.033 23:50:05 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@954 -- # '[' -z 100028 ']' 00:16:18.033 23:50:05 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@958 -- # kill -0 100028 00:16:18.033 23:50:05 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@959 -- # uname 00:16:18.033 23:50:06 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:16:18.033 23:50:06 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 100028 00:16:18.033 23:50:06 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:16:18.033 23:50:06 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:16:18.033 killing process with pid 100028 00:16:18.033 23:50:06 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@972 -- # echo 'killing process with pid 100028' 00:16:18.033 23:50:06 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@973 -- # kill 100028 00:16:18.033 23:50:06 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@978 -- # wait 100028 00:16:18.602 23:50:06 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@326 -- # trap - SIGINT SIGTERM EXIT 00:16:18.602 00:16:18.602 real 0m4.424s 00:16:18.602 user 0m6.224s 00:16:18.602 sys 0m1.371s 00:16:18.602 23:50:06 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@1130 -- # xtrace_disable 00:16:18.602 23:50:06 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@10 -- # set +x 00:16:18.602 ************************************ 00:16:18.602 END TEST bdev_nbd 00:16:18.602 ************************************ 00:16:18.602 23:50:06 blockdev_raid5f -- bdev/blockdev.sh@800 -- # [[ y == y ]] 00:16:18.602 23:50:06 blockdev_raid5f -- bdev/blockdev.sh@801 -- # '[' raid5f = nvme ']' 00:16:18.602 23:50:06 blockdev_raid5f -- bdev/blockdev.sh@801 -- # '[' raid5f = gpt ']' 00:16:18.602 23:50:06 blockdev_raid5f -- bdev/blockdev.sh@805 -- # run_test bdev_fio fio_test_suite '' 00:16:18.602 23:50:06 blockdev_raid5f -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:16:18.602 23:50:06 blockdev_raid5f -- common/autotest_common.sh@1111 -- # xtrace_disable 00:16:18.602 23:50:06 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:16:18.602 ************************************ 00:16:18.602 START TEST bdev_fio 00:16:18.602 ************************************ 00:16:18.602 23:50:06 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1129 -- # fio_test_suite '' 00:16:18.602 23:50:06 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@330 -- # local env_context 00:16:18.602 23:50:06 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@334 -- # pushd /home/vagrant/spdk_repo/spdk/test/bdev 00:16:18.602 /home/vagrant/spdk_repo/spdk/test/bdev /home/vagrant/spdk_repo/spdk 00:16:18.602 23:50:06 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@335 -- # trap 'rm -f ./*.state; popd; exit 1' SIGINT SIGTERM EXIT 00:16:18.602 23:50:06 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@338 -- # echo '' 00:16:18.602 23:50:06 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@338 -- # sed s/--env-context=// 00:16:18.602 23:50:06 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@338 -- # env_context= 00:16:18.602 23:50:06 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@339 -- # fio_config_gen /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio verify AIO '' 00:16:18.602 23:50:06 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1284 -- # local config_file=/home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio 00:16:18.602 23:50:06 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1285 -- # local workload=verify 00:16:18.602 23:50:06 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1286 -- # local bdev_type=AIO 00:16:18.602 23:50:06 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1287 -- # local env_context= 00:16:18.602 23:50:06 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1288 -- # local fio_dir=/usr/src/fio 00:16:18.602 23:50:06 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1290 -- # '[' -e /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio ']' 00:16:18.602 23:50:06 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1295 -- # '[' -z verify ']' 00:16:18.602 23:50:06 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1299 -- # '[' -n '' ']' 00:16:18.602 23:50:06 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1303 -- # touch /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio 00:16:18.602 23:50:06 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1305 -- # cat 00:16:18.602 23:50:06 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1317 -- # '[' verify == verify ']' 00:16:18.602 23:50:06 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1318 -- # cat 00:16:18.602 23:50:06 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1327 -- # '[' AIO == AIO ']' 00:16:18.602 23:50:06 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1328 -- # /usr/src/fio/fio --version 00:16:18.602 23:50:06 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1328 -- # [[ fio-3.35 == *\f\i\o\-\3* ]] 00:16:18.602 23:50:06 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1329 -- # echo serialize_overlap=1 00:16:18.602 23:50:06 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@340 -- # for b in "${bdevs_name[@]}" 00:16:18.602 23:50:06 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@341 -- # echo '[job_raid5f]' 00:16:18.602 23:50:06 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@342 -- # echo filename=raid5f 00:16:18.602 23:50:06 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@346 -- # local 'fio_params=--ioengine=spdk_bdev --iodepth=8 --bs=4k --runtime=10 /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio --verify_state_save=0 --spdk_json_conf=/home/vagrant/spdk_repo/spdk/test/bdev/bdev.json' 00:16:18.602 23:50:06 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@348 -- # run_test bdev_fio_rw_verify fio_bdev --ioengine=spdk_bdev --iodepth=8 --bs=4k --runtime=10 /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio --verify_state_save=0 --spdk_json_conf=/home/vagrant/spdk_repo/spdk/test/bdev/bdev.json --spdk_mem=0 --aux-path=/home/vagrant/spdk_repo/spdk/../output 00:16:18.602 23:50:06 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1105 -- # '[' 11 -le 1 ']' 00:16:18.602 23:50:06 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1111 -- # xtrace_disable 00:16:18.602 23:50:06 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@10 -- # set +x 00:16:18.602 ************************************ 00:16:18.602 START TEST bdev_fio_rw_verify 00:16:18.602 ************************************ 00:16:18.602 23:50:06 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1129 -- # fio_bdev --ioengine=spdk_bdev --iodepth=8 --bs=4k --runtime=10 /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio --verify_state_save=0 --spdk_json_conf=/home/vagrant/spdk_repo/spdk/test/bdev/bdev.json --spdk_mem=0 --aux-path=/home/vagrant/spdk_repo/spdk/../output 00:16:18.603 23:50:06 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1360 -- # fio_plugin /home/vagrant/spdk_repo/spdk/build/fio/spdk_bdev --ioengine=spdk_bdev --iodepth=8 --bs=4k --runtime=10 /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio --verify_state_save=0 --spdk_json_conf=/home/vagrant/spdk_repo/spdk/test/bdev/bdev.json --spdk_mem=0 --aux-path=/home/vagrant/spdk_repo/spdk/../output 00:16:18.603 23:50:06 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1341 -- # local fio_dir=/usr/src/fio 00:16:18.603 23:50:06 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1343 -- # sanitizers=('libasan' 'libclang_rt.asan') 00:16:18.603 23:50:06 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1343 -- # local sanitizers 00:16:18.603 23:50:06 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1344 -- # local plugin=/home/vagrant/spdk_repo/spdk/build/fio/spdk_bdev 00:16:18.603 23:50:06 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1345 -- # shift 00:16:18.603 23:50:06 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1347 -- # local asan_lib= 00:16:18.603 23:50:06 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1348 -- # for sanitizer in "${sanitizers[@]}" 00:16:18.603 23:50:06 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1349 -- # grep libasan 00:16:18.603 23:50:06 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1349 -- # ldd /home/vagrant/spdk_repo/spdk/build/fio/spdk_bdev 00:16:18.603 23:50:06 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1349 -- # awk '{print $3}' 00:16:18.863 23:50:06 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1349 -- # asan_lib=/usr/lib64/libasan.so.8 00:16:18.863 23:50:06 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1350 -- # [[ -n /usr/lib64/libasan.so.8 ]] 00:16:18.863 23:50:06 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1351 -- # break 00:16:18.863 23:50:06 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1356 -- # LD_PRELOAD='/usr/lib64/libasan.so.8 /home/vagrant/spdk_repo/spdk/build/fio/spdk_bdev' 00:16:18.863 23:50:06 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1356 -- # /usr/src/fio/fio --ioengine=spdk_bdev --iodepth=8 --bs=4k --runtime=10 /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio --verify_state_save=0 --spdk_json_conf=/home/vagrant/spdk_repo/spdk/test/bdev/bdev.json --spdk_mem=0 --aux-path=/home/vagrant/spdk_repo/spdk/../output 00:16:18.863 job_raid5f: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=spdk_bdev, iodepth=8 00:16:18.863 fio-3.35 00:16:18.863 Starting 1 thread 00:16:31.078 00:16:31.078 job_raid5f: (groupid=0, jobs=1): err= 0: pid=100218: Tue Nov 26 23:50:17 2024 00:16:31.078 read: IOPS=12.6k, BW=49.3MiB/s (51.7MB/s)(493MiB/10000msec) 00:16:31.078 slat (nsec): min=17605, max=44037, avg=18878.45, stdev=1549.92 00:16:31.078 clat (usec): min=11, max=289, avg=129.36, stdev=44.36 00:16:31.078 lat (usec): min=30, max=308, avg=148.24, stdev=44.54 00:16:31.078 clat percentiles (usec): 00:16:31.078 | 50.000th=[ 135], 99.000th=[ 206], 99.900th=[ 231], 99.990th=[ 260], 00:16:31.078 | 99.999th=[ 273] 00:16:31.078 write: IOPS=13.2k, BW=51.6MiB/s (54.1MB/s)(509MiB/9876msec); 0 zone resets 00:16:31.078 slat (usec): min=7, max=279, avg=15.84, stdev= 3.53 00:16:31.078 clat (usec): min=58, max=1782, avg=290.77, stdev=39.89 00:16:31.078 lat (usec): min=74, max=2062, avg=306.61, stdev=40.90 00:16:31.078 clat percentiles (usec): 00:16:31.078 | 50.000th=[ 297], 99.000th=[ 363], 99.900th=[ 594], 99.990th=[ 1123], 00:16:31.078 | 99.999th=[ 1696] 00:16:31.078 bw ( KiB/s): min=50512, max=54576, per=98.90%, avg=52239.58, stdev=1353.23, samples=19 00:16:31.078 iops : min=12628, max=13644, avg=13059.89, stdev=338.31, samples=19 00:16:31.078 lat (usec) : 20=0.01%, 50=0.01%, 100=16.61%, 250=39.76%, 500=43.55% 00:16:31.078 lat (usec) : 750=0.05%, 1000=0.02% 00:16:31.078 lat (msec) : 2=0.01% 00:16:31.078 cpu : usr=98.95%, sys=0.37%, ctx=19, majf=0, minf=13378 00:16:31.078 IO depths : 1=7.6%, 2=19.9%, 4=55.2%, 8=17.4%, 16=0.0%, 32=0.0%, >=64=0.0% 00:16:31.078 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:16:31.078 complete : 0=0.0%, 4=90.0%, 8=10.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:16:31.078 issued rwts: total=126131,130414,0,0 short=0,0,0,0 dropped=0,0,0,0 00:16:31.078 latency : target=0, window=0, percentile=100.00%, depth=8 00:16:31.078 00:16:31.078 Run status group 0 (all jobs): 00:16:31.078 READ: bw=49.3MiB/s (51.7MB/s), 49.3MiB/s-49.3MiB/s (51.7MB/s-51.7MB/s), io=493MiB (517MB), run=10000-10000msec 00:16:31.078 WRITE: bw=51.6MiB/s (54.1MB/s), 51.6MiB/s-51.6MiB/s (54.1MB/s-54.1MB/s), io=509MiB (534MB), run=9876-9876msec 00:16:31.078 ----------------------------------------------------- 00:16:31.078 Suppressions used: 00:16:31.078 count bytes template 00:16:31.078 1 7 /usr/src/fio/parse.c 00:16:31.078 363 34848 /usr/src/fio/iolog.c 00:16:31.078 1 8 libtcmalloc_minimal.so 00:16:31.078 1 904 libcrypto.so 00:16:31.078 ----------------------------------------------------- 00:16:31.078 00:16:31.078 00:16:31.078 real 0m11.426s 00:16:31.078 user 0m11.634s 00:16:31.078 sys 0m0.699s 00:16:31.078 23:50:18 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1130 -- # xtrace_disable 00:16:31.078 23:50:18 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@10 -- # set +x 00:16:31.078 ************************************ 00:16:31.078 END TEST bdev_fio_rw_verify 00:16:31.078 ************************************ 00:16:31.078 23:50:18 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@349 -- # rm -f 00:16:31.078 23:50:18 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@350 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio 00:16:31.078 23:50:18 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@353 -- # fio_config_gen /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio trim '' '' 00:16:31.078 23:50:18 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1284 -- # local config_file=/home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio 00:16:31.078 23:50:18 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1285 -- # local workload=trim 00:16:31.078 23:50:18 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1286 -- # local bdev_type= 00:16:31.078 23:50:18 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1287 -- # local env_context= 00:16:31.078 23:50:18 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1288 -- # local fio_dir=/usr/src/fio 00:16:31.078 23:50:18 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1290 -- # '[' -e /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio ']' 00:16:31.078 23:50:18 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1295 -- # '[' -z trim ']' 00:16:31.078 23:50:18 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1299 -- # '[' -n '' ']' 00:16:31.078 23:50:18 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1303 -- # touch /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio 00:16:31.078 23:50:18 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1305 -- # cat 00:16:31.078 23:50:18 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1317 -- # '[' trim == verify ']' 00:16:31.078 23:50:18 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1332 -- # '[' trim == trim ']' 00:16:31.078 23:50:18 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1333 -- # echo rw=trimwrite 00:16:31.079 23:50:18 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@354 -- # printf '%s\n' '{' ' "name": "raid5f",' ' "aliases": [' ' "ba0693a7-890f-4e27-87d5-edfe708395e9"' ' ],' ' "product_name": "Raid Volume",' ' "block_size": 512,' ' "num_blocks": 131072,' ' "uuid": "ba0693a7-890f-4e27-87d5-edfe708395e9",' ' "assigned_rate_limits": {' ' "rw_ios_per_sec": 0,' ' "rw_mbytes_per_sec": 0,' ' "r_mbytes_per_sec": 0,' ' "w_mbytes_per_sec": 0' ' },' ' "claimed": false,' ' "zoned": false,' ' "supported_io_types": {' ' "read": true,' ' "write": true,' ' "unmap": false,' ' "flush": false,' ' "reset": true,' ' "nvme_admin": false,' ' "nvme_io": false,' ' "nvme_io_md": false,' ' "write_zeroes": true,' ' "zcopy": false,' ' "get_zone_info": false,' ' "zone_management": false,' ' "zone_append": false,' ' "compare": false,' ' "compare_and_write": false,' ' "abort": false,' ' "seek_hole": false,' ' "seek_data": false,' ' "copy": false,' ' "nvme_iov_md": false' ' },' ' "driver_specific": {' ' "raid": {' ' "uuid": "ba0693a7-890f-4e27-87d5-edfe708395e9",' ' "strip_size_kb": 2,' ' "state": "online",' ' "raid_level": "raid5f",' ' "superblock": false,' ' "num_base_bdevs": 3,' ' "num_base_bdevs_discovered": 3,' ' "num_base_bdevs_operational": 3,' ' "base_bdevs_list": [' ' {' ' "name": "Malloc0",' ' "uuid": "7123a51f-862c-49a9-af5a-0bf8b025a1e3",' ' "is_configured": true,' ' "data_offset": 0,' ' "data_size": 65536' ' },' ' {' ' "name": "Malloc1",' ' "uuid": "b8891a8f-118b-4b0f-86f1-1be5692fb209",' ' "is_configured": true,' ' "data_offset": 0,' ' "data_size": 65536' ' },' ' {' ' "name": "Malloc2",' ' "uuid": "8c3c9d09-7a50-4d4d-a9cd-5c176f376d6f",' ' "is_configured": true,' ' "data_offset": 0,' ' "data_size": 65536' ' }' ' ]' ' }' ' }' '}' 00:16:31.079 23:50:18 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@354 -- # jq -r 'select(.supported_io_types.unmap == true) | .name' 00:16:31.079 23:50:18 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@354 -- # [[ -n '' ]] 00:16:31.079 23:50:18 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@360 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio 00:16:31.079 /home/vagrant/spdk_repo/spdk 00:16:31.079 23:50:18 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@361 -- # popd 00:16:31.079 23:50:18 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@362 -- # trap - SIGINT SIGTERM EXIT 00:16:31.079 23:50:18 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@363 -- # return 0 00:16:31.079 00:16:31.079 real 0m11.741s 00:16:31.079 user 0m11.767s 00:16:31.079 sys 0m0.853s 00:16:31.079 23:50:18 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1130 -- # xtrace_disable 00:16:31.079 23:50:18 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@10 -- # set +x 00:16:31.079 ************************************ 00:16:31.079 END TEST bdev_fio 00:16:31.079 ************************************ 00:16:31.079 23:50:18 blockdev_raid5f -- bdev/blockdev.sh@812 -- # trap cleanup SIGINT SIGTERM EXIT 00:16:31.079 23:50:18 blockdev_raid5f -- bdev/blockdev.sh@814 -- # run_test bdev_verify /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json -q 128 -o 4096 -w verify -t 5 -C -m 0x3 '' 00:16:31.079 23:50:18 blockdev_raid5f -- common/autotest_common.sh@1105 -- # '[' 16 -le 1 ']' 00:16:31.079 23:50:18 blockdev_raid5f -- common/autotest_common.sh@1111 -- # xtrace_disable 00:16:31.079 23:50:18 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:16:31.079 ************************************ 00:16:31.079 START TEST bdev_verify 00:16:31.079 ************************************ 00:16:31.079 23:50:18 blockdev_raid5f.bdev_verify -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json -q 128 -o 4096 -w verify -t 5 -C -m 0x3 '' 00:16:31.079 [2024-11-26 23:50:18.416003] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:16:31.079 [2024-11-26 23:50:18.416133] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x3 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid100370 ] 00:16:31.079 [2024-11-26 23:50:18.574842] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 2 00:16:31.079 [2024-11-26 23:50:18.627428] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:16:31.079 [2024-11-26 23:50:18.627524] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:16:31.079 Running I/O for 5 seconds... 00:16:32.959 10610.00 IOPS, 41.45 MiB/s [2024-11-26T23:50:22.030Z] 10620.00 IOPS, 41.48 MiB/s [2024-11-26T23:50:22.970Z] 10693.33 IOPS, 41.77 MiB/s [2024-11-26T23:50:23.915Z] 10705.75 IOPS, 41.82 MiB/s [2024-11-26T23:50:23.915Z] 10732.60 IOPS, 41.92 MiB/s 00:16:35.784 Latency(us) 00:16:35.784 [2024-11-26T23:50:23.916Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:16:35.784 Job: raid5f (Core Mask 0x1, workload: verify, depth: 128, IO size: 4096) 00:16:35.784 Verification LBA range: start 0x0 length 0x2000 00:16:35.784 raid5f : 5.02 6482.54 25.32 0.00 0.00 29668.58 228.95 20834.15 00:16:35.784 Job: raid5f (Core Mask 0x2, workload: verify, depth: 128, IO size: 4096) 00:16:35.784 Verification LBA range: start 0x2000 length 0x2000 00:16:35.784 raid5f : 5.03 4251.61 16.61 0.00 0.00 44993.39 255.78 32739.38 00:16:35.784 [2024-11-26T23:50:23.916Z] =================================================================================================================== 00:16:35.784 [2024-11-26T23:50:23.916Z] Total : 10734.16 41.93 0.00 0.00 35741.44 228.95 32739.38 00:16:36.355 00:16:36.355 real 0m5.938s 00:16:36.355 user 0m10.994s 00:16:36.355 sys 0m0.303s 00:16:36.355 23:50:24 blockdev_raid5f.bdev_verify -- common/autotest_common.sh@1130 -- # xtrace_disable 00:16:36.355 23:50:24 blockdev_raid5f.bdev_verify -- common/autotest_common.sh@10 -- # set +x 00:16:36.355 ************************************ 00:16:36.355 END TEST bdev_verify 00:16:36.355 ************************************ 00:16:36.355 23:50:24 blockdev_raid5f -- bdev/blockdev.sh@815 -- # run_test bdev_verify_big_io /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json -q 128 -o 65536 -w verify -t 5 -C -m 0x3 '' 00:16:36.355 23:50:24 blockdev_raid5f -- common/autotest_common.sh@1105 -- # '[' 16 -le 1 ']' 00:16:36.355 23:50:24 blockdev_raid5f -- common/autotest_common.sh@1111 -- # xtrace_disable 00:16:36.355 23:50:24 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:16:36.355 ************************************ 00:16:36.355 START TEST bdev_verify_big_io 00:16:36.355 ************************************ 00:16:36.355 23:50:24 blockdev_raid5f.bdev_verify_big_io -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json -q 128 -o 65536 -w verify -t 5 -C -m 0x3 '' 00:16:36.355 [2024-11-26 23:50:24.426049] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:16:36.355 [2024-11-26 23:50:24.426166] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x3 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid100456 ] 00:16:36.616 [2024-11-26 23:50:24.584948] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 2 00:16:36.616 [2024-11-26 23:50:24.631622] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:16:36.616 [2024-11-26 23:50:24.631667] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:16:36.876 Running I/O for 5 seconds... 00:16:39.201 633.00 IOPS, 39.56 MiB/s [2024-11-26T23:50:28.272Z] 760.00 IOPS, 47.50 MiB/s [2024-11-26T23:50:29.218Z] 760.67 IOPS, 47.54 MiB/s [2024-11-26T23:50:30.182Z] 777.00 IOPS, 48.56 MiB/s [2024-11-26T23:50:30.182Z] 774.00 IOPS, 48.38 MiB/s 00:16:42.050 Latency(us) 00:16:42.050 [2024-11-26T23:50:30.182Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:16:42.050 Job: raid5f (Core Mask 0x1, workload: verify, depth: 128, IO size: 65536) 00:16:42.050 Verification LBA range: start 0x0 length 0x200 00:16:42.050 raid5f : 5.21 450.29 28.14 0.00 0.00 7050386.87 172.60 304041.25 00:16:42.050 Job: raid5f (Core Mask 0x2, workload: verify, depth: 128, IO size: 65536) 00:16:42.050 Verification LBA range: start 0x200 length 0x200 00:16:42.050 raid5f : 5.20 341.87 21.37 0.00 0.00 9279474.28 305.86 386462.07 00:16:42.050 [2024-11-26T23:50:30.182Z] =================================================================================================================== 00:16:42.050 [2024-11-26T23:50:30.182Z] Total : 792.16 49.51 0.00 0.00 8010958.23 172.60 386462.07 00:16:42.620 00:16:42.620 real 0m6.132s 00:16:42.620 user 0m11.388s 00:16:42.620 sys 0m0.297s 00:16:42.620 23:50:30 blockdev_raid5f.bdev_verify_big_io -- common/autotest_common.sh@1130 -- # xtrace_disable 00:16:42.620 23:50:30 blockdev_raid5f.bdev_verify_big_io -- common/autotest_common.sh@10 -- # set +x 00:16:42.620 ************************************ 00:16:42.620 END TEST bdev_verify_big_io 00:16:42.620 ************************************ 00:16:42.620 23:50:30 blockdev_raid5f -- bdev/blockdev.sh@816 -- # run_test bdev_write_zeroes /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json -q 128 -o 4096 -w write_zeroes -t 1 '' 00:16:42.620 23:50:30 blockdev_raid5f -- common/autotest_common.sh@1105 -- # '[' 13 -le 1 ']' 00:16:42.620 23:50:30 blockdev_raid5f -- common/autotest_common.sh@1111 -- # xtrace_disable 00:16:42.620 23:50:30 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:16:42.620 ************************************ 00:16:42.620 START TEST bdev_write_zeroes 00:16:42.620 ************************************ 00:16:42.620 23:50:30 blockdev_raid5f.bdev_write_zeroes -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json -q 128 -o 4096 -w write_zeroes -t 1 '' 00:16:42.620 [2024-11-26 23:50:30.633234] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:16:42.620 [2024-11-26 23:50:30.633387] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid100539 ] 00:16:42.880 [2024-11-26 23:50:30.788483] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:16:42.880 [2024-11-26 23:50:30.834934] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:16:43.141 Running I/O for 1 seconds... 00:16:44.083 29631.00 IOPS, 115.75 MiB/s 00:16:44.083 Latency(us) 00:16:44.083 [2024-11-26T23:50:32.215Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:16:44.083 Job: raid5f (Core Mask 0x1, workload: write_zeroes, depth: 128, IO size: 4096) 00:16:44.083 raid5f : 1.01 29599.37 115.62 0.00 0.00 4311.78 1373.68 5838.14 00:16:44.083 [2024-11-26T23:50:32.215Z] =================================================================================================================== 00:16:44.083 [2024-11-26T23:50:32.215Z] Total : 29599.37 115.62 0.00 0.00 4311.78 1373.68 5838.14 00:16:44.343 00:16:44.343 real 0m1.895s 00:16:44.343 user 0m1.506s 00:16:44.343 sys 0m0.277s 00:16:44.343 23:50:32 blockdev_raid5f.bdev_write_zeroes -- common/autotest_common.sh@1130 -- # xtrace_disable 00:16:44.343 23:50:32 blockdev_raid5f.bdev_write_zeroes -- common/autotest_common.sh@10 -- # set +x 00:16:44.343 ************************************ 00:16:44.343 END TEST bdev_write_zeroes 00:16:44.343 ************************************ 00:16:44.602 23:50:32 blockdev_raid5f -- bdev/blockdev.sh@819 -- # run_test bdev_json_nonenclosed /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/nonenclosed.json -q 128 -o 4096 -w write_zeroes -t 1 '' 00:16:44.602 23:50:32 blockdev_raid5f -- common/autotest_common.sh@1105 -- # '[' 13 -le 1 ']' 00:16:44.602 23:50:32 blockdev_raid5f -- common/autotest_common.sh@1111 -- # xtrace_disable 00:16:44.602 23:50:32 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:16:44.602 ************************************ 00:16:44.602 START TEST bdev_json_nonenclosed 00:16:44.602 ************************************ 00:16:44.602 23:50:32 blockdev_raid5f.bdev_json_nonenclosed -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/nonenclosed.json -q 128 -o 4096 -w write_zeroes -t 1 '' 00:16:44.602 [2024-11-26 23:50:32.601750] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:16:44.602 [2024-11-26 23:50:32.601858] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid100581 ] 00:16:44.863 [2024-11-26 23:50:32.756644] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:16:44.863 [2024-11-26 23:50:32.803184] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:16:44.863 [2024-11-26 23:50:32.803303] json_config.c: 608:json_config_prepare_ctx: *ERROR*: Invalid JSON configuration: not enclosed in {}. 00:16:44.863 [2024-11-26 23:50:32.803325] rpc.c: 190:spdk_rpc_server_finish: *ERROR*: No server listening on provided address: 00:16:44.863 [2024-11-26 23:50:32.803350] app.c:1064:spdk_app_stop: *WARNING*: spdk_app_stop'd on non-zero 00:16:44.863 00:16:44.863 real 0m0.393s 00:16:44.863 user 0m0.169s 00:16:44.863 sys 0m0.120s 00:16:44.863 23:50:32 blockdev_raid5f.bdev_json_nonenclosed -- common/autotest_common.sh@1130 -- # xtrace_disable 00:16:44.863 23:50:32 blockdev_raid5f.bdev_json_nonenclosed -- common/autotest_common.sh@10 -- # set +x 00:16:44.863 ************************************ 00:16:44.863 END TEST bdev_json_nonenclosed 00:16:44.863 ************************************ 00:16:44.863 23:50:32 blockdev_raid5f -- bdev/blockdev.sh@822 -- # run_test bdev_json_nonarray /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/nonarray.json -q 128 -o 4096 -w write_zeroes -t 1 '' 00:16:44.863 23:50:32 blockdev_raid5f -- common/autotest_common.sh@1105 -- # '[' 13 -le 1 ']' 00:16:44.863 23:50:32 blockdev_raid5f -- common/autotest_common.sh@1111 -- # xtrace_disable 00:16:44.863 23:50:32 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:16:44.863 ************************************ 00:16:44.863 START TEST bdev_json_nonarray 00:16:44.863 ************************************ 00:16:44.863 23:50:32 blockdev_raid5f.bdev_json_nonarray -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/nonarray.json -q 128 -o 4096 -w write_zeroes -t 1 '' 00:16:45.123 [2024-11-26 23:50:33.067442] Starting SPDK v25.01-pre git sha1 2f2acf4eb / DPDK 22.11.4 initialization... 00:16:45.123 [2024-11-26 23:50:33.067560] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid100607 ] 00:16:45.123 [2024-11-26 23:50:33.221748] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:16:45.383 [2024-11-26 23:50:33.272619] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:16:45.383 [2024-11-26 23:50:33.272745] json_config.c: 614:json_config_prepare_ctx: *ERROR*: Invalid JSON configuration: 'subsystems' should be an array. 00:16:45.383 [2024-11-26 23:50:33.272774] rpc.c: 190:spdk_rpc_server_finish: *ERROR*: No server listening on provided address: 00:16:45.383 [2024-11-26 23:50:33.272788] app.c:1064:spdk_app_stop: *WARNING*: spdk_app_stop'd on non-zero 00:16:45.383 00:16:45.383 real 0m0.395s 00:16:45.383 user 0m0.162s 00:16:45.383 sys 0m0.129s 00:16:45.383 23:50:33 blockdev_raid5f.bdev_json_nonarray -- common/autotest_common.sh@1130 -- # xtrace_disable 00:16:45.383 23:50:33 blockdev_raid5f.bdev_json_nonarray -- common/autotest_common.sh@10 -- # set +x 00:16:45.383 ************************************ 00:16:45.383 END TEST bdev_json_nonarray 00:16:45.383 ************************************ 00:16:45.383 23:50:33 blockdev_raid5f -- bdev/blockdev.sh@824 -- # [[ raid5f == bdev ]] 00:16:45.383 23:50:33 blockdev_raid5f -- bdev/blockdev.sh@832 -- # [[ raid5f == gpt ]] 00:16:45.383 23:50:33 blockdev_raid5f -- bdev/blockdev.sh@836 -- # [[ raid5f == crypto_sw ]] 00:16:45.383 23:50:33 blockdev_raid5f -- bdev/blockdev.sh@848 -- # trap - SIGINT SIGTERM EXIT 00:16:45.383 23:50:33 blockdev_raid5f -- bdev/blockdev.sh@849 -- # cleanup 00:16:45.383 23:50:33 blockdev_raid5f -- bdev/blockdev.sh@23 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/aiofile 00:16:45.383 23:50:33 blockdev_raid5f -- bdev/blockdev.sh@24 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json 00:16:45.383 23:50:33 blockdev_raid5f -- bdev/blockdev.sh@26 -- # [[ raid5f == rbd ]] 00:16:45.383 23:50:33 blockdev_raid5f -- bdev/blockdev.sh@30 -- # [[ raid5f == daos ]] 00:16:45.383 23:50:33 blockdev_raid5f -- bdev/blockdev.sh@34 -- # [[ raid5f = \g\p\t ]] 00:16:45.383 23:50:33 blockdev_raid5f -- bdev/blockdev.sh@40 -- # [[ raid5f == xnvme ]] 00:16:45.383 00:16:45.383 real 0m36.218s 00:16:45.383 user 0m48.531s 00:16:45.383 sys 0m5.295s 00:16:45.383 23:50:33 blockdev_raid5f -- common/autotest_common.sh@1130 -- # xtrace_disable 00:16:45.383 23:50:33 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:16:45.383 ************************************ 00:16:45.383 END TEST blockdev_raid5f 00:16:45.383 ************************************ 00:16:45.642 23:50:33 -- spdk/autotest.sh@194 -- # uname -s 00:16:45.642 23:50:33 -- spdk/autotest.sh@194 -- # [[ Linux == Linux ]] 00:16:45.642 23:50:33 -- spdk/autotest.sh@195 -- # [[ 0 -eq 1 ]] 00:16:45.642 23:50:33 -- spdk/autotest.sh@195 -- # [[ 0 -eq 1 ]] 00:16:45.642 23:50:33 -- spdk/autotest.sh@207 -- # '[' 0 -eq 1 ']' 00:16:45.642 23:50:33 -- spdk/autotest.sh@256 -- # '[' 0 -eq 1 ']' 00:16:45.642 23:50:33 -- spdk/autotest.sh@260 -- # timing_exit lib 00:16:45.642 23:50:33 -- common/autotest_common.sh@732 -- # xtrace_disable 00:16:45.642 23:50:33 -- common/autotest_common.sh@10 -- # set +x 00:16:45.642 23:50:33 -- spdk/autotest.sh@262 -- # '[' 0 -eq 1 ']' 00:16:45.642 23:50:33 -- spdk/autotest.sh@267 -- # '[' 0 -eq 1 ']' 00:16:45.642 23:50:33 -- spdk/autotest.sh@276 -- # '[' 0 -eq 1 ']' 00:16:45.642 23:50:33 -- spdk/autotest.sh@311 -- # '[' 0 -eq 1 ']' 00:16:45.642 23:50:33 -- spdk/autotest.sh@315 -- # '[' 0 -eq 1 ']' 00:16:45.642 23:50:33 -- spdk/autotest.sh@319 -- # '[' 0 -eq 1 ']' 00:16:45.642 23:50:33 -- spdk/autotest.sh@324 -- # '[' 0 -eq 1 ']' 00:16:45.642 23:50:33 -- spdk/autotest.sh@333 -- # '[' 0 -eq 1 ']' 00:16:45.642 23:50:33 -- spdk/autotest.sh@338 -- # '[' 0 -eq 1 ']' 00:16:45.642 23:50:33 -- spdk/autotest.sh@342 -- # '[' 0 -eq 1 ']' 00:16:45.642 23:50:33 -- spdk/autotest.sh@346 -- # '[' 0 -eq 1 ']' 00:16:45.642 23:50:33 -- spdk/autotest.sh@350 -- # '[' 0 -eq 1 ']' 00:16:45.642 23:50:33 -- spdk/autotest.sh@355 -- # '[' 0 -eq 1 ']' 00:16:45.642 23:50:33 -- spdk/autotest.sh@359 -- # '[' 0 -eq 1 ']' 00:16:45.642 23:50:33 -- spdk/autotest.sh@366 -- # [[ 0 -eq 1 ]] 00:16:45.642 23:50:33 -- spdk/autotest.sh@370 -- # [[ 0 -eq 1 ]] 00:16:45.642 23:50:33 -- spdk/autotest.sh@374 -- # [[ 0 -eq 1 ]] 00:16:45.642 23:50:33 -- spdk/autotest.sh@378 -- # [[ '' -eq 1 ]] 00:16:45.642 23:50:33 -- spdk/autotest.sh@385 -- # trap - SIGINT SIGTERM EXIT 00:16:45.642 23:50:33 -- spdk/autotest.sh@387 -- # timing_enter post_cleanup 00:16:45.642 23:50:33 -- common/autotest_common.sh@726 -- # xtrace_disable 00:16:45.642 23:50:33 -- common/autotest_common.sh@10 -- # set +x 00:16:45.642 23:50:33 -- spdk/autotest.sh@388 -- # autotest_cleanup 00:16:45.642 23:50:33 -- common/autotest_common.sh@1396 -- # local autotest_es=0 00:16:45.642 23:50:33 -- common/autotest_common.sh@1397 -- # xtrace_disable 00:16:45.642 23:50:33 -- common/autotest_common.sh@10 -- # set +x 00:16:48.182 INFO: APP EXITING 00:16:48.182 INFO: killing all VMs 00:16:48.182 INFO: killing vhost app 00:16:48.182 INFO: EXIT DONE 00:16:48.442 0000:00:03.0 (1af4 1001): Active devices: mount@vda:vda2,mount@vda:vda3,mount@vda:vda5, so not binding PCI dev 00:16:48.442 Waiting for block devices as requested 00:16:48.702 0000:00:11.0 (1b36 0010): uio_pci_generic -> nvme 00:16:48.702 0000:00:10.0 (1b36 0010): uio_pci_generic -> nvme 00:16:49.642 0000:00:03.0 (1af4 1001): Active devices: mount@vda:vda2,mount@vda:vda3,mount@vda:vda5, so not binding PCI dev 00:16:49.642 Cleaning 00:16:49.642 Removing: /var/run/dpdk/spdk0/config 00:16:49.642 Removing: /var/run/dpdk/spdk0/fbarray_memseg-2048k-0-0 00:16:49.642 Removing: /var/run/dpdk/spdk0/fbarray_memseg-2048k-0-1 00:16:49.642 Removing: /var/run/dpdk/spdk0/fbarray_memseg-2048k-0-2 00:16:49.642 Removing: /var/run/dpdk/spdk0/fbarray_memseg-2048k-0-3 00:16:49.642 Removing: /var/run/dpdk/spdk0/fbarray_memzone 00:16:49.642 Removing: /var/run/dpdk/spdk0/hugepage_info 00:16:49.642 Removing: /dev/shm/spdk_tgt_trace.pid68854 00:16:49.642 Removing: /var/run/dpdk/spdk0 00:16:49.642 Removing: /var/run/dpdk/spdk_pid100207 00:16:49.642 Removing: /var/run/dpdk/spdk_pid100370 00:16:49.642 Removing: /var/run/dpdk/spdk_pid100456 00:16:49.642 Removing: /var/run/dpdk/spdk_pid100539 00:16:49.642 Removing: /var/run/dpdk/spdk_pid100581 00:16:49.642 Removing: /var/run/dpdk/spdk_pid100607 00:16:49.642 Removing: /var/run/dpdk/spdk_pid68690 00:16:49.642 Removing: /var/run/dpdk/spdk_pid68854 00:16:49.642 Removing: /var/run/dpdk/spdk_pid69061 00:16:49.642 Removing: /var/run/dpdk/spdk_pid69143 00:16:49.642 Removing: /var/run/dpdk/spdk_pid69177 00:16:49.642 Removing: /var/run/dpdk/spdk_pid69283 00:16:49.642 Removing: /var/run/dpdk/spdk_pid69301 00:16:49.642 Removing: /var/run/dpdk/spdk_pid69489 00:16:49.642 Removing: /var/run/dpdk/spdk_pid69564 00:16:49.642 Removing: /var/run/dpdk/spdk_pid69642 00:16:49.642 Removing: /var/run/dpdk/spdk_pid69742 00:16:49.642 Removing: /var/run/dpdk/spdk_pid69828 00:16:49.642 Removing: /var/run/dpdk/spdk_pid69862 00:16:49.642 Removing: /var/run/dpdk/spdk_pid69903 00:16:49.642 Removing: /var/run/dpdk/spdk_pid69969 00:16:49.642 Removing: /var/run/dpdk/spdk_pid70086 00:16:49.642 Removing: /var/run/dpdk/spdk_pid70506 00:16:49.902 Removing: /var/run/dpdk/spdk_pid70559 00:16:49.902 Removing: /var/run/dpdk/spdk_pid70600 00:16:49.902 Removing: /var/run/dpdk/spdk_pid70616 00:16:49.902 Removing: /var/run/dpdk/spdk_pid70685 00:16:49.902 Removing: /var/run/dpdk/spdk_pid70695 00:16:49.902 Removing: /var/run/dpdk/spdk_pid70762 00:16:49.902 Removing: /var/run/dpdk/spdk_pid70778 00:16:49.902 Removing: /var/run/dpdk/spdk_pid70820 00:16:49.902 Removing: /var/run/dpdk/spdk_pid70838 00:16:49.902 Removing: /var/run/dpdk/spdk_pid70880 00:16:49.902 Removing: /var/run/dpdk/spdk_pid70898 00:16:49.902 Removing: /var/run/dpdk/spdk_pid71036 00:16:49.902 Removing: /var/run/dpdk/spdk_pid71067 00:16:49.902 Removing: /var/run/dpdk/spdk_pid71151 00:16:49.902 Removing: /var/run/dpdk/spdk_pid72316 00:16:49.902 Removing: /var/run/dpdk/spdk_pid72511 00:16:49.902 Removing: /var/run/dpdk/spdk_pid72640 00:16:49.902 Removing: /var/run/dpdk/spdk_pid73239 00:16:49.903 Removing: /var/run/dpdk/spdk_pid73440 00:16:49.903 Removing: /var/run/dpdk/spdk_pid73569 00:16:49.903 Removing: /var/run/dpdk/spdk_pid74179 00:16:49.903 Removing: /var/run/dpdk/spdk_pid74487 00:16:49.903 Removing: /var/run/dpdk/spdk_pid74616 00:16:49.903 Removing: /var/run/dpdk/spdk_pid75957 00:16:49.903 Removing: /var/run/dpdk/spdk_pid76199 00:16:49.903 Removing: /var/run/dpdk/spdk_pid76328 00:16:49.903 Removing: /var/run/dpdk/spdk_pid77658 00:16:49.903 Removing: /var/run/dpdk/spdk_pid77900 00:16:49.903 Removing: /var/run/dpdk/spdk_pid78029 00:16:49.903 Removing: /var/run/dpdk/spdk_pid79370 00:16:49.903 Removing: /var/run/dpdk/spdk_pid79805 00:16:49.903 Removing: /var/run/dpdk/spdk_pid79938 00:16:49.903 Removing: /var/run/dpdk/spdk_pid81371 00:16:49.903 Removing: /var/run/dpdk/spdk_pid81619 00:16:49.903 Removing: /var/run/dpdk/spdk_pid81755 00:16:49.903 Removing: /var/run/dpdk/spdk_pid83175 00:16:49.903 Removing: /var/run/dpdk/spdk_pid83422 00:16:49.903 Removing: /var/run/dpdk/spdk_pid83563 00:16:49.903 Removing: /var/run/dpdk/spdk_pid84983 00:16:49.903 Removing: /var/run/dpdk/spdk_pid85465 00:16:49.903 Removing: /var/run/dpdk/spdk_pid85600 00:16:49.903 Removing: /var/run/dpdk/spdk_pid85728 00:16:49.903 Removing: /var/run/dpdk/spdk_pid86130 00:16:49.903 Removing: /var/run/dpdk/spdk_pid86839 00:16:49.903 Removing: /var/run/dpdk/spdk_pid87201 00:16:49.903 Removing: /var/run/dpdk/spdk_pid87876 00:16:49.903 Removing: /var/run/dpdk/spdk_pid88300 00:16:49.903 Removing: /var/run/dpdk/spdk_pid89036 00:16:49.903 Removing: /var/run/dpdk/spdk_pid89429 00:16:49.903 Removing: /var/run/dpdk/spdk_pid91345 00:16:49.903 Removing: /var/run/dpdk/spdk_pid91772 00:16:49.903 Removing: /var/run/dpdk/spdk_pid92190 00:16:49.903 Removing: /var/run/dpdk/spdk_pid94212 00:16:49.903 Removing: /var/run/dpdk/spdk_pid94682 00:16:49.903 Removing: /var/run/dpdk/spdk_pid95180 00:16:50.163 Removing: /var/run/dpdk/spdk_pid96208 00:16:50.163 Removing: /var/run/dpdk/spdk_pid96521 00:16:50.163 Removing: /var/run/dpdk/spdk_pid97430 00:16:50.163 Removing: /var/run/dpdk/spdk_pid97742 00:16:50.163 Removing: /var/run/dpdk/spdk_pid98657 00:16:50.163 Removing: /var/run/dpdk/spdk_pid98974 00:16:50.163 Removing: /var/run/dpdk/spdk_pid99645 00:16:50.163 Removing: /var/run/dpdk/spdk_pid99903 00:16:50.163 Removing: /var/run/dpdk/spdk_pid99943 00:16:50.163 Removing: /var/run/dpdk/spdk_pid99974 00:16:50.163 Clean 00:16:50.163 23:50:38 -- common/autotest_common.sh@1453 -- # return 0 00:16:50.163 23:50:38 -- spdk/autotest.sh@389 -- # timing_exit post_cleanup 00:16:50.163 23:50:38 -- common/autotest_common.sh@732 -- # xtrace_disable 00:16:50.163 23:50:38 -- common/autotest_common.sh@10 -- # set +x 00:16:50.163 23:50:38 -- spdk/autotest.sh@391 -- # timing_exit autotest 00:16:50.163 23:50:38 -- common/autotest_common.sh@732 -- # xtrace_disable 00:16:50.163 23:50:38 -- common/autotest_common.sh@10 -- # set +x 00:16:50.163 23:50:38 -- spdk/autotest.sh@392 -- # chmod a+r /home/vagrant/spdk_repo/spdk/../output/timing.txt 00:16:50.163 23:50:38 -- spdk/autotest.sh@394 -- # [[ -f /home/vagrant/spdk_repo/spdk/../output/udev.log ]] 00:16:50.163 23:50:38 -- spdk/autotest.sh@394 -- # rm -f /home/vagrant/spdk_repo/spdk/../output/udev.log 00:16:50.163 23:50:38 -- spdk/autotest.sh@396 -- # [[ y == y ]] 00:16:50.163 23:50:38 -- spdk/autotest.sh@398 -- # hostname 00:16:50.163 23:50:38 -- spdk/autotest.sh@398 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -c --no-external -d /home/vagrant/spdk_repo/spdk -t fedora39-cloud-1721788873-2326 -o /home/vagrant/spdk_repo/spdk/../output/cov_test.info 00:16:50.424 geninfo: WARNING: invalid characters removed from testname! 00:17:16.993 23:51:04 -- spdk/autotest.sh@399 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -a /home/vagrant/spdk_repo/spdk/../output/cov_base.info -a /home/vagrant/spdk_repo/spdk/../output/cov_test.info -o /home/vagrant/spdk_repo/spdk/../output/cov_total.info 00:17:18.902 23:51:06 -- spdk/autotest.sh@400 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -r /home/vagrant/spdk_repo/spdk/../output/cov_total.info '*/dpdk/*' -o /home/vagrant/spdk_repo/spdk/../output/cov_total.info 00:17:20.811 23:51:08 -- spdk/autotest.sh@404 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -r /home/vagrant/spdk_repo/spdk/../output/cov_total.info --ignore-errors unused,unused '/usr/*' -o /home/vagrant/spdk_repo/spdk/../output/cov_total.info 00:17:22.720 23:51:10 -- spdk/autotest.sh@405 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -r /home/vagrant/spdk_repo/spdk/../output/cov_total.info '*/examples/vmd/*' -o /home/vagrant/spdk_repo/spdk/../output/cov_total.info 00:17:24.634 23:51:12 -- spdk/autotest.sh@406 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -r /home/vagrant/spdk_repo/spdk/../output/cov_total.info '*/app/spdk_lspci/*' -o /home/vagrant/spdk_repo/spdk/../output/cov_total.info 00:17:26.544 23:51:14 -- spdk/autotest.sh@407 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -r /home/vagrant/spdk_repo/spdk/../output/cov_total.info '*/app/spdk_top/*' -o /home/vagrant/spdk_repo/spdk/../output/cov_total.info 00:17:28.454 23:51:16 -- spdk/autotest.sh@408 -- # rm -f cov_base.info cov_test.info OLD_STDOUT OLD_STDERR 00:17:28.454 23:51:16 -- spdk/autorun.sh@1 -- $ timing_finish 00:17:28.454 23:51:16 -- common/autotest_common.sh@738 -- $ [[ -e /home/vagrant/spdk_repo/spdk/../output/timing.txt ]] 00:17:28.454 23:51:16 -- common/autotest_common.sh@740 -- $ flamegraph=/usr/local/FlameGraph/flamegraph.pl 00:17:28.454 23:51:16 -- common/autotest_common.sh@741 -- $ [[ -x /usr/local/FlameGraph/flamegraph.pl ]] 00:17:28.454 23:51:16 -- common/autotest_common.sh@744 -- $ /usr/local/FlameGraph/flamegraph.pl --title 'Build Timing' --nametype Step: --countname seconds /home/vagrant/spdk_repo/spdk/../output/timing.txt 00:17:28.454 + [[ -n 6157 ]] 00:17:28.454 + sudo kill 6157 00:17:28.465 [Pipeline] } 00:17:28.481 [Pipeline] // timeout 00:17:28.488 [Pipeline] } 00:17:28.505 [Pipeline] // stage 00:17:28.512 [Pipeline] } 00:17:28.527 [Pipeline] // catchError 00:17:28.536 [Pipeline] stage 00:17:28.538 [Pipeline] { (Stop VM) 00:17:28.550 [Pipeline] sh 00:17:28.862 + vagrant halt 00:17:31.404 ==> default: Halting domain... 00:17:39.552 [Pipeline] sh 00:17:39.838 + vagrant destroy -f 00:17:42.385 ==> default: Removing domain... 00:17:42.399 [Pipeline] sh 00:17:42.684 + mv output /var/jenkins/workspace/raid-vg-autotest/output 00:17:42.695 [Pipeline] } 00:17:42.710 [Pipeline] // stage 00:17:42.715 [Pipeline] } 00:17:42.730 [Pipeline] // dir 00:17:42.735 [Pipeline] } 00:17:42.749 [Pipeline] // wrap 00:17:42.755 [Pipeline] } 00:17:42.768 [Pipeline] // catchError 00:17:42.777 [Pipeline] stage 00:17:42.779 [Pipeline] { (Epilogue) 00:17:42.792 [Pipeline] sh 00:17:43.077 + jbp/jenkins/jjb-config/jobs/scripts/compress_artifacts.sh 00:17:47.312 [Pipeline] catchError 00:17:47.315 [Pipeline] { 00:17:47.329 [Pipeline] sh 00:17:47.615 + jbp/jenkins/jjb-config/jobs/scripts/check_artifacts_size.sh 00:17:47.615 Artifacts sizes are good 00:17:47.625 [Pipeline] } 00:17:47.640 [Pipeline] // catchError 00:17:47.651 [Pipeline] archiveArtifacts 00:17:47.659 Archiving artifacts 00:17:47.787 [Pipeline] cleanWs 00:17:47.840 [WS-CLEANUP] Deleting project workspace... 00:17:47.840 [WS-CLEANUP] Deferred wipeout is used... 00:17:47.847 [WS-CLEANUP] done 00:17:47.849 [Pipeline] } 00:17:47.863 [Pipeline] // stage 00:17:47.867 [Pipeline] } 00:17:47.880 [Pipeline] // node 00:17:47.884 [Pipeline] End of Pipeline 00:17:47.921 Finished: SUCCESS